Mar 2 13:17:45.223046 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 2 11:01:37 -00 2026 Mar 2 13:17:45.223079 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5b1ae8ad250cf3ddd00dc7c63ded260e5b82ee29f2cdc578a6ade4cab26e6a0b Mar 2 13:17:45.223097 kernel: BIOS-provided physical RAM map: Mar 2 13:17:45.223107 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 2 13:17:45.223117 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 2 13:17:45.223126 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 2 13:17:45.223138 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Mar 2 13:17:45.223148 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 2 13:17:45.223158 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Mar 2 13:17:45.223168 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Mar 2 13:17:45.223183 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Mar 2 13:17:45.223193 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Mar 2 13:17:45.223204 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Mar 2 13:17:45.223214 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Mar 2 13:17:45.223226 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Mar 2 13:17:45.223238 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 2 13:17:45.223252 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Mar 2 13:17:45.223263 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Mar 2 13:17:45.223274 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 2 13:17:45.223285 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Mar 2 13:17:45.223295 kernel: NX (Execute Disable) protection: active Mar 2 13:17:45.223306 kernel: APIC: Static calls initialized Mar 2 13:17:45.223316 kernel: efi: EFI v2.7 by EDK II Mar 2 13:17:45.223360 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b675198 Mar 2 13:17:45.223371 kernel: SMBIOS 2.8 present. Mar 2 13:17:45.223381 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 Mar 2 13:17:45.223392 kernel: Hypervisor detected: KVM Mar 2 13:17:45.223408 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 2 13:17:45.223419 kernel: kvm-clock: using sched offset of 13524010791 cycles Mar 2 13:17:45.223431 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 2 13:17:45.223442 kernel: tsc: Detected 2445.424 MHz processor Mar 2 13:17:45.223453 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 2 13:17:45.223465 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 2 13:17:45.223475 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 Mar 2 13:17:45.223487 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Mar 2 13:17:45.223499 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 2 13:17:45.223516 kernel: Using GB pages for direct mapping Mar 2 13:17:45.223526 kernel: Secure boot disabled Mar 2 13:17:45.223538 kernel: ACPI: Early table checksum verification disabled Mar 2 13:17:45.223549 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Mar 2 13:17:45.223568 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Mar 2 13:17:45.223581 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223592 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223609 kernel: ACPI: FACS 0x000000009CBDD000 000040 Mar 2 13:17:45.223620 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223631 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223643 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223654 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 2 13:17:45.223666 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Mar 2 13:17:45.223679 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Mar 2 13:17:45.223694 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Mar 2 13:17:45.223706 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Mar 2 13:17:45.223718 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Mar 2 13:17:45.223729 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Mar 2 13:17:45.223741 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Mar 2 13:17:45.223753 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Mar 2 13:17:45.223763 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Mar 2 13:17:45.223821 kernel: No NUMA configuration found Mar 2 13:17:45.223833 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Mar 2 13:17:45.223851 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Mar 2 13:17:45.223863 kernel: Zone ranges: Mar 2 13:17:45.223874 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 2 13:17:45.223886 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Mar 2 13:17:45.223897 kernel: Normal empty Mar 2 13:17:45.223909 kernel: Movable zone start for each node Mar 2 13:17:45.223921 kernel: Early memory node ranges Mar 2 13:17:45.223932 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 2 13:17:45.223944 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Mar 2 13:17:45.223956 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Mar 2 13:17:45.223972 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Mar 2 13:17:45.223984 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Mar 2 13:17:45.223995 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Mar 2 13:17:45.224006 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Mar 2 13:17:45.224018 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 2 13:17:45.224029 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 2 13:17:45.224041 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Mar 2 13:17:45.224053 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 2 13:17:45.224065 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Mar 2 13:17:45.224081 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Mar 2 13:17:45.224093 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Mar 2 13:17:45.224105 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 2 13:17:45.224115 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 2 13:17:45.224127 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 2 13:17:45.224139 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 2 13:17:45.224150 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 2 13:17:45.224162 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 2 13:17:45.224174 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 2 13:17:45.224190 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 2 13:17:45.224202 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 2 13:17:45.224213 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 2 13:17:45.224225 kernel: TSC deadline timer available Mar 2 13:17:45.224237 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Mar 2 13:17:45.224248 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 2 13:17:45.224260 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 2 13:17:45.224273 kernel: kvm-guest: setup PV sched yield Mar 2 13:17:45.224285 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Mar 2 13:17:45.224302 kernel: Booting paravirtualized kernel on KVM Mar 2 13:17:45.224315 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 2 13:17:45.224365 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 2 13:17:45.224379 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u524288 Mar 2 13:17:45.224391 kernel: pcpu-alloc: s196328 r8192 d28952 u524288 alloc=1*2097152 Mar 2 13:17:45.224403 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 2 13:17:45.224414 kernel: kvm-guest: PV spinlocks enabled Mar 2 13:17:45.224450 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 2 13:17:45.224463 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5b1ae8ad250cf3ddd00dc7c63ded260e5b82ee29f2cdc578a6ade4cab26e6a0b Mar 2 13:17:45.224505 kernel: random: crng init done Mar 2 13:17:45.224515 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 2 13:17:45.224526 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 2 13:17:45.224536 kernel: Fallback order for Node 0: 0 Mar 2 13:17:45.224546 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Mar 2 13:17:45.224582 kernel: Policy zone: DMA32 Mar 2 13:17:45.224593 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 2 13:17:45.224630 kernel: Memory: 2400616K/2567000K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42892K init, 2304K bss, 166124K reserved, 0K cma-reserved) Mar 2 13:17:45.224650 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 2 13:17:45.224662 kernel: ftrace: allocating 37996 entries in 149 pages Mar 2 13:17:45.224675 kernel: ftrace: allocated 149 pages with 4 groups Mar 2 13:17:45.224710 kernel: Dynamic Preempt: voluntary Mar 2 13:17:45.224723 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 2 13:17:45.224753 kernel: rcu: RCU event tracing is enabled. Mar 2 13:17:45.224869 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 2 13:17:45.224887 kernel: Trampoline variant of Tasks RCU enabled. Mar 2 13:17:45.224900 kernel: Rude variant of Tasks RCU enabled. Mar 2 13:17:45.224912 kernel: Tracing variant of Tasks RCU enabled. Mar 2 13:17:45.224924 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 2 13:17:45.224936 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 2 13:17:45.224955 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 2 13:17:45.224967 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 2 13:17:45.224979 kernel: Console: colour dummy device 80x25 Mar 2 13:17:45.224991 kernel: printk: console [ttyS0] enabled Mar 2 13:17:45.225003 kernel: ACPI: Core revision 20230628 Mar 2 13:17:45.225020 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 2 13:17:45.225032 kernel: APIC: Switch to symmetric I/O mode setup Mar 2 13:17:45.225044 kernel: x2apic enabled Mar 2 13:17:45.225056 kernel: APIC: Switched APIC routing to: physical x2apic Mar 2 13:17:45.225069 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 2 13:17:45.225081 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 2 13:17:45.225094 kernel: kvm-guest: setup PV IPIs Mar 2 13:17:45.225106 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 2 13:17:45.225118 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Mar 2 13:17:45.225136 kernel: Calibrating delay loop (skipped) preset value.. 4890.84 BogoMIPS (lpj=2445424) Mar 2 13:17:45.225148 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 2 13:17:45.225161 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 2 13:17:45.225173 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 2 13:17:45.225185 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 2 13:17:45.225198 kernel: Spectre V2 : Mitigation: Retpolines Mar 2 13:17:45.225210 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Mar 2 13:17:45.225222 kernel: Speculative Store Bypass: Vulnerable Mar 2 13:17:45.225236 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 2 13:17:45.225252 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 2 13:17:45.225266 kernel: active return thunk: srso_alias_return_thunk Mar 2 13:17:45.225278 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 2 13:17:45.225290 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Mar 2 13:17:45.225302 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Mar 2 13:17:45.225314 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 2 13:17:45.225363 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 2 13:17:45.225376 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 2 13:17:45.225395 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 2 13:17:45.225409 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 2 13:17:45.225422 kernel: Freeing SMP alternatives memory: 32K Mar 2 13:17:45.225434 kernel: pid_max: default: 32768 minimum: 301 Mar 2 13:17:45.225445 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 2 13:17:45.225458 kernel: landlock: Up and running. Mar 2 13:17:45.225470 kernel: SELinux: Initializing. Mar 2 13:17:45.225481 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 13:17:45.225494 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 13:17:45.225511 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Mar 2 13:17:45.225524 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:17:45.225536 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:17:45.225548 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 2 13:17:45.225561 kernel: Performance Events: PMU not available due to virtualization, using software events only. Mar 2 13:17:45.225573 kernel: signal: max sigframe size: 1776 Mar 2 13:17:45.225586 kernel: rcu: Hierarchical SRCU implementation. Mar 2 13:17:45.225599 kernel: rcu: Max phase no-delay instances is 400. Mar 2 13:17:45.225610 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 2 13:17:45.225629 kernel: smp: Bringing up secondary CPUs ... Mar 2 13:17:45.225641 kernel: smpboot: x86: Booting SMP configuration: Mar 2 13:17:45.225653 kernel: .... node #0, CPUs: #1 #2 #3 Mar 2 13:17:45.225666 kernel: smp: Brought up 1 node, 4 CPUs Mar 2 13:17:45.225677 kernel: smpboot: Max logical packages: 1 Mar 2 13:17:45.225690 kernel: smpboot: Total of 4 processors activated (19563.39 BogoMIPS) Mar 2 13:17:45.225702 kernel: devtmpfs: initialized Mar 2 13:17:45.225714 kernel: x86/mm: Memory block size: 128MB Mar 2 13:17:45.225726 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Mar 2 13:17:45.225744 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Mar 2 13:17:45.225756 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Mar 2 13:17:45.225769 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Mar 2 13:17:45.225831 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Mar 2 13:17:45.225845 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 2 13:17:45.225857 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 2 13:17:45.225869 kernel: pinctrl core: initialized pinctrl subsystem Mar 2 13:17:45.225882 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 2 13:17:45.225893 kernel: audit: initializing netlink subsys (disabled) Mar 2 13:17:45.225911 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 2 13:17:45.225924 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 2 13:17:45.225935 kernel: audit: type=2000 audit(1772457462.076:1): state=initialized audit_enabled=0 res=1 Mar 2 13:17:45.225948 kernel: cpuidle: using governor menu Mar 2 13:17:45.225960 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 2 13:17:45.225972 kernel: dca service started, version 1.12.1 Mar 2 13:17:45.225984 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Mar 2 13:17:45.225997 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Mar 2 13:17:45.226010 kernel: PCI: Using configuration type 1 for base access Mar 2 13:17:45.226028 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 2 13:17:45.226041 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 2 13:17:45.226053 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 2 13:17:45.226065 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 2 13:17:45.226078 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 2 13:17:45.226089 kernel: ACPI: Added _OSI(Module Device) Mar 2 13:17:45.226101 kernel: ACPI: Added _OSI(Processor Device) Mar 2 13:17:45.226114 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 2 13:17:45.226125 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 2 13:17:45.226144 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 2 13:17:45.226156 kernel: ACPI: Interpreter enabled Mar 2 13:17:45.226168 kernel: ACPI: PM: (supports S0 S3 S5) Mar 2 13:17:45.226181 kernel: ACPI: Using IOAPIC for interrupt routing Mar 2 13:17:45.226192 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 2 13:17:45.226205 kernel: PCI: Using E820 reservations for host bridge windows Mar 2 13:17:45.226218 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 2 13:17:45.226229 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 2 13:17:45.226538 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 2 13:17:45.226765 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 2 13:17:45.227028 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 2 13:17:45.227049 kernel: PCI host bridge to bus 0000:00 Mar 2 13:17:45.227252 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 2 13:17:45.227480 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 2 13:17:45.227668 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 2 13:17:45.227917 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Mar 2 13:17:45.228114 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Mar 2 13:17:45.228299 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] Mar 2 13:17:45.228523 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 2 13:17:45.228878 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 2 13:17:45.229105 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Mar 2 13:17:45.229291 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Mar 2 13:17:45.229513 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Mar 2 13:17:45.229723 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Mar 2 13:17:45.229981 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Mar 2 13:17:45.230212 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 2 13:17:45.230474 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Mar 2 13:17:45.230681 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Mar 2 13:17:45.230943 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Mar 2 13:17:45.231147 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Mar 2 13:17:45.231397 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Mar 2 13:17:45.231604 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Mar 2 13:17:45.231858 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Mar 2 13:17:45.232067 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Mar 2 13:17:45.232281 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Mar 2 13:17:45.232526 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Mar 2 13:17:45.232730 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Mar 2 13:17:45.232985 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Mar 2 13:17:45.233188 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Mar 2 13:17:45.233441 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 2 13:17:45.233834 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 2 13:17:45.234088 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 2 13:17:45.234301 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Mar 2 13:17:45.234543 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Mar 2 13:17:45.234763 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 2 13:17:45.235021 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Mar 2 13:17:45.235041 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 2 13:17:45.235055 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 2 13:17:45.235067 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 2 13:17:45.235087 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 2 13:17:45.235099 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 2 13:17:45.235112 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 2 13:17:45.235124 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 2 13:17:45.235136 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 2 13:17:45.235149 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 2 13:17:45.235161 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 2 13:17:45.235174 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 2 13:17:45.235187 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 2 13:17:45.235204 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 2 13:17:45.235217 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 2 13:17:45.235229 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 2 13:17:45.235242 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 2 13:17:45.235255 kernel: iommu: Default domain type: Translated Mar 2 13:17:45.235267 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 2 13:17:45.235281 kernel: efivars: Registered efivars operations Mar 2 13:17:45.235294 kernel: PCI: Using ACPI for IRQ routing Mar 2 13:17:45.235307 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 2 13:17:45.235356 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Mar 2 13:17:45.235371 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Mar 2 13:17:45.235382 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Mar 2 13:17:45.235395 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Mar 2 13:17:45.235599 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 2 13:17:45.235854 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 2 13:17:45.236060 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 2 13:17:45.236080 kernel: vgaarb: loaded Mar 2 13:17:45.236092 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 2 13:17:45.236112 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 2 13:17:45.236124 kernel: clocksource: Switched to clocksource kvm-clock Mar 2 13:17:45.236138 kernel: VFS: Disk quotas dquot_6.6.0 Mar 2 13:17:45.236151 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 2 13:17:45.236163 kernel: pnp: PnP ACPI init Mar 2 13:17:45.236409 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Mar 2 13:17:45.236430 kernel: pnp: PnP ACPI: found 6 devices Mar 2 13:17:45.236442 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 2 13:17:45.236461 kernel: NET: Registered PF_INET protocol family Mar 2 13:17:45.236473 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 2 13:17:45.236486 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 2 13:17:45.236499 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 2 13:17:45.236511 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 2 13:17:45.236524 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 2 13:17:45.236536 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 2 13:17:45.236548 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 13:17:45.236562 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 13:17:45.236578 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 2 13:17:45.236591 kernel: NET: Registered PF_XDP protocol family Mar 2 13:17:45.236851 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Mar 2 13:17:45.237058 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Mar 2 13:17:45.237250 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 2 13:17:45.237481 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 2 13:17:45.237668 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 2 13:17:45.237916 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Mar 2 13:17:45.238114 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Mar 2 13:17:45.238299 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] Mar 2 13:17:45.238317 kernel: PCI: CLS 0 bytes, default 64 Mar 2 13:17:45.238367 kernel: Initialise system trusted keyrings Mar 2 13:17:45.238379 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 2 13:17:45.238392 kernel: Key type asymmetric registered Mar 2 13:17:45.238405 kernel: Asymmetric key parser 'x509' registered Mar 2 13:17:45.238417 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 2 13:17:45.238437 kernel: io scheduler mq-deadline registered Mar 2 13:17:45.238451 kernel: io scheduler kyber registered Mar 2 13:17:45.238464 kernel: io scheduler bfq registered Mar 2 13:17:45.238497 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 2 13:17:45.238512 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 2 13:17:45.238526 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 2 13:17:45.238539 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 2 13:17:45.238552 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 2 13:17:45.238566 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 2 13:17:45.238579 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 2 13:17:45.238598 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 2 13:17:45.238611 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 2 13:17:45.239246 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 2 13:17:45.239290 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 2 13:17:45.239608 kernel: rtc_cmos 00:04: registered as rtc0 Mar 2 13:17:45.240013 kernel: rtc_cmos 00:04: setting system clock to 2026-03-02T13:17:44 UTC (1772457464) Mar 2 13:17:45.240214 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Mar 2 13:17:45.240239 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 2 13:17:45.240254 kernel: efifb: probing for efifb Mar 2 13:17:45.240266 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Mar 2 13:17:45.240278 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Mar 2 13:17:45.240292 kernel: efifb: scrolling: redraw Mar 2 13:17:45.240303 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Mar 2 13:17:45.240316 kernel: Console: switching to colour frame buffer device 100x37 Mar 2 13:17:45.240363 kernel: fb0: EFI VGA frame buffer device Mar 2 13:17:45.240377 kernel: pstore: Using crash dump compression: deflate Mar 2 13:17:45.240396 kernel: pstore: Registered efi_pstore as persistent store backend Mar 2 13:17:45.240409 kernel: NET: Registered PF_INET6 protocol family Mar 2 13:17:45.240421 kernel: Segment Routing with IPv6 Mar 2 13:17:45.240434 kernel: In-situ OAM (IOAM) with IPv6 Mar 2 13:17:45.240445 kernel: NET: Registered PF_PACKET protocol family Mar 2 13:17:45.240458 kernel: Key type dns_resolver registered Mar 2 13:17:45.240471 kernel: IPI shorthand broadcast: enabled Mar 2 13:17:45.240512 kernel: sched_clock: Marking stable (1358027189, 859381983)->(3348059288, -1130650116) Mar 2 13:17:45.240530 kernel: registered taskstats version 1 Mar 2 13:17:45.240543 kernel: Loading compiled-in X.509 certificates Mar 2 13:17:45.240561 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: adc4961784537911a77ff0c4d6bd9b9639a51d45' Mar 2 13:17:45.240574 kernel: Key type .fscrypt registered Mar 2 13:17:45.240586 kernel: Key type fscrypt-provisioning registered Mar 2 13:17:45.240600 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 2 13:17:45.240612 kernel: ima: Allocated hash algorithm: sha1 Mar 2 13:17:45.240626 kernel: ima: No architecture policies found Mar 2 13:17:45.240639 kernel: clk: Disabling unused clocks Mar 2 13:17:45.240651 kernel: Freeing unused kernel image (initmem) memory: 42892K Mar 2 13:17:45.240669 kernel: Write protecting the kernel read-only data: 36864k Mar 2 13:17:45.240682 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Mar 2 13:17:45.240696 kernel: Run /init as init process Mar 2 13:17:45.240708 kernel: with arguments: Mar 2 13:17:45.240721 kernel: /init Mar 2 13:17:45.240735 kernel: with environment: Mar 2 13:17:45.240747 kernel: HOME=/ Mar 2 13:17:45.240760 kernel: TERM=linux Mar 2 13:17:45.240821 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 2 13:17:45.240845 systemd[1]: Detected virtualization kvm. Mar 2 13:17:45.240860 systemd[1]: Detected architecture x86-64. Mar 2 13:17:45.240873 systemd[1]: Running in initrd. Mar 2 13:17:45.240886 systemd[1]: No hostname configured, using default hostname. Mar 2 13:17:45.240900 systemd[1]: Hostname set to . Mar 2 13:17:45.240913 systemd[1]: Initializing machine ID from VM UUID. Mar 2 13:17:45.240927 systemd[1]: Queued start job for default target initrd.target. Mar 2 13:17:45.240948 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:17:45.240961 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:17:45.240976 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 2 13:17:45.240990 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 13:17:45.241004 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 2 13:17:45.241027 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 2 13:17:45.241043 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 2 13:17:45.241057 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 2 13:17:45.241070 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:17:45.241085 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:17:45.241098 systemd[1]: Reached target paths.target - Path Units. Mar 2 13:17:45.241112 systemd[1]: Reached target slices.target - Slice Units. Mar 2 13:17:45.241130 systemd[1]: Reached target swap.target - Swaps. Mar 2 13:17:45.241144 systemd[1]: Reached target timers.target - Timer Units. Mar 2 13:17:45.241158 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 13:17:45.241171 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 13:17:45.241186 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 13:17:45.241200 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 2 13:17:45.241214 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:17:45.241227 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 13:17:45.241246 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:17:45.241259 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 13:17:45.241274 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 2 13:17:45.241288 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 13:17:45.241302 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 2 13:17:45.241316 systemd[1]: Starting systemd-fsck-usr.service... Mar 2 13:17:45.241361 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 13:17:45.241375 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 13:17:45.241390 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:17:45.241442 systemd-journald[194]: Collecting audit messages is disabled. Mar 2 13:17:45.241475 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 2 13:17:45.241489 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:17:45.241503 systemd[1]: Finished systemd-fsck-usr.service. Mar 2 13:17:45.241524 systemd-journald[194]: Journal started Mar 2 13:17:45.241551 systemd-journald[194]: Runtime Journal (/run/log/journal/782142d9b18247ff89804a683de43a4e) is 6.0M, max 48.3M, 42.2M free. Mar 2 13:17:45.242275 systemd-modules-load[195]: Inserted module 'overlay' Mar 2 13:17:45.253001 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 13:17:45.253711 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:45.274424 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 13:17:45.287419 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 13:17:45.298257 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 13:17:45.308048 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:17:45.312599 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 2 13:17:45.321872 kernel: Bridge firewalling registered Mar 2 13:17:45.321921 systemd-modules-load[195]: Inserted module 'br_netfilter' Mar 2 13:17:45.325102 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 13:17:45.332007 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:17:45.340265 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:17:45.362427 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 2 13:17:45.367634 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:17:45.376527 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 13:17:45.388275 dracut-cmdline[221]: dracut-dracut-053 Mar 2 13:17:45.388275 dracut-cmdline[221]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=5b1ae8ad250cf3ddd00dc7c63ded260e5b82ee29f2cdc578a6ade4cab26e6a0b Mar 2 13:17:45.393624 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:17:45.409842 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:17:45.432150 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 13:17:45.481991 systemd-resolved[262]: Positive Trust Anchors: Mar 2 13:17:45.482035 systemd-resolved[262]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 13:17:45.482075 systemd-resolved[262]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 13:17:45.485424 systemd-resolved[262]: Defaulting to hostname 'linux'. Mar 2 13:17:45.487165 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 13:17:45.491613 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:17:45.567879 kernel: SCSI subsystem initialized Mar 2 13:17:45.578885 kernel: Loading iSCSI transport class v2.0-870. Mar 2 13:17:45.605893 kernel: iscsi: registered transport (tcp) Mar 2 13:17:45.646024 kernel: iscsi: registered transport (qla4xxx) Mar 2 13:17:45.646122 kernel: QLogic iSCSI HBA Driver Mar 2 13:17:45.715522 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 2 13:17:45.725999 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 2 13:17:45.777245 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 2 13:17:45.777368 kernel: device-mapper: uevent: version 1.0.3 Mar 2 13:17:45.779958 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 2 13:17:45.852878 kernel: raid6: avx2x4 gen() 28234 MB/s Mar 2 13:17:45.870856 kernel: raid6: avx2x2 gen() 23151 MB/s Mar 2 13:17:45.890181 kernel: raid6: avx2x1 gen() 19023 MB/s Mar 2 13:17:45.890229 kernel: raid6: using algorithm avx2x4 gen() 28234 MB/s Mar 2 13:17:45.925613 kernel: raid6: .... xor() 5041 MB/s, rmw enabled Mar 2 13:17:45.925712 kernel: raid6: using avx2x2 recovery algorithm Mar 2 13:17:45.960872 kernel: xor: automatically using best checksumming function avx Mar 2 13:17:46.265873 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 2 13:17:46.285244 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 2 13:17:46.298141 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:17:46.325502 systemd-udevd[415]: Using default interface naming scheme 'v255'. Mar 2 13:17:46.333046 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:17:46.348992 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 2 13:17:46.367158 dracut-pre-trigger[426]: rd.md=0: removing MD RAID activation Mar 2 13:17:46.419741 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 13:17:46.436447 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 13:17:46.536953 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:17:46.559976 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 2 13:17:46.587864 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 2 13:17:46.596828 kernel: cryptd: max_cpu_qlen set to 1000 Mar 2 13:17:46.606695 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 2 13:17:46.627971 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 2 13:17:46.628231 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 2 13:17:46.625246 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 13:17:46.646882 kernel: GPT:9289727 != 19775487 Mar 2 13:17:46.646906 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 2 13:17:46.646915 kernel: GPT:9289727 != 19775487 Mar 2 13:17:46.646932 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 2 13:17:46.646942 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 13:17:46.646962 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:17:46.651682 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 13:17:46.684829 kernel: AVX2 version of gcm_enc/dec engaged. Mar 2 13:17:46.684887 kernel: libata version 3.00 loaded. Mar 2 13:17:46.687223 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 2 13:17:46.710261 kernel: AES CTR mode by8 optimization enabled Mar 2 13:17:46.710293 kernel: ahci 0000:00:1f.2: version 3.0 Mar 2 13:17:46.712240 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 2 13:17:46.712456 kernel: BTRFS: device fsid a0930b2b-aeed-42a5-bf2f-ec141dfc71d3 devid 1 transid 35 /dev/vda3 scanned by (udev-worker) (475) Mar 2 13:17:46.712486 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (469) Mar 2 13:17:46.718250 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 2 13:17:46.729941 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 2 13:17:46.730189 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 2 13:17:46.730432 kernel: scsi host0: ahci Mar 2 13:17:46.730651 kernel: scsi host1: ahci Mar 2 13:17:46.736828 kernel: scsi host2: ahci Mar 2 13:17:46.739809 kernel: scsi host3: ahci Mar 2 13:17:46.741828 kernel: scsi host4: ahci Mar 2 13:17:46.747166 kernel: scsi host5: ahci Mar 2 13:17:46.747425 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Mar 2 13:17:46.747444 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Mar 2 13:17:46.752186 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Mar 2 13:17:46.755137 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 2 13:17:46.766923 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Mar 2 13:17:46.766949 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Mar 2 13:17:46.766966 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Mar 2 13:17:46.780373 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 2 13:17:46.807827 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 13:17:46.825101 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 2 13:17:46.835111 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 2 13:17:46.865131 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 2 13:17:46.874365 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 13:17:46.886823 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 13:17:46.886847 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 13:17:46.886857 disk-uuid[551]: Primary Header is updated. Mar 2 13:17:46.886857 disk-uuid[551]: Secondary Entries is updated. Mar 2 13:17:46.886857 disk-uuid[551]: Secondary Header is updated. Mar 2 13:17:46.874469 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:17:46.886923 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 13:17:46.914666 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 13:17:46.914855 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:46.929964 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:17:46.947991 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:17:46.974260 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:47.019089 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 13:17:47.040443 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:17:47.070848 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 2 13:17:47.075862 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 2 13:17:47.075915 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 2 13:17:47.078885 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 2 13:17:47.078911 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 2 13:17:47.083873 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 2 13:17:47.083909 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 2 13:17:47.086363 kernel: ata3.00: applying bridge limits Mar 2 13:17:47.088028 kernel: ata3.00: configured for UDMA/100 Mar 2 13:17:47.088828 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 2 13:17:47.155697 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 2 13:17:47.156111 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 2 13:17:47.170855 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 2 13:17:47.890881 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 2 13:17:47.891594 disk-uuid[552]: The operation has completed successfully. Mar 2 13:17:47.934470 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 2 13:17:47.934624 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 2 13:17:47.953005 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 2 13:17:47.960730 sh[594]: Success Mar 2 13:17:47.980863 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Mar 2 13:17:48.027220 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 2 13:17:48.041756 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 2 13:17:48.045760 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 2 13:17:48.069136 kernel: BTRFS info (device dm-0): first mount of filesystem a0930b2b-aeed-42a5-bf2f-ec141dfc71d3 Mar 2 13:17:48.069191 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:17:48.069219 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 2 13:17:48.072761 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 2 13:17:48.075212 kernel: BTRFS info (device dm-0): using free space tree Mar 2 13:17:48.087435 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 2 13:17:48.090607 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 2 13:17:48.099086 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 2 13:17:48.106998 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 2 13:17:48.122277 kernel: BTRFS info (device vda6): first mount of filesystem 59abb777-1ea9-43fd-8326-9ccf988e79fa Mar 2 13:17:48.122324 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:17:48.122378 kernel: BTRFS info (device vda6): using free space tree Mar 2 13:17:48.127867 kernel: BTRFS info (device vda6): auto enabling async discard Mar 2 13:17:48.139107 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 2 13:17:48.144044 kernel: BTRFS info (device vda6): last unmount of filesystem 59abb777-1ea9-43fd-8326-9ccf988e79fa Mar 2 13:17:48.157183 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 2 13:17:48.173004 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 2 13:17:48.261218 ignition[676]: Ignition 2.19.0 Mar 2 13:17:48.261253 ignition[676]: Stage: fetch-offline Mar 2 13:17:48.261312 ignition[676]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:48.261327 ignition[676]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:48.261486 ignition[676]: parsed url from cmdline: "" Mar 2 13:17:48.261494 ignition[676]: no config URL provided Mar 2 13:17:48.261504 ignition[676]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 13:17:48.261517 ignition[676]: no config at "/usr/lib/ignition/user.ign" Mar 2 13:17:48.261557 ignition[676]: op(1): [started] loading QEMU firmware config module Mar 2 13:17:48.261565 ignition[676]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 2 13:17:48.270826 ignition[676]: op(1): [finished] loading QEMU firmware config module Mar 2 13:17:48.331099 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 13:17:48.355870 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 13:17:48.406098 systemd-networkd[782]: lo: Link UP Mar 2 13:17:48.406137 systemd-networkd[782]: lo: Gained carrier Mar 2 13:17:48.408556 systemd-networkd[782]: Enumeration completed Mar 2 13:17:48.409843 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 13:17:48.409850 systemd-networkd[782]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 13:17:48.410972 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 13:17:48.416737 systemd[1]: Reached target network.target - Network. Mar 2 13:17:48.417201 systemd-networkd[782]: eth0: Link UP Mar 2 13:17:48.417207 systemd-networkd[782]: eth0: Gained carrier Mar 2 13:17:48.417219 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 13:17:48.459969 systemd-networkd[782]: eth0: DHCPv4 address 10.0.0.160/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 13:17:48.557280 ignition[676]: parsing config with SHA512: 9096524952308294dd4e2f6be2b3ddb7f7f900838454d61d3f7d7e946bc5dd4e26faa0955b4d8223d818bd922c792bb5ba026c307811b8ba995676d3a27121ba Mar 2 13:17:48.566249 unknown[676]: fetched base config from "system" Mar 2 13:17:48.566267 unknown[676]: fetched user config from "qemu" Mar 2 13:17:48.567094 ignition[676]: fetch-offline: fetch-offline passed Mar 2 13:17:48.567187 ignition[676]: Ignition finished successfully Mar 2 13:17:48.582301 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 13:17:48.592535 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 2 13:17:48.616160 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 2 13:17:48.665572 ignition[787]: Ignition 2.19.0 Mar 2 13:17:48.665608 ignition[787]: Stage: kargs Mar 2 13:17:48.665883 ignition[787]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:48.665899 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:48.667114 ignition[787]: kargs: kargs passed Mar 2 13:17:48.667175 ignition[787]: Ignition finished successfully Mar 2 13:17:48.687941 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 2 13:17:48.706193 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 2 13:17:48.775730 ignition[795]: Ignition 2.19.0 Mar 2 13:17:48.775765 ignition[795]: Stage: disks Mar 2 13:17:48.776076 ignition[795]: no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:48.781145 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 2 13:17:48.776095 ignition[795]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:48.785740 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 2 13:17:48.777428 ignition[795]: disks: disks passed Mar 2 13:17:48.790646 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 13:17:48.777495 ignition[795]: Ignition finished successfully Mar 2 13:17:48.798843 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 13:17:48.802094 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 13:17:48.805142 systemd[1]: Reached target basic.target - Basic System. Mar 2 13:17:48.824098 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 2 13:17:48.869229 systemd-fsck[805]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 2 13:17:48.876765 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 2 13:17:48.929535 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 2 13:17:49.155326 kernel: EXT4-fs (vda9): mounted filesystem 84e86976-7918-44d3-a6f5-d0f90ce6c152 r/w with ordered data mode. Quota mode: none. Mar 2 13:17:49.159144 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 2 13:17:49.164229 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 2 13:17:49.188021 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 13:17:49.199758 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 2 13:17:49.203946 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 2 13:17:49.246992 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (813) Mar 2 13:17:49.247029 kernel: BTRFS info (device vda6): first mount of filesystem 59abb777-1ea9-43fd-8326-9ccf988e79fa Mar 2 13:17:49.247078 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:17:49.204014 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 2 13:17:49.269723 kernel: BTRFS info (device vda6): using free space tree Mar 2 13:17:49.269852 kernel: BTRFS info (device vda6): auto enabling async discard Mar 2 13:17:49.204061 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 13:17:49.240187 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 2 13:17:49.279133 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 2 13:17:49.288069 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 13:17:49.367070 initrd-setup-root[838]: cut: /sysroot/etc/passwd: No such file or directory Mar 2 13:17:49.373413 initrd-setup-root[845]: cut: /sysroot/etc/group: No such file or directory Mar 2 13:17:49.378081 initrd-setup-root[852]: cut: /sysroot/etc/shadow: No such file or directory Mar 2 13:17:49.386910 initrd-setup-root[859]: cut: /sysroot/etc/gshadow: No such file or directory Mar 2 13:17:49.544508 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 2 13:17:49.553000 systemd-networkd[782]: eth0: Gained IPv6LL Mar 2 13:17:49.557171 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 2 13:17:49.558303 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 2 13:17:49.579810 kernel: BTRFS info (device vda6): last unmount of filesystem 59abb777-1ea9-43fd-8326-9ccf988e79fa Mar 2 13:17:49.568228 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 2 13:17:49.622722 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 2 13:17:49.635896 ignition[926]: INFO : Ignition 2.19.0 Mar 2 13:17:49.635896 ignition[926]: INFO : Stage: mount Mar 2 13:17:49.641230 ignition[926]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:49.641230 ignition[926]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:49.641230 ignition[926]: INFO : mount: mount passed Mar 2 13:17:49.641230 ignition[926]: INFO : Ignition finished successfully Mar 2 13:17:49.639743 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 2 13:17:49.660198 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 2 13:17:50.173278 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 13:17:50.192922 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (941) Mar 2 13:17:50.201280 kernel: BTRFS info (device vda6): first mount of filesystem 59abb777-1ea9-43fd-8326-9ccf988e79fa Mar 2 13:17:50.201371 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 2 13:17:50.201394 kernel: BTRFS info (device vda6): using free space tree Mar 2 13:17:50.214886 kernel: BTRFS info (device vda6): auto enabling async discard Mar 2 13:17:50.219874 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 13:17:50.338980 ignition[958]: INFO : Ignition 2.19.0 Mar 2 13:17:50.338980 ignition[958]: INFO : Stage: files Mar 2 13:17:50.346524 ignition[958]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:50.346524 ignition[958]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:50.346524 ignition[958]: DEBUG : files: compiled without relabeling support, skipping Mar 2 13:17:50.346524 ignition[958]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 2 13:17:50.346524 ignition[958]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 2 13:17:50.372559 ignition[958]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 2 13:17:50.372559 ignition[958]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 2 13:17:50.372559 ignition[958]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 2 13:17:50.372559 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 2 13:17:50.372559 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 2 13:17:50.372559 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 13:17:50.372559 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Mar 2 13:17:50.349188 unknown[958]: wrote ssh authorized keys file for user: core Mar 2 13:17:50.444695 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 2 13:17:50.599960 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Mar 2 13:17:50.599960 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 2 13:17:50.615057 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Mar 2 13:17:50.743871 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Mar 2 13:17:50.860331 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 2 13:17:50.865843 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-x86-64.raw: attempt #1 Mar 2 13:17:51.133880 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Mar 2 13:17:51.483938 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-x86-64.raw" Mar 2 13:17:51.483938 ignition[958]: INFO : files: op(d): [started] processing unit "containerd.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(d): [finished] processing unit "containerd.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(11): [started] processing unit "coreos-metadata.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(11): op(12): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(11): op(12): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(11): [finished] processing unit "coreos-metadata.service" Mar 2 13:17:51.496533 ignition[958]: INFO : files: op(13): [started] setting preset to disabled for "coreos-metadata.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: op(13): op(14): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: op(13): op(14): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: op(13): [finished] setting preset to disabled for "coreos-metadata.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: op(15): [started] setting preset to enabled for "prepare-helm.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: op(15): [finished] setting preset to enabled for "prepare-helm.service" Mar 2 13:17:51.580719 ignition[958]: INFO : files: createResultFile: createFiles: op(16): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 2 13:17:51.580719 ignition[958]: INFO : files: createResultFile: createFiles: op(16): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 2 13:17:51.580719 ignition[958]: INFO : files: files passed Mar 2 13:17:51.580719 ignition[958]: INFO : Ignition finished successfully Mar 2 13:17:51.540310 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 2 13:17:51.583226 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 2 13:17:51.591976 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 2 13:17:51.665136 initrd-setup-root-after-ignition[984]: grep: /sysroot/oem/oem-release: No such file or directory Mar 2 13:17:51.616520 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 13:17:51.675505 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:17:51.675505 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:17:51.622833 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 2 13:17:51.700492 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 13:17:51.647026 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 2 13:17:51.676177 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 2 13:17:51.676378 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 2 13:17:51.686254 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 2 13:17:51.696168 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 2 13:17:51.705140 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 2 13:17:51.705314 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 2 13:17:51.710597 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 2 13:17:51.738107 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 2 13:17:51.852893 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 13:17:51.872024 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 2 13:17:51.884544 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:17:51.889066 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:17:51.897737 systemd[1]: Stopped target timers.target - Timer Units. Mar 2 13:17:51.905397 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 2 13:17:51.905517 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 13:17:51.913521 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 2 13:17:51.921046 systemd[1]: Stopped target basic.target - Basic System. Mar 2 13:17:51.928924 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 2 13:17:51.936732 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 13:17:51.936950 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 2 13:17:51.937613 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 2 13:17:51.939561 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 13:17:51.940905 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 2 13:17:51.942729 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 2 13:17:51.944665 systemd[1]: Stopped target swap.target - Swaps. Mar 2 13:17:51.946504 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 2 13:17:51.946617 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 2 13:17:51.948448 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:17:51.949725 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:17:51.951643 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 2 13:17:51.951948 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:17:51.952275 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 2 13:17:51.952439 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 2 13:17:51.954901 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 2 13:17:51.955074 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 13:17:51.955559 systemd[1]: Stopped target paths.target - Path Units. Mar 2 13:17:51.956748 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 2 13:17:51.960934 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:17:52.135189 ignition[1012]: INFO : Ignition 2.19.0 Mar 2 13:17:52.135189 ignition[1012]: INFO : Stage: umount Mar 2 13:17:52.135189 ignition[1012]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 13:17:52.135189 ignition[1012]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 2 13:17:52.135189 ignition[1012]: INFO : umount: umount passed Mar 2 13:17:52.135189 ignition[1012]: INFO : Ignition finished successfully Mar 2 13:17:51.961398 systemd[1]: Stopped target slices.target - Slice Units. Mar 2 13:17:51.962639 systemd[1]: Stopped target sockets.target - Socket Units. Mar 2 13:17:51.964558 systemd[1]: iscsid.socket: Deactivated successfully. Mar 2 13:17:51.964671 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 13:17:51.965899 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 2 13:17:51.966003 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 13:17:51.966532 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 2 13:17:51.966659 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 13:17:51.967819 systemd[1]: ignition-files.service: Deactivated successfully. Mar 2 13:17:51.967942 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 2 13:17:52.021656 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 2 13:17:52.027576 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 2 13:17:52.034045 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 2 13:17:52.034289 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:17:52.042721 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 2 13:17:52.042936 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 13:17:52.055561 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 2 13:17:52.057743 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 2 13:17:52.066974 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 2 13:17:52.130142 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 2 13:17:52.130329 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 2 13:17:52.135418 systemd[1]: Stopped target network.target - Network. Mar 2 13:17:52.142151 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 2 13:17:52.142227 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 2 13:17:52.143642 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 2 13:17:52.143705 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 2 13:17:52.146710 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 2 13:17:52.146861 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 2 13:17:52.147184 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 2 13:17:52.147231 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 2 13:17:52.148652 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 2 13:17:52.149884 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 2 13:17:52.171907 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 2 13:17:52.172041 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 2 13:17:52.176921 systemd-networkd[782]: eth0: DHCPv6 lease lost Mar 2 13:17:52.179905 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 2 13:17:52.179989 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 2 13:17:52.186710 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 2 13:17:52.186912 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 2 13:17:52.194391 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 2 13:17:52.194534 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 2 13:17:52.203980 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 2 13:17:52.460320 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Mar 2 13:17:52.204036 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:17:52.222023 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 2 13:17:52.227610 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 2 13:17:52.227686 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 13:17:52.227898 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 13:17:52.227947 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:17:52.229588 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 2 13:17:52.229634 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 2 13:17:52.231509 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 2 13:17:52.231555 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:17:52.232901 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:17:52.253632 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 2 13:17:52.253817 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 2 13:17:52.263565 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 2 13:17:52.263863 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:17:52.271845 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 2 13:17:52.271931 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 2 13:17:52.278470 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 2 13:17:52.278527 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:17:52.286404 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 2 13:17:52.286482 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 2 13:17:52.294474 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 2 13:17:52.294546 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 2 13:17:52.302570 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 13:17:52.302679 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 13:17:52.328045 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 2 13:17:52.335428 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 2 13:17:52.335506 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:17:52.343736 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 2 13:17:52.343859 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:17:52.352157 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 2 13:17:52.352224 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:17:52.356953 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 13:17:52.357017 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:52.365130 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 2 13:17:52.365290 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 2 13:17:52.373451 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 2 13:17:52.400109 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 2 13:17:52.414598 systemd[1]: Switching root. Mar 2 13:17:52.606561 systemd-journald[194]: Journal stopped Mar 2 13:17:53.981754 kernel: SELinux: policy capability network_peer_controls=1 Mar 2 13:17:53.981866 kernel: SELinux: policy capability open_perms=1 Mar 2 13:17:53.981879 kernel: SELinux: policy capability extended_socket_class=1 Mar 2 13:17:53.981890 kernel: SELinux: policy capability always_check_network=0 Mar 2 13:17:53.981907 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 2 13:17:53.981916 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 2 13:17:53.981927 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 2 13:17:53.981937 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 2 13:17:53.981953 kernel: audit: type=1403 audit(1772457472.737:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 2 13:17:53.981972 systemd[1]: Successfully loaded SELinux policy in 59.334ms. Mar 2 13:17:53.981999 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 19.989ms. Mar 2 13:17:53.982011 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 2 13:17:53.982022 systemd[1]: Detected virtualization kvm. Mar 2 13:17:53.982033 systemd[1]: Detected architecture x86-64. Mar 2 13:17:53.982046 systemd[1]: Detected first boot. Mar 2 13:17:53.982056 systemd[1]: Initializing machine ID from VM UUID. Mar 2 13:17:53.982067 zram_generator::config[1072]: No configuration found. Mar 2 13:17:53.982081 systemd[1]: Populated /etc with preset unit settings. Mar 2 13:17:53.982092 systemd[1]: Queued start job for default target multi-user.target. Mar 2 13:17:53.982104 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 2 13:17:53.982115 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 2 13:17:53.982126 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 2 13:17:53.982137 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 2 13:17:53.982148 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 2 13:17:53.982159 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 2 13:17:53.982171 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 2 13:17:53.982184 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 2 13:17:53.982195 systemd[1]: Created slice user.slice - User and Session Slice. Mar 2 13:17:53.982206 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 13:17:53.982216 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 13:17:53.982227 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 2 13:17:53.982238 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 2 13:17:53.982249 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 2 13:17:53.982260 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 13:17:53.982274 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 2 13:17:53.982285 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 13:17:53.982296 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 2 13:17:53.982306 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 13:17:53.982317 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 13:17:53.982328 systemd[1]: Reached target slices.target - Slice Units. Mar 2 13:17:53.982338 systemd[1]: Reached target swap.target - Swaps. Mar 2 13:17:53.982349 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 2 13:17:53.982394 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 2 13:17:53.982405 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 13:17:53.982416 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 2 13:17:53.982449 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 13:17:53.982461 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 13:17:53.982471 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 13:17:53.982482 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 2 13:17:53.982493 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 2 13:17:53.982503 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 2 13:17:53.982514 systemd[1]: Mounting media.mount - External Media Directory... Mar 2 13:17:53.982529 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:17:53.982540 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 2 13:17:53.982551 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 2 13:17:53.982561 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 2 13:17:53.982573 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 2 13:17:53.982584 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:17:53.982594 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 13:17:53.982606 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 2 13:17:53.982619 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:17:53.982630 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 13:17:53.982640 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:17:53.982651 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 2 13:17:53.982662 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:17:53.982673 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 2 13:17:53.982684 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Mar 2 13:17:53.982695 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Mar 2 13:17:53.982708 kernel: fuse: init (API version 7.39) Mar 2 13:17:53.982719 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 13:17:53.982729 kernel: ACPI: bus type drm_connector registered Mar 2 13:17:53.982739 kernel: loop: module loaded Mar 2 13:17:53.982750 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 13:17:53.982761 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 13:17:53.982807 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 2 13:17:53.982841 systemd-journald[1171]: Collecting audit messages is disabled. Mar 2 13:17:53.982872 systemd-journald[1171]: Journal started Mar 2 13:17:53.982902 systemd-journald[1171]: Runtime Journal (/run/log/journal/782142d9b18247ff89804a683de43a4e) is 6.0M, max 48.3M, 42.2M free. Mar 2 13:17:53.991011 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 13:17:53.991074 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:17:54.003858 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 13:17:54.009215 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 2 13:17:54.013228 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 2 13:17:54.017294 systemd[1]: Mounted media.mount - External Media Directory. Mar 2 13:17:54.020958 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 2 13:17:54.024616 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 2 13:17:54.028512 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 2 13:17:54.032554 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 2 13:17:54.037684 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 13:17:54.042693 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 2 13:17:54.043151 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 2 13:17:54.047327 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:17:54.047673 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:17:54.051396 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 13:17:54.051706 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 13:17:54.055455 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:17:54.055762 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:17:54.064588 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 2 13:17:54.064986 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 2 13:17:54.068445 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:17:54.069100 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:17:54.072941 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 13:17:54.077012 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 13:17:54.081517 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 2 13:17:54.099638 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 13:17:54.117007 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 2 13:17:54.123016 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 2 13:17:54.127067 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 2 13:17:54.129533 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 2 13:17:54.136952 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 2 13:17:54.141286 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 13:17:54.149213 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 2 13:17:54.153447 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 13:17:54.156066 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:17:54.162288 systemd-journald[1171]: Time spent on flushing to /var/log/journal/782142d9b18247ff89804a683de43a4e is 20.759ms for 974 entries. Mar 2 13:17:54.162288 systemd-journald[1171]: System Journal (/var/log/journal/782142d9b18247ff89804a683de43a4e) is 8.0M, max 195.6M, 187.6M free. Mar 2 13:17:54.209045 systemd-journald[1171]: Received client request to flush runtime journal. Mar 2 13:17:54.166004 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 13:17:54.175288 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 13:17:54.180580 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 2 13:17:54.185458 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 2 13:17:54.191133 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 2 13:17:54.203722 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 2 13:17:54.220045 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 2 13:17:54.225058 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 2 13:17:54.230133 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:17:54.236687 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Mar 2 13:17:54.237065 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Mar 2 13:17:54.238969 udevadm[1222]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 2 13:17:54.245859 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 13:17:54.266152 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 2 13:17:54.300867 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 2 13:17:54.317017 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 13:17:54.341237 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Mar 2 13:17:54.341272 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Mar 2 13:17:54.347275 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 13:17:54.767545 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 2 13:17:54.785179 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 13:17:54.822261 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Mar 2 13:17:54.844545 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 13:17:54.862968 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 13:17:54.874939 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 2 13:17:54.889423 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Mar 2 13:17:54.910849 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1249) Mar 2 13:17:54.970029 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 2 13:17:54.981657 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 2 13:17:54.984946 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Mar 2 13:17:55.000879 kernel: ACPI: button: Power Button [PWRF] Mar 2 13:17:55.027163 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 2 13:17:55.027243 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Mar 2 13:17:55.027596 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 2 13:17:55.034432 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 2 13:17:55.040958 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 2 13:17:55.089999 systemd-networkd[1252]: lo: Link UP Mar 2 13:17:55.090031 systemd-networkd[1252]: lo: Gained carrier Mar 2 13:17:55.094181 systemd-networkd[1252]: Enumeration completed Mar 2 13:17:55.094424 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 13:17:55.095203 systemd-networkd[1252]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 13:17:55.095209 systemd-networkd[1252]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 13:17:55.096617 systemd-networkd[1252]: eth0: Link UP Mar 2 13:17:55.096622 systemd-networkd[1252]: eth0: Gained carrier Mar 2 13:17:55.096636 systemd-networkd[1252]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 13:17:55.109703 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 2 13:17:55.131985 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:17:55.146042 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 13:17:55.147130 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:55.156877 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 13:17:55.184836 kernel: mousedev: PS/2 mouse device common for all mice Mar 2 13:17:55.186428 systemd-networkd[1252]: eth0: DHCPv4 address 10.0.0.160/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 2 13:17:55.204942 kernel: kvm_amd: TSC scaling supported Mar 2 13:17:55.205033 kernel: kvm_amd: Nested Virtualization enabled Mar 2 13:17:55.205062 kernel: kvm_amd: Nested Paging enabled Mar 2 13:17:55.205090 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 2 13:17:55.208866 kernel: kvm_amd: PMU virtualization is disabled Mar 2 13:17:55.267866 kernel: EDAC MC: Ver: 3.0.0 Mar 2 13:17:55.280707 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 13:17:55.305492 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 2 13:17:55.335140 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 2 13:17:55.345654 lvm[1289]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 2 13:17:56.849118 systemd-networkd[1252]: eth0: Gained IPv6LL Mar 2 13:17:57.237997 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 2 13:17:57.247186 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 2 13:17:57.252244 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 13:17:57.268985 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 2 13:17:57.278437 lvm[1295]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 2 13:17:57.330594 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 2 13:17:57.338059 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 13:17:57.343265 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 2 13:17:57.343345 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 13:17:57.347299 systemd[1]: Reached target machines.target - Containers. Mar 2 13:17:57.352570 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 2 13:17:57.369126 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 2 13:17:57.379967 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 2 13:17:57.384878 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:17:57.386618 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 2 13:17:57.393483 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 2 13:17:57.400925 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 2 13:17:57.407868 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 2 13:17:57.496979 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 2 13:17:57.505337 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 2 13:17:57.507506 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 2 13:17:57.533979 kernel: loop0: detected capacity change from 0 to 142488 Mar 2 13:17:57.572449 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 2 13:17:57.604886 kernel: loop1: detected capacity change from 0 to 140768 Mar 2 13:17:57.669813 kernel: loop2: detected capacity change from 0 to 228704 Mar 2 13:17:57.726830 kernel: loop3: detected capacity change from 0 to 142488 Mar 2 13:17:57.958218 kernel: loop4: detected capacity change from 0 to 140768 Mar 2 13:17:58.041353 kernel: loop5: detected capacity change from 0 to 228704 Mar 2 13:17:58.079126 (sd-merge)[1316]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 2 13:17:58.081585 (sd-merge)[1316]: Merged extensions into '/usr'. Mar 2 13:17:58.241230 systemd[1]: Reloading requested from client PID 1304 ('systemd-sysext') (unit systemd-sysext.service)... Mar 2 13:17:58.241316 systemd[1]: Reloading... Mar 2 13:17:58.439982 zram_generator::config[1349]: No configuration found. Mar 2 13:17:59.058646 ldconfig[1301]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 2 13:17:59.072870 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 2 13:17:59.221336 systemd[1]: Reloading finished in 979 ms. Mar 2 13:17:59.248863 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 2 13:17:59.254170 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 2 13:17:59.277932 systemd[1]: Starting ensure-sysext.service... Mar 2 13:17:59.286326 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 13:17:59.303093 systemd[1]: Reloading requested from client PID 1387 ('systemctl') (unit ensure-sysext.service)... Mar 2 13:17:59.303134 systemd[1]: Reloading... Mar 2 13:18:00.024884 zram_generator::config[1416]: No configuration found. Mar 2 13:18:00.025849 systemd-tmpfiles[1388]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 13:18:00.026359 systemd-tmpfiles[1388]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 2 13:18:00.028521 systemd-tmpfiles[1388]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 2 13:18:00.029086 systemd-tmpfiles[1388]: ACLs are not supported, ignoring. Mar 2 13:18:00.030081 systemd-tmpfiles[1388]: ACLs are not supported, ignoring. Mar 2 13:18:00.036873 systemd-tmpfiles[1388]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 13:18:00.036977 systemd-tmpfiles[1388]: Skipping /boot Mar 2 13:18:00.057586 systemd-tmpfiles[1388]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 13:18:00.057628 systemd-tmpfiles[1388]: Skipping /boot Mar 2 13:18:00.195521 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 2 13:18:00.311732 systemd[1]: Reloading finished in 1008 ms. Mar 2 13:18:00.394254 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 13:18:00.422166 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 2 13:18:00.435343 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 2 13:18:00.447067 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 2 13:18:00.474414 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 13:18:00.509295 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 2 13:18:00.546323 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.546726 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:18:00.550629 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:18:00.566090 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:18:00.575077 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:18:00.594872 augenrules[1488]: No rules Mar 2 13:18:00.579448 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:18:00.579677 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.582508 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 2 13:18:00.589276 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:18:00.589573 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:18:00.600478 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 2 13:18:00.606181 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:18:00.606502 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:18:00.618610 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 2 13:18:00.623536 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:18:00.623951 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:18:00.642553 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.642916 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:18:00.650110 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:18:00.655723 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:18:00.661654 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:18:00.665506 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:18:00.670067 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 2 13:18:00.673910 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 13:18:00.674017 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.676768 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 2 13:18:00.682213 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:18:00.682335 systemd-resolved[1469]: Positive Trust Anchors: Mar 2 13:18:00.682344 systemd-resolved[1469]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 13:18:00.682407 systemd-resolved[1469]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 13:18:00.682540 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:18:00.687686 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:18:00.687693 systemd-resolved[1469]: Defaulting to hostname 'linux'. Mar 2 13:18:00.688005 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:18:00.693508 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 13:18:00.698322 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:18:00.698682 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:18:00.708865 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 2 13:18:00.723161 systemd[1]: Reached target network.target - Network. Mar 2 13:18:00.726710 systemd[1]: Reached target network-online.target - Network is Online. Mar 2 13:18:00.731013 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 13:18:00.735584 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.735969 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 13:18:00.753201 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 13:18:00.758628 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 13:18:00.763845 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 13:18:00.769483 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 13:18:00.773302 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 13:18:00.773538 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 13:18:00.773676 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 2 13:18:00.775281 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 13:18:00.775544 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 13:18:00.781070 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 13:18:00.781444 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 13:18:00.792233 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 13:18:00.792572 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 13:18:00.799544 systemd[1]: Finished ensure-sysext.service. Mar 2 13:18:00.803282 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 13:18:00.803762 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 13:18:00.815098 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 13:18:00.815263 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 13:18:00.830055 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 2 13:18:01.054590 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 2 13:18:01.059652 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 13:18:01.063946 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 2 13:18:01.577576 systemd-resolved[1469]: Clock change detected. Flushing caches. Mar 2 13:18:01.582109 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 2 13:18:01.582251 systemd-timesyncd[1534]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 2 13:18:01.582306 systemd-timesyncd[1534]: Initial clock synchronization to Mon 2026-03-02 13:18:01.577357 UTC. Mar 2 13:18:01.587288 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 2 13:18:01.592153 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 2 13:18:01.592289 systemd[1]: Reached target paths.target - Path Units. Mar 2 13:18:01.595715 systemd[1]: Reached target time-set.target - System Time Set. Mar 2 13:18:01.599819 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 2 13:18:01.604001 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 2 13:18:01.608758 systemd[1]: Reached target timers.target - Timer Units. Mar 2 13:18:01.613886 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 2 13:18:01.620558 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 2 13:18:01.627696 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 2 13:18:01.638344 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 2 13:18:01.642422 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 13:18:01.646304 systemd[1]: Reached target basic.target - Basic System. Mar 2 13:18:01.650255 systemd[1]: System is tainted: cgroupsv1 Mar 2 13:18:01.650328 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 2 13:18:01.650364 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 2 13:18:01.653612 systemd[1]: Starting containerd.service - containerd container runtime... Mar 2 13:18:01.659455 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 2 13:18:01.667963 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 2 13:18:01.674400 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 2 13:18:01.680149 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 2 13:18:01.684595 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 2 13:18:01.692509 jq[1542]: false Mar 2 13:18:01.695316 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:01.701373 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 2 13:18:01.705279 extend-filesystems[1543]: Found loop3 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found loop4 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found loop5 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found sr0 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda1 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda2 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda3 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found usr Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda4 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda6 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda7 Mar 2 13:18:01.705279 extend-filesystems[1543]: Found vda9 Mar 2 13:18:01.705279 extend-filesystems[1543]: Checking size of /dev/vda9 Mar 2 13:18:01.721609 dbus-daemon[1540]: [system] SELinux support is enabled Mar 2 13:18:01.732394 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 2 13:18:01.758056 extend-filesystems[1543]: Resized partition /dev/vda9 Mar 2 13:18:01.754439 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 2 13:18:01.763822 extend-filesystems[1561]: resize2fs 1.47.1 (20-May-2024) Mar 2 13:18:01.792304 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 2 13:18:01.792330 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1565) Mar 2 13:18:01.764461 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 2 13:18:01.787407 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 2 13:18:01.810515 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 2 13:18:01.817235 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 2 13:18:01.820049 systemd[1]: Starting update-engine.service - Update Engine... Mar 2 13:18:01.834898 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 2 13:18:01.859955 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 2 13:18:01.860043 update_engine[1585]: I20260302 13:18:01.855392 1585 main.cc:92] Flatcar Update Engine starting Mar 2 13:18:01.860043 update_engine[1585]: I20260302 13:18:01.858336 1585 update_check_scheduler.cc:74] Next update check in 4m17s Mar 2 13:18:01.842036 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 2 13:18:01.860576 jq[1586]: true Mar 2 13:18:01.858907 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 2 13:18:01.859415 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 2 13:18:01.865042 extend-filesystems[1561]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 2 13:18:01.865042 extend-filesystems[1561]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 2 13:18:01.865042 extend-filesystems[1561]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 2 13:18:01.863862 systemd[1]: motdgen.service: Deactivated successfully. Mar 2 13:18:01.888639 extend-filesystems[1543]: Resized filesystem in /dev/vda9 Mar 2 13:18:01.873969 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 2 13:18:01.892453 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 2 13:18:01.892895 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 2 13:18:01.898225 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 2 13:18:01.904679 systemd-logind[1577]: Watching system buttons on /dev/input/event1 (Power Button) Mar 2 13:18:01.904711 systemd-logind[1577]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 2 13:18:01.906564 systemd-logind[1577]: New seat seat0. Mar 2 13:18:01.908752 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 2 13:18:01.909151 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 2 13:18:01.914907 systemd[1]: Started systemd-logind.service - User Login Management. Mar 2 13:18:01.943348 jq[1603]: true Mar 2 13:18:01.953036 (ntainerd)[1604]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 2 13:18:01.960132 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 2 13:18:01.960597 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 2 13:18:01.975348 dbus-daemon[1540]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 2 13:18:01.979669 tar[1595]: linux-amd64/LICENSE Mar 2 13:18:01.979669 tar[1595]: linux-amd64/helm Mar 2 13:18:01.986540 systemd[1]: Started update-engine.service - Update Engine. Mar 2 13:18:01.998877 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 2 13:18:01.999042 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 2 13:18:01.999266 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 2 13:18:02.004304 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 2 13:18:02.004540 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 2 13:18:02.011349 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 2 13:18:02.900603 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 2 13:18:02.961122 sshd_keygen[1583]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 2 13:18:02.966006 bash[1637]: Updated "/home/core/.ssh/authorized_keys" Mar 2 13:18:02.967662 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 2 13:18:02.976549 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 2 13:18:02.995404 locksmithd[1638]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 2 13:18:03.020118 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 2 13:18:03.113537 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 2 13:18:03.144073 systemd[1]: issuegen.service: Deactivated successfully. Mar 2 13:18:03.145269 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 2 13:18:03.165734 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 2 13:18:03.454297 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 2 13:18:03.472723 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 2 13:18:03.479621 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 2 13:18:03.483825 systemd[1]: Reached target getty.target - Login Prompts. Mar 2 13:18:04.442796 containerd[1604]: time="2026-03-02T13:18:04.442107105Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 2 13:18:04.744981 containerd[1604]: time="2026-03-02T13:18:04.743950441Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.751330 containerd[1604]: time="2026-03-02T13:18:04.751156638Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 2 13:18:04.751330 containerd[1604]: time="2026-03-02T13:18:04.751248059Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 2 13:18:04.751330 containerd[1604]: time="2026-03-02T13:18:04.751313331Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 2 13:18:04.752069 containerd[1604]: time="2026-03-02T13:18:04.751608843Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 2 13:18:04.752069 containerd[1604]: time="2026-03-02T13:18:04.751664997Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.752069 containerd[1604]: time="2026-03-02T13:18:04.751830246Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 2 13:18:04.752069 containerd[1604]: time="2026-03-02T13:18:04.751849382Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.753327 containerd[1604]: time="2026-03-02T13:18:04.752347222Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 2 13:18:04.753327 containerd[1604]: time="2026-03-02T13:18:04.752375875Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.753327 containerd[1604]: time="2026-03-02T13:18:04.752399228Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 2 13:18:04.753327 containerd[1604]: time="2026-03-02T13:18:04.752410069Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.753327 containerd[1604]: time="2026-03-02T13:18:04.752652421Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.755111 containerd[1604]: time="2026-03-02T13:18:04.754938349Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 2 13:18:04.755841 containerd[1604]: time="2026-03-02T13:18:04.755409549Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 2 13:18:04.755841 containerd[1604]: time="2026-03-02T13:18:04.755441549Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 2 13:18:04.755841 containerd[1604]: time="2026-03-02T13:18:04.755640701Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 2 13:18:04.755841 containerd[1604]: time="2026-03-02T13:18:04.755794107Z" level=info msg="metadata content store policy set" policy=shared Mar 2 13:18:04.766862 containerd[1604]: time="2026-03-02T13:18:04.766758185Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 2 13:18:04.767241 containerd[1604]: time="2026-03-02T13:18:04.767131883Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 2 13:18:04.767282 containerd[1604]: time="2026-03-02T13:18:04.767246678Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 2 13:18:04.767282 containerd[1604]: time="2026-03-02T13:18:04.767271043Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 2 13:18:04.767335 containerd[1604]: time="2026-03-02T13:18:04.767318832Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 2 13:18:04.767972 containerd[1604]: time="2026-03-02T13:18:04.767911559Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 2 13:18:04.781110 containerd[1604]: time="2026-03-02T13:18:04.781029645Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781443648Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781474425Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781494804Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781515101Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781534167Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781551920Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781606031Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781647880Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781687623Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781704315Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781743478Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781896544Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781921150Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.782859 containerd[1604]: time="2026-03-02T13:18:04.781939494Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.781960103Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.781977705Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782022259Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782039671Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782085427Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782104573Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782221460Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782279188Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782323622Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782365359Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782387570Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782451801Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782469694Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783314 containerd[1604]: time="2026-03-02T13:18:04.782484532Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782619174Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782691118Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782708810Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782723588Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782736032Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782824447Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782918422Z" level=info msg="NRI interface is disabled by configuration." Mar 2 13:18:04.783805 containerd[1604]: time="2026-03-02T13:18:04.782957024Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 2 13:18:04.786588 containerd[1604]: time="2026-03-02T13:18:04.784341409Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 2 13:18:04.786588 containerd[1604]: time="2026-03-02T13:18:04.786445488Z" level=info msg="Connect containerd service" Mar 2 13:18:04.787241 containerd[1604]: time="2026-03-02T13:18:04.786753042Z" level=info msg="using legacy CRI server" Mar 2 13:18:04.787241 containerd[1604]: time="2026-03-02T13:18:04.786827592Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 2 13:18:04.787853 containerd[1604]: time="2026-03-02T13:18:04.787671337Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 2 13:18:04.789461 containerd[1604]: time="2026-03-02T13:18:04.789382403Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 13:18:04.789906 containerd[1604]: time="2026-03-02T13:18:04.789724534Z" level=info msg="Start subscribing containerd event" Mar 2 13:18:04.790333 containerd[1604]: time="2026-03-02T13:18:04.790259603Z" level=info msg="Start recovering state" Mar 2 13:18:04.790733 containerd[1604]: time="2026-03-02T13:18:04.790642398Z" level=info msg="Start event monitor" Mar 2 13:18:04.790925 containerd[1604]: time="2026-03-02T13:18:04.790850597Z" level=info msg="Start snapshots syncer" Mar 2 13:18:04.790925 containerd[1604]: time="2026-03-02T13:18:04.790891440Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 2 13:18:04.790999 containerd[1604]: time="2026-03-02T13:18:04.790965689Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 2 13:18:04.794993 containerd[1604]: time="2026-03-02T13:18:04.793249026Z" level=info msg="Start cni network conf syncer for default" Mar 2 13:18:04.794993 containerd[1604]: time="2026-03-02T13:18:04.793273811Z" level=info msg="Start streaming server" Mar 2 13:18:04.794993 containerd[1604]: time="2026-03-02T13:18:04.793536542Z" level=info msg="containerd successfully booted in 0.354690s" Mar 2 13:18:04.794820 systemd[1]: Started containerd.service - containerd container runtime. Mar 2 13:18:04.944012 kernel: hrtimer: interrupt took 4886435 ns Mar 2 13:18:05.179077 tar[1595]: linux-amd64/README.md Mar 2 13:18:05.218507 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 2 13:18:08.458265 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:08.463831 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 2 13:18:08.468625 systemd[1]: Startup finished in 9.455s (kernel) + 15.275s (userspace) = 24.730s. Mar 2 13:18:08.472386 (kubelet)[1688]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:18:09.751003 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 2 13:18:09.761627 systemd[1]: Started sshd@0-10.0.0.160:22-10.0.0.1:43118.service - OpenSSH per-connection server daemon (10.0.0.1:43118). Mar 2 13:18:09.883670 sshd[1698]: Accepted publickey for core from 10.0.0.1 port 43118 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:09.891434 sshd[1698]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:09.910500 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 2 13:18:09.918536 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 2 13:18:09.922273 systemd-logind[1577]: New session 1 of user core. Mar 2 13:18:10.296861 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 2 13:18:10.305547 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 2 13:18:10.338148 (systemd)[1705]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 2 13:18:10.550925 systemd[1705]: Queued start job for default target default.target. Mar 2 13:18:10.552300 systemd[1705]: Created slice app.slice - User Application Slice. Mar 2 13:18:10.552452 systemd[1705]: Reached target paths.target - Paths. Mar 2 13:18:10.552474 systemd[1705]: Reached target timers.target - Timers. Mar 2 13:18:10.567495 systemd[1705]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 2 13:18:11.249422 systemd[1705]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 2 13:18:11.249575 systemd[1705]: Reached target sockets.target - Sockets. Mar 2 13:18:11.249596 systemd[1705]: Reached target basic.target - Basic System. Mar 2 13:18:11.249661 systemd[1705]: Reached target default.target - Main User Target. Mar 2 13:18:11.249792 systemd[1705]: Startup finished in 896ms. Mar 2 13:18:11.250071 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 2 13:18:11.257982 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 2 13:18:11.309787 kubelet[1688]: E0302 13:18:11.309531 1688 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:18:11.313939 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:18:11.314365 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:18:11.341229 systemd[1]: Started sshd@1-10.0.0.160:22-10.0.0.1:43132.service - OpenSSH per-connection server daemon (10.0.0.1:43132). Mar 2 13:18:11.387593 sshd[1719]: Accepted publickey for core from 10.0.0.1 port 43132 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:11.391105 sshd[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:11.398895 systemd-logind[1577]: New session 2 of user core. Mar 2 13:18:11.404559 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 2 13:18:11.480727 sshd[1719]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:11.490046 systemd[1]: Started sshd@2-10.0.0.160:22-10.0.0.1:43144.service - OpenSSH per-connection server daemon (10.0.0.1:43144). Mar 2 13:18:11.491939 systemd[1]: sshd@1-10.0.0.160:22-10.0.0.1:43132.service: Deactivated successfully. Mar 2 13:18:11.497224 systemd[1]: session-2.scope: Deactivated successfully. Mar 2 13:18:11.550106 systemd-logind[1577]: Session 2 logged out. Waiting for processes to exit. Mar 2 13:18:11.563094 systemd-logind[1577]: Removed session 2. Mar 2 13:18:11.605461 sshd[1724]: Accepted publickey for core from 10.0.0.1 port 43144 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:11.611287 sshd[1724]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:11.644672 systemd-logind[1577]: New session 3 of user core. Mar 2 13:18:11.656815 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 2 13:18:11.766953 sshd[1724]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:11.791231 systemd[1]: Started sshd@3-10.0.0.160:22-10.0.0.1:43146.service - OpenSSH per-connection server daemon (10.0.0.1:43146). Mar 2 13:18:11.793413 systemd[1]: sshd@2-10.0.0.160:22-10.0.0.1:43144.service: Deactivated successfully. Mar 2 13:18:11.808576 systemd-logind[1577]: Session 3 logged out. Waiting for processes to exit. Mar 2 13:18:11.810393 systemd[1]: session-3.scope: Deactivated successfully. Mar 2 13:18:11.817793 systemd-logind[1577]: Removed session 3. Mar 2 13:18:11.918345 sshd[1732]: Accepted publickey for core from 10.0.0.1 port 43146 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:11.921022 sshd[1732]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:11.947285 systemd-logind[1577]: New session 4 of user core. Mar 2 13:18:11.976736 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 2 13:18:12.103987 sshd[1732]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:12.139739 systemd[1]: Started sshd@4-10.0.0.160:22-10.0.0.1:33560.service - OpenSSH per-connection server daemon (10.0.0.1:33560). Mar 2 13:18:12.142793 systemd[1]: sshd@3-10.0.0.160:22-10.0.0.1:43146.service: Deactivated successfully. Mar 2 13:18:12.150707 systemd[1]: session-4.scope: Deactivated successfully. Mar 2 13:18:12.159069 systemd-logind[1577]: Session 4 logged out. Waiting for processes to exit. Mar 2 13:18:12.168236 systemd-logind[1577]: Removed session 4. Mar 2 13:18:12.211445 sshd[1741]: Accepted publickey for core from 10.0.0.1 port 33560 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:12.223251 sshd[1741]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:12.284873 systemd-logind[1577]: New session 5 of user core. Mar 2 13:18:12.300568 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 2 13:18:12.387225 sudo[1747]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 2 13:18:12.387704 sudo[1747]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 13:18:12.406492 sudo[1747]: pam_unix(sudo:session): session closed for user root Mar 2 13:18:12.411940 sshd[1741]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:12.422671 systemd[1]: Started sshd@5-10.0.0.160:22-10.0.0.1:33562.service - OpenSSH per-connection server daemon (10.0.0.1:33562). Mar 2 13:18:12.423886 systemd[1]: sshd@4-10.0.0.160:22-10.0.0.1:33560.service: Deactivated successfully. Mar 2 13:18:12.437253 systemd-logind[1577]: Session 5 logged out. Waiting for processes to exit. Mar 2 13:18:12.437790 systemd[1]: session-5.scope: Deactivated successfully. Mar 2 13:18:12.439277 systemd-logind[1577]: Removed session 5. Mar 2 13:18:12.473457 sshd[1749]: Accepted publickey for core from 10.0.0.1 port 33562 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:12.475826 sshd[1749]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:12.485472 systemd-logind[1577]: New session 6 of user core. Mar 2 13:18:12.495689 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 2 13:18:12.562286 sudo[1757]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 2 13:18:12.562636 sudo[1757]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 13:18:12.568917 sudo[1757]: pam_unix(sudo:session): session closed for user root Mar 2 13:18:12.578349 sudo[1756]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 2 13:18:12.578787 sudo[1756]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 13:18:12.605568 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 2 13:18:12.609489 auditctl[1760]: No rules Mar 2 13:18:12.610881 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 13:18:12.611320 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 2 13:18:12.616620 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 2 13:18:12.690280 augenrules[1779]: No rules Mar 2 13:18:12.693935 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 2 13:18:12.696148 sudo[1756]: pam_unix(sudo:session): session closed for user root Mar 2 13:18:12.699076 sshd[1749]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:12.710785 systemd[1]: Started sshd@6-10.0.0.160:22-10.0.0.1:33570.service - OpenSSH per-connection server daemon (10.0.0.1:33570). Mar 2 13:18:12.712111 systemd[1]: sshd@5-10.0.0.160:22-10.0.0.1:33562.service: Deactivated successfully. Mar 2 13:18:12.715944 systemd-logind[1577]: Session 6 logged out. Waiting for processes to exit. Mar 2 13:18:12.716944 systemd[1]: session-6.scope: Deactivated successfully. Mar 2 13:18:12.721686 systemd-logind[1577]: Removed session 6. Mar 2 13:18:12.765045 sshd[1785]: Accepted publickey for core from 10.0.0.1 port 33570 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:18:12.768979 sshd[1785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:18:12.824053 systemd-logind[1577]: New session 7 of user core. Mar 2 13:18:12.850908 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 2 13:18:12.960896 sudo[1792]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 2 13:18:12.961466 sudo[1792]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 13:18:16.945553 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 2 13:18:16.949677 (dockerd)[1810]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 2 13:18:20.580712 dockerd[1810]: time="2026-03-02T13:18:20.580406060Z" level=info msg="Starting up" Mar 2 13:18:21.408972 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 2 13:18:21.425720 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:21.549688 systemd[1]: var-lib-docker-metacopy\x2dcheck989951734-merged.mount: Deactivated successfully. Mar 2 13:18:22.168664 dockerd[1810]: time="2026-03-02T13:18:22.168336164Z" level=info msg="Loading containers: start." Mar 2 13:18:22.746804 kernel: Initializing XFRM netlink socket Mar 2 13:18:22.919362 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:22.925513 (kubelet)[1921]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:18:22.928745 systemd-networkd[1252]: docker0: Link UP Mar 2 13:18:22.979891 dockerd[1810]: time="2026-03-02T13:18:22.978752847Z" level=info msg="Loading containers: done." Mar 2 13:18:23.035499 dockerd[1810]: time="2026-03-02T13:18:23.035325628Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 2 13:18:23.035706 dockerd[1810]: time="2026-03-02T13:18:23.035531383Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 2 13:18:23.035924 dockerd[1810]: time="2026-03-02T13:18:23.035859927Z" level=info msg="Daemon has completed initialization" Mar 2 13:18:23.070603 kubelet[1921]: E0302 13:18:23.070471 1921 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:18:23.078569 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:18:23.079347 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:18:23.138100 dockerd[1810]: time="2026-03-02T13:18:23.137875266Z" level=info msg="API listen on /run/docker.sock" Mar 2 13:18:23.139208 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 2 13:18:23.947066 containerd[1604]: time="2026-03-02T13:18:23.946142449Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\"" Mar 2 13:18:24.714496 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3071460461.mount: Deactivated successfully. Mar 2 13:18:27.657298 containerd[1604]: time="2026-03-02T13:18:27.657097649Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:27.658331 containerd[1604]: time="2026-03-02T13:18:27.657559575Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.9: active requests=0, bytes read=30116186" Mar 2 13:18:27.659603 containerd[1604]: time="2026-03-02T13:18:27.659529541Z" level=info msg="ImageCreate event name:\"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:27.664075 containerd[1604]: time="2026-03-02T13:18:27.663987584Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:27.665657 containerd[1604]: time="2026-03-02T13:18:27.665570043Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.9\" with image id \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\", size \"30112785\" in 3.719283303s" Mar 2 13:18:27.665712 containerd[1604]: time="2026-03-02T13:18:27.665660131Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\" returns image reference \"sha256:d3c49e1d7c1cb22893888d0d7a4142c80e16023143fdd2c0225a362ec08ab4a4\"" Mar 2 13:18:27.667804 containerd[1604]: time="2026-03-02T13:18:27.667550193Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\"" Mar 2 13:18:29.955481 containerd[1604]: time="2026-03-02T13:18:29.955084476Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:29.956984 containerd[1604]: time="2026-03-02T13:18:29.956925661Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.9: active requests=0, bytes read=26021810" Mar 2 13:18:29.958982 containerd[1604]: time="2026-03-02T13:18:29.958902946Z" level=info msg="ImageCreate event name:\"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:29.964513 containerd[1604]: time="2026-03-02T13:18:29.964400659Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:29.966267 containerd[1604]: time="2026-03-02T13:18:29.966143214Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.9\" with image id \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\", size \"27678758\" in 2.298554799s" Mar 2 13:18:29.966267 containerd[1604]: time="2026-03-02T13:18:29.966250935Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\" returns image reference \"sha256:bdbe897c17b593b8163eebd3c55c6723711b8b775bf7e554da6d75d33d114e98\"" Mar 2 13:18:29.968017 containerd[1604]: time="2026-03-02T13:18:29.967957993Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\"" Mar 2 13:18:31.162471 containerd[1604]: time="2026-03-02T13:18:31.162338825Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:31.163658 containerd[1604]: time="2026-03-02T13:18:31.163605789Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.9: active requests=0, bytes read=20162746" Mar 2 13:18:31.165667 containerd[1604]: time="2026-03-02T13:18:31.165567675Z" level=info msg="ImageCreate event name:\"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:31.171144 containerd[1604]: time="2026-03-02T13:18:31.171064529Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:31.172990 containerd[1604]: time="2026-03-02T13:18:31.172912708Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.9\" with image id \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\", size \"21819712\" in 1.204892218s" Mar 2 13:18:31.172990 containerd[1604]: time="2026-03-02T13:18:31.172971858Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\" returns image reference \"sha256:04e9a75bd404b7d5d286565ebcd5e8d5a2be3355e6cb0c3f1ab9db53fe6f180a\"" Mar 2 13:18:31.173880 containerd[1604]: time="2026-03-02T13:18:31.173632953Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\"" Mar 2 13:18:32.394260 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4181930954.mount: Deactivated successfully. Mar 2 13:18:33.162129 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 2 13:18:33.171367 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:33.397639 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:33.403417 (kubelet)[2069]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:18:33.441490 containerd[1604]: time="2026-03-02T13:18:33.441316677Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:33.443322 containerd[1604]: time="2026-03-02T13:18:33.443220151Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.9: active requests=0, bytes read=31828647" Mar 2 13:18:33.444967 containerd[1604]: time="2026-03-02T13:18:33.444874629Z" level=info msg="ImageCreate event name:\"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:33.448444 containerd[1604]: time="2026-03-02T13:18:33.448348250Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:33.449378 containerd[1604]: time="2026-03-02T13:18:33.449303987Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.9\" with image id \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\", repo tag \"registry.k8s.io/kube-proxy:v1.33.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\", size \"31827666\" in 2.275631751s" Mar 2 13:18:33.449378 containerd[1604]: time="2026-03-02T13:18:33.449349933Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\" returns image reference \"sha256:36d290108190a8d792e275b3e6ba8f1c0def0fd717573d69c3970816d945510a\"" Mar 2 13:18:33.450548 containerd[1604]: time="2026-03-02T13:18:33.450463618Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Mar 2 13:18:33.521930 kubelet[2069]: E0302 13:18:33.521872 2069 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:18:33.527533 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:18:33.527940 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:18:33.851665 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1505174033.mount: Deactivated successfully. Mar 2 13:18:36.754579 containerd[1604]: time="2026-03-02T13:18:36.754405841Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:36.756276 containerd[1604]: time="2026-03-02T13:18:36.756140351Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Mar 2 13:18:36.757639 containerd[1604]: time="2026-03-02T13:18:36.757572025Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:36.763822 containerd[1604]: time="2026-03-02T13:18:36.763698368Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:36.765588 containerd[1604]: time="2026-03-02T13:18:36.765490914Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 3.314975139s" Mar 2 13:18:36.765588 containerd[1604]: time="2026-03-02T13:18:36.765541187Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Mar 2 13:18:36.766966 containerd[1604]: time="2026-03-02T13:18:36.766878323Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 2 13:18:37.469295 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1082824903.mount: Deactivated successfully. Mar 2 13:18:37.477271 containerd[1604]: time="2026-03-02T13:18:37.477127203Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:37.478359 containerd[1604]: time="2026-03-02T13:18:37.478276749Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Mar 2 13:18:37.480348 containerd[1604]: time="2026-03-02T13:18:37.480227346Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:37.483707 containerd[1604]: time="2026-03-02T13:18:37.483602822Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:37.485043 containerd[1604]: time="2026-03-02T13:18:37.484938731Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 717.996951ms" Mar 2 13:18:37.485043 containerd[1604]: time="2026-03-02T13:18:37.484999314Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 2 13:18:37.486322 containerd[1604]: time="2026-03-02T13:18:37.486277848Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Mar 2 13:18:37.963350 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2815006778.mount: Deactivated successfully. Mar 2 13:18:40.623987 containerd[1604]: time="2026-03-02T13:18:40.623809571Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:40.625001 containerd[1604]: time="2026-03-02T13:18:40.624540035Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=23718840" Mar 2 13:18:40.626443 containerd[1604]: time="2026-03-02T13:18:40.626390875Z" level=info msg="ImageCreate event name:\"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:40.637858 containerd[1604]: time="2026-03-02T13:18:40.637696500Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:18:40.638880 containerd[1604]: time="2026-03-02T13:18:40.638730897Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"23716032\" in 3.152401252s" Mar 2 13:18:40.638880 containerd[1604]: time="2026-03-02T13:18:40.638847716Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:8cb12dd0c3e42c6d0175d09a060358cbb68a3ecc2ba4dbb00327c7d760e1425d\"" Mar 2 13:18:43.669610 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 2 13:18:43.690482 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:44.164396 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:44.174922 (kubelet)[2234]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 13:18:44.525358 kubelet[2234]: E0302 13:18:44.524665 2234 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 13:18:44.536463 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 13:18:44.537118 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 13:18:44.883343 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:44.893520 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:44.958959 systemd[1]: Reloading requested from client PID 2253 ('systemctl') (unit session-7.scope)... Mar 2 13:18:44.959008 systemd[1]: Reloading... Mar 2 13:18:45.137429 zram_generator::config[2298]: No configuration found. Mar 2 13:18:45.526647 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 2 13:18:45.760123 systemd[1]: Reloading finished in 800 ms. Mar 2 13:18:45.884654 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:45.885484 (kubelet)[2340]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 13:18:45.888861 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:45.894461 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 13:18:45.894881 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:45.897140 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:46.129398 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:46.130647 (kubelet)[2356]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 13:18:46.300531 kubelet[2356]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:18:46.300531 kubelet[2356]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 13:18:46.300531 kubelet[2356]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:18:46.301561 kubelet[2356]: I0302 13:18:46.301314 2356 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 13:18:46.656630 kubelet[2356]: I0302 13:18:46.656474 2356 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 2 13:18:46.656630 kubelet[2356]: I0302 13:18:46.656556 2356 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 13:18:46.657479 kubelet[2356]: I0302 13:18:46.657439 2356 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 13:18:46.710473 kubelet[2356]: E0302 13:18:46.710326 2356 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.160:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 13:18:46.733963 kubelet[2356]: I0302 13:18:46.733849 2356 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 13:18:46.750661 update_engine[1585]: I20260302 13:18:46.750427 1585 update_attempter.cc:509] Updating boot flags... Mar 2 13:18:46.758692 kubelet[2356]: E0302 13:18:46.757510 2356 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 2 13:18:46.759137 kubelet[2356]: I0302 13:18:46.758977 2356 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 2 13:18:46.769019 kubelet[2356]: I0302 13:18:46.768928 2356 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 2 13:18:46.769847 kubelet[2356]: I0302 13:18:46.769686 2356 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 13:18:46.770285 kubelet[2356]: I0302 13:18:46.769830 2356 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 2 13:18:46.770285 kubelet[2356]: I0302 13:18:46.770260 2356 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 13:18:46.770285 kubelet[2356]: I0302 13:18:46.770273 2356 container_manager_linux.go:303] "Creating device plugin manager" Mar 2 13:18:46.770755 kubelet[2356]: I0302 13:18:46.770661 2356 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:18:46.777845 kubelet[2356]: I0302 13:18:46.777723 2356 kubelet.go:480] "Attempting to sync node with API server" Mar 2 13:18:46.777845 kubelet[2356]: I0302 13:18:46.777794 2356 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 13:18:46.777965 kubelet[2356]: I0302 13:18:46.777920 2356 kubelet.go:386] "Adding apiserver pod source" Mar 2 13:18:46.780891 kubelet[2356]: I0302 13:18:46.780819 2356 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 13:18:46.791487 kubelet[2356]: E0302 13:18:46.791338 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.160:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:18:46.792843 kubelet[2356]: I0302 13:18:46.791560 2356 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 2 13:18:46.793715 kubelet[2356]: I0302 13:18:46.793631 2356 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 13:18:46.793715 kubelet[2356]: E0302 13:18:46.793665 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.160:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:18:46.795811 kubelet[2356]: W0302 13:18:46.795698 2356 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 2 13:18:46.803543 kubelet[2356]: I0302 13:18:46.803403 2356 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 2 13:18:46.803543 kubelet[2356]: I0302 13:18:46.803508 2356 server.go:1289] "Started kubelet" Mar 2 13:18:46.804947 kubelet[2356]: I0302 13:18:46.804713 2356 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 13:18:46.806982 kubelet[2356]: I0302 13:18:46.805376 2356 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 13:18:46.806982 kubelet[2356]: I0302 13:18:46.805804 2356 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 13:18:46.807912 kubelet[2356]: I0302 13:18:46.807877 2356 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 13:18:46.809942 kubelet[2356]: I0302 13:18:46.809874 2356 server.go:317] "Adding debug handlers to kubelet server" Mar 2 13:18:46.810711 kubelet[2356]: E0302 13:18:46.809268 2356 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.160:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.160:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.189908bad09639ef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:18:46.803446255 +0000 UTC m=+0.665498013,LastTimestamp:2026-03-02 13:18:46.803446255 +0000 UTC m=+0.665498013,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:18:46.813422 kubelet[2356]: I0302 13:18:46.813341 2356 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 13:18:46.817468 kubelet[2356]: E0302 13:18:46.817362 2356 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 13:18:46.820914 kubelet[2356]: E0302 13:18:46.820820 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:46.821090 kubelet[2356]: I0302 13:18:46.820996 2356 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 2 13:18:46.821909 kubelet[2356]: I0302 13:18:46.821868 2356 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 2 13:18:46.821976 kubelet[2356]: I0302 13:18:46.821963 2356 reconciler.go:26] "Reconciler: start to sync state" Mar 2 13:18:46.854611 kubelet[2356]: I0302 13:18:46.854533 2356 factory.go:223] Registration of the systemd container factory successfully Mar 2 13:18:46.854877 kubelet[2356]: I0302 13:18:46.854691 2356 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 13:18:46.855664 kubelet[2356]: E0302 13:18:46.855570 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.160:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:18:46.858330 kubelet[2356]: E0302 13:18:46.858247 2356 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.160:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.160:6443: connect: connection refused" interval="200ms" Mar 2 13:18:46.864730 kubelet[2356]: I0302 13:18:46.864655 2356 factory.go:223] Registration of the containerd container factory successfully Mar 2 13:18:47.112301 kubelet[2356]: E0302 13:18:47.088941 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:47.112301 kubelet[2356]: E0302 13:18:47.089309 2356 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.160:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.160:6443: connect: connection refused" interval="400ms" Mar 2 13:18:47.178587 kubelet[2356]: I0302 13:18:47.178290 2356 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 2 13:18:47.180358 kubelet[2356]: I0302 13:18:47.180287 2356 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 2 13:18:47.180503 kubelet[2356]: I0302 13:18:47.180445 2356 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 2 13:18:47.181592 kubelet[2356]: I0302 13:18:47.181531 2356 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 13:18:47.181673 kubelet[2356]: I0302 13:18:47.181611 2356 kubelet.go:2436] "Starting kubelet main sync loop" Mar 2 13:18:47.181917 kubelet[2356]: E0302 13:18:47.181704 2356 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 13:18:47.184433 kubelet[2356]: E0302 13:18:47.184328 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.160:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:18:47.191715 kubelet[2356]: E0302 13:18:47.189717 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:47.199302 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2391) Mar 2 13:18:47.204387 kubelet[2356]: I0302 13:18:47.204332 2356 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 13:18:47.204387 kubelet[2356]: I0302 13:18:47.204366 2356 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 13:18:47.204481 kubelet[2356]: I0302 13:18:47.204409 2356 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:18:47.209275 kubelet[2356]: I0302 13:18:47.207438 2356 policy_none.go:49] "None policy: Start" Mar 2 13:18:47.209275 kubelet[2356]: I0302 13:18:47.207512 2356 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 2 13:18:47.209275 kubelet[2356]: I0302 13:18:47.207570 2356 state_mem.go:35] "Initializing new in-memory state store" Mar 2 13:18:47.232255 kubelet[2356]: E0302 13:18:47.231400 2356 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 13:18:47.232255 kubelet[2356]: I0302 13:18:47.232056 2356 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 13:18:47.237018 kubelet[2356]: I0302 13:18:47.232437 2356 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 13:18:47.244582 kubelet[2356]: I0302 13:18:47.244557 2356 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 13:18:47.244828 kubelet[2356]: E0302 13:18:47.244744 2356 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 13:18:47.245061 kubelet[2356]: E0302 13:18:47.244955 2356 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 2 13:18:47.283298 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2391) Mar 2 13:18:47.317233 kubelet[2356]: E0302 13:18:47.315846 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:47.317233 kubelet[2356]: E0302 13:18:47.315953 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:47.331061 kubelet[2356]: E0302 13:18:47.330956 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:47.347894 kubelet[2356]: I0302 13:18:47.347829 2356 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:18:47.348364 kubelet[2356]: E0302 13:18:47.348325 2356 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.160:6443/api/v1/nodes\": dial tcp 10.0.0.160:6443: connect: connection refused" node="localhost" Mar 2 13:18:47.390983 kubelet[2356]: I0302 13:18:47.390758 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:47.390983 kubelet[2356]: I0302 13:18:47.390847 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:47.391254 kubelet[2356]: I0302 13:18:47.390986 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:47.391254 kubelet[2356]: I0302 13:18:47.391051 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:47.391254 kubelet[2356]: I0302 13:18:47.391087 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:47.391254 kubelet[2356]: I0302 13:18:47.391119 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:47.391254 kubelet[2356]: I0302 13:18:47.391158 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e944e4cb17af904786c3a2e01e298498-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"e944e4cb17af904786c3a2e01e298498\") " pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:47.391443 kubelet[2356]: I0302 13:18:47.391290 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:47.391443 kubelet[2356]: I0302 13:18:47.391322 2356 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:47.495047 kubelet[2356]: E0302 13:18:47.494932 2356 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.160:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.160:6443: connect: connection refused" interval="800ms" Mar 2 13:18:47.550454 kubelet[2356]: I0302 13:18:47.550404 2356 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:18:47.551149 kubelet[2356]: E0302 13:18:47.551070 2356 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.160:6443/api/v1/nodes\": dial tcp 10.0.0.160:6443: connect: connection refused" node="localhost" Mar 2 13:18:47.617233 kubelet[2356]: E0302 13:18:47.617101 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:47.617409 kubelet[2356]: E0302 13:18:47.617343 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:47.618429 containerd[1604]: time="2026-03-02T13:18:47.618371537Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2935ee7a8401bad51289f6726bd9fc6e,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:47.618887 containerd[1604]: time="2026-03-02T13:18:47.618371633Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8747e1f8a49a618fbc1324a8fe2d3754,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:47.632252 kubelet[2356]: E0302 13:18:47.632135 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:47.633132 containerd[1604]: time="2026-03-02T13:18:47.633060227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:e944e4cb17af904786c3a2e01e298498,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:47.953665 kubelet[2356]: I0302 13:18:47.953603 2356 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:18:47.954460 kubelet[2356]: E0302 13:18:47.954374 2356 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.160:6443/api/v1/nodes\": dial tcp 10.0.0.160:6443: connect: connection refused" node="localhost" Mar 2 13:18:48.042063 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1159823642.mount: Deactivated successfully. Mar 2 13:18:48.052116 containerd[1604]: time="2026-03-02T13:18:48.052023014Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:18:48.055691 containerd[1604]: time="2026-03-02T13:18:48.055617161Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Mar 2 13:18:48.056866 containerd[1604]: time="2026-03-02T13:18:48.056683329Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:18:48.057984 containerd[1604]: time="2026-03-02T13:18:48.057922521Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:18:48.059285 containerd[1604]: time="2026-03-02T13:18:48.059211917Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 2 13:18:48.060543 containerd[1604]: time="2026-03-02T13:18:48.060459730Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:18:48.061544 containerd[1604]: time="2026-03-02T13:18:48.061472598Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 2 13:18:48.069062 containerd[1604]: time="2026-03-02T13:18:48.067797068Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 13:18:48.069746 containerd[1604]: time="2026-03-02T13:18:48.069702354Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 451.091189ms" Mar 2 13:18:48.070686 containerd[1604]: time="2026-03-02T13:18:48.070628951Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 451.855423ms" Mar 2 13:18:48.074621 containerd[1604]: time="2026-03-02T13:18:48.074476184Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 441.304728ms" Mar 2 13:18:48.085137 kubelet[2356]: E0302 13:18:48.085087 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.160:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 13:18:48.099968 kubelet[2356]: E0302 13:18:48.099421 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.160:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 13:18:48.122925 kubelet[2356]: E0302 13:18:48.122856 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.160:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 13:18:48.194563 containerd[1604]: time="2026-03-02T13:18:48.194346830Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:48.194563 containerd[1604]: time="2026-03-02T13:18:48.194428523Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:48.194563 containerd[1604]: time="2026-03-02T13:18:48.194457597Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.195224 containerd[1604]: time="2026-03-02T13:18:48.195076702Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.196530 containerd[1604]: time="2026-03-02T13:18:48.196304024Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:48.196530 containerd[1604]: time="2026-03-02T13:18:48.196376390Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:48.196530 containerd[1604]: time="2026-03-02T13:18:48.196391638Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.196530 containerd[1604]: time="2026-03-02T13:18:48.196470085Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.198003 containerd[1604]: time="2026-03-02T13:18:48.197832412Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:48.198230 containerd[1604]: time="2026-03-02T13:18:48.197953998Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:48.198230 containerd[1604]: time="2026-03-02T13:18:48.198011155Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.201715 containerd[1604]: time="2026-03-02T13:18:48.201590310Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:48.292339 containerd[1604]: time="2026-03-02T13:18:48.292061074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:8747e1f8a49a618fbc1324a8fe2d3754,Namespace:kube-system,Attempt:0,} returns sandbox id \"b39cdf8ea7c182d4923e06a59c4f9880171ab30be9333cbe318c48e2a1014198\"" Mar 2 13:18:48.295362 kubelet[2356]: E0302 13:18:48.295148 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:48.296338 kubelet[2356]: E0302 13:18:48.296310 2356 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.160:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.160:6443: connect: connection refused" interval="1.6s" Mar 2 13:18:48.300365 containerd[1604]: time="2026-03-02T13:18:48.300156310Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:e944e4cb17af904786c3a2e01e298498,Namespace:kube-system,Attempt:0,} returns sandbox id \"07dfd42a908cf64a3d8b9fce4d4cacec8b728ab4ae1d54b27bfa31535fe04460\"" Mar 2 13:18:48.300568 containerd[1604]: time="2026-03-02T13:18:48.300284719Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2935ee7a8401bad51289f6726bd9fc6e,Namespace:kube-system,Attempt:0,} returns sandbox id \"041fe95a77fcc244034fcda51173f2e9dde8d95ca76e742b774a998aa2500e22\"" Mar 2 13:18:48.301444 kubelet[2356]: E0302 13:18:48.301423 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:48.301742 kubelet[2356]: E0302 13:18:48.301688 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:48.302785 containerd[1604]: time="2026-03-02T13:18:48.302707373Z" level=info msg="CreateContainer within sandbox \"b39cdf8ea7c182d4923e06a59c4f9880171ab30be9333cbe318c48e2a1014198\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 2 13:18:48.306685 containerd[1604]: time="2026-03-02T13:18:48.306646621Z" level=info msg="CreateContainer within sandbox \"07dfd42a908cf64a3d8b9fce4d4cacec8b728ab4ae1d54b27bfa31535fe04460\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 2 13:18:48.309701 containerd[1604]: time="2026-03-02T13:18:48.309560730Z" level=info msg="CreateContainer within sandbox \"041fe95a77fcc244034fcda51173f2e9dde8d95ca76e742b774a998aa2500e22\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 2 13:18:48.323487 containerd[1604]: time="2026-03-02T13:18:48.323399895Z" level=info msg="CreateContainer within sandbox \"b39cdf8ea7c182d4923e06a59c4f9880171ab30be9333cbe318c48e2a1014198\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"335c507ffe03710ede6daf7df6f68c2c36550cc2b7d81f6c304d3f013d65e05b\"" Mar 2 13:18:48.324369 containerd[1604]: time="2026-03-02T13:18:48.324287193Z" level=info msg="StartContainer for \"335c507ffe03710ede6daf7df6f68c2c36550cc2b7d81f6c304d3f013d65e05b\"" Mar 2 13:18:48.332235 containerd[1604]: time="2026-03-02T13:18:48.332048443Z" level=info msg="CreateContainer within sandbox \"07dfd42a908cf64a3d8b9fce4d4cacec8b728ab4ae1d54b27bfa31535fe04460\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c3cf40c529758514bb44842c93b9e743148a0c8fbffefe3b59a3bd80bae2579a\"" Mar 2 13:18:48.332602 containerd[1604]: time="2026-03-02T13:18:48.332565688Z" level=info msg="StartContainer for \"c3cf40c529758514bb44842c93b9e743148a0c8fbffefe3b59a3bd80bae2579a\"" Mar 2 13:18:48.337821 containerd[1604]: time="2026-03-02T13:18:48.337703136Z" level=info msg="CreateContainer within sandbox \"041fe95a77fcc244034fcda51173f2e9dde8d95ca76e742b774a998aa2500e22\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"1886c25294f23e30c491b8d8af6ab46c39695b4350d85f2786060094b62cf5aa\"" Mar 2 13:18:48.338434 containerd[1604]: time="2026-03-02T13:18:48.338390811Z" level=info msg="StartContainer for \"1886c25294f23e30c491b8d8af6ab46c39695b4350d85f2786060094b62cf5aa\"" Mar 2 13:18:48.380103 kubelet[2356]: E0302 13:18:48.380050 2356 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.160:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.160:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 13:18:48.418809 containerd[1604]: time="2026-03-02T13:18:48.418738354Z" level=info msg="StartContainer for \"335c507ffe03710ede6daf7df6f68c2c36550cc2b7d81f6c304d3f013d65e05b\" returns successfully" Mar 2 13:18:48.428218 containerd[1604]: time="2026-03-02T13:18:48.425321489Z" level=info msg="StartContainer for \"1886c25294f23e30c491b8d8af6ab46c39695b4350d85f2786060094b62cf5aa\" returns successfully" Mar 2 13:18:48.428309 containerd[1604]: time="2026-03-02T13:18:48.428229716Z" level=info msg="StartContainer for \"c3cf40c529758514bb44842c93b9e743148a0c8fbffefe3b59a3bd80bae2579a\" returns successfully" Mar 2 13:18:48.760908 kubelet[2356]: I0302 13:18:48.758313 2356 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:18:49.198227 kubelet[2356]: E0302 13:18:49.198142 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:49.198380 kubelet[2356]: E0302 13:18:49.198338 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:49.200614 kubelet[2356]: E0302 13:18:49.200566 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:49.200783 kubelet[2356]: E0302 13:18:49.200723 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:49.202645 kubelet[2356]: E0302 13:18:49.202592 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:49.202749 kubelet[2356]: E0302 13:18:49.202710 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:50.096226 kubelet[2356]: E0302 13:18:50.095066 2356 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 2 13:18:50.182973 kubelet[2356]: I0302 13:18:50.182926 2356 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 13:18:50.182973 kubelet[2356]: E0302 13:18:50.182968 2356 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Mar 2 13:18:50.194951 kubelet[2356]: E0302 13:18:50.194866 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.204671 kubelet[2356]: E0302 13:18:50.204651 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:50.204730 kubelet[2356]: E0302 13:18:50.204701 2356 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Mar 2 13:18:50.204894 kubelet[2356]: E0302 13:18:50.204833 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:50.204894 kubelet[2356]: E0302 13:18:50.204833 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:50.224758 kubelet[2356]: E0302 13:18:50.224485 2356 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.189908bad09639ef default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-03-02 13:18:46.803446255 +0000 UTC m=+0.665498013,LastTimestamp:2026-03-02 13:18:46.803446255 +0000 UTC m=+0.665498013,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 2 13:18:50.295522 kubelet[2356]: E0302 13:18:50.295433 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.396119 kubelet[2356]: E0302 13:18:50.395897 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.496311 kubelet[2356]: E0302 13:18:50.496258 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.596953 kubelet[2356]: E0302 13:18:50.596906 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.697586 kubelet[2356]: E0302 13:18:50.697435 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.797695 kubelet[2356]: E0302 13:18:50.797582 2356 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 2 13:18:50.921902 kubelet[2356]: I0302 13:18:50.921794 2356 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:50.930320 kubelet[2356]: E0302 13:18:50.930233 2356 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:50.930320 kubelet[2356]: I0302 13:18:50.930272 2356 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:50.932243 kubelet[2356]: E0302 13:18:50.932120 2356 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:50.932243 kubelet[2356]: I0302 13:18:50.932150 2356 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:50.933510 kubelet[2356]: E0302 13:18:50.933464 2356 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:51.205864 kubelet[2356]: I0302 13:18:51.205815 2356 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:51.219225 kubelet[2356]: E0302 13:18:51.219103 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:51.788597 kubelet[2356]: I0302 13:18:51.788500 2356 apiserver.go:52] "Watching apiserver" Mar 2 13:18:51.822715 kubelet[2356]: I0302 13:18:51.822577 2356 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 2 13:18:52.207752 kubelet[2356]: E0302 13:18:52.207640 2356 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:52.601075 systemd[1]: Reloading requested from client PID 2652 ('systemctl') (unit session-7.scope)... Mar 2 13:18:52.601104 systemd[1]: Reloading... Mar 2 13:18:52.699248 zram_generator::config[2691]: No configuration found. Mar 2 13:18:52.812637 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 2 13:18:52.895253 systemd[1]: Reloading finished in 293 ms. Mar 2 13:18:52.937805 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:52.964480 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 13:18:52.964893 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:52.970533 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 13:18:53.130897 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 13:18:53.136300 (kubelet)[2746]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 13:18:53.187817 kubelet[2746]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:18:53.187817 kubelet[2746]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 13:18:53.187817 kubelet[2746]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 13:18:53.188362 kubelet[2746]: I0302 13:18:53.187984 2746 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 13:18:53.196220 kubelet[2746]: I0302 13:18:53.196097 2746 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 2 13:18:53.196220 kubelet[2746]: I0302 13:18:53.196147 2746 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 13:18:53.196579 kubelet[2746]: I0302 13:18:53.196518 2746 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 13:18:53.198396 kubelet[2746]: I0302 13:18:53.198362 2746 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 2 13:18:53.201315 kubelet[2746]: I0302 13:18:53.201269 2746 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 13:18:53.205586 kubelet[2746]: E0302 13:18:53.205551 2746 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 2 13:18:53.205631 kubelet[2746]: I0302 13:18:53.205588 2746 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 2 13:18:53.210728 kubelet[2746]: I0302 13:18:53.210652 2746 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 2 13:18:53.211319 kubelet[2746]: I0302 13:18:53.211246 2746 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 13:18:53.211431 kubelet[2746]: I0302 13:18:53.211285 2746 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Mar 2 13:18:53.211431 kubelet[2746]: I0302 13:18:53.211410 2746 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 13:18:53.211431 kubelet[2746]: I0302 13:18:53.211419 2746 container_manager_linux.go:303] "Creating device plugin manager" Mar 2 13:18:53.211572 kubelet[2746]: I0302 13:18:53.211465 2746 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:18:53.211733 kubelet[2746]: I0302 13:18:53.211676 2746 kubelet.go:480] "Attempting to sync node with API server" Mar 2 13:18:53.211733 kubelet[2746]: I0302 13:18:53.211700 2746 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 13:18:53.211733 kubelet[2746]: I0302 13:18:53.211723 2746 kubelet.go:386] "Adding apiserver pod source" Mar 2 13:18:53.211820 kubelet[2746]: I0302 13:18:53.211737 2746 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 13:18:53.212821 kubelet[2746]: I0302 13:18:53.212721 2746 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 2 13:18:53.214268 kubelet[2746]: I0302 13:18:53.213340 2746 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 13:18:53.223428 kubelet[2746]: I0302 13:18:53.223380 2746 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 2 13:18:53.223519 kubelet[2746]: I0302 13:18:53.223442 2746 server.go:1289] "Started kubelet" Mar 2 13:18:53.223565 kubelet[2746]: I0302 13:18:53.223542 2746 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 13:18:53.224253 kubelet[2746]: I0302 13:18:53.224085 2746 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 13:18:53.228795 kubelet[2746]: I0302 13:18:53.228736 2746 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 13:18:53.229517 kubelet[2746]: I0302 13:18:53.229459 2746 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 13:18:53.233717 kubelet[2746]: I0302 13:18:53.231414 2746 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 13:18:53.233717 kubelet[2746]: I0302 13:18:53.232240 2746 server.go:317] "Adding debug handlers to kubelet server" Mar 2 13:18:53.233900 kubelet[2746]: I0302 13:18:53.233884 2746 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 2 13:18:53.234155 kubelet[2746]: I0302 13:18:53.233992 2746 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 2 13:18:53.234709 kubelet[2746]: I0302 13:18:53.234695 2746 reconciler.go:26] "Reconciler: start to sync state" Mar 2 13:18:53.235849 kubelet[2746]: I0302 13:18:53.235809 2746 factory.go:223] Registration of the systemd container factory successfully Mar 2 13:18:53.235976 kubelet[2746]: I0302 13:18:53.235911 2746 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 13:18:53.243731 kubelet[2746]: I0302 13:18:53.243670 2746 factory.go:223] Registration of the containerd container factory successfully Mar 2 13:18:53.245283 kubelet[2746]: E0302 13:18:53.245251 2746 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 13:18:53.251742 kubelet[2746]: I0302 13:18:53.251701 2746 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 2 13:18:53.253894 kubelet[2746]: I0302 13:18:53.253875 2746 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 2 13:18:53.254136 kubelet[2746]: I0302 13:18:53.254003 2746 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 2 13:18:53.254281 kubelet[2746]: I0302 13:18:53.254031 2746 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 13:18:53.254621 kubelet[2746]: I0302 13:18:53.254340 2746 kubelet.go:2436] "Starting kubelet main sync loop" Mar 2 13:18:53.254621 kubelet[2746]: E0302 13:18:53.254398 2746 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 13:18:53.293336 kubelet[2746]: I0302 13:18:53.293268 2746 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 13:18:53.293336 kubelet[2746]: I0302 13:18:53.293297 2746 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 13:18:53.293336 kubelet[2746]: I0302 13:18:53.293316 2746 state_mem.go:36] "Initialized new in-memory state store" Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293507 2746 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293526 2746 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293552 2746 policy_none.go:49] "None policy: Start" Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293573 2746 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293585 2746 state_mem.go:35] "Initializing new in-memory state store" Mar 2 13:18:53.294887 kubelet[2746]: I0302 13:18:53.293692 2746 state_mem.go:75] "Updated machine memory state" Mar 2 13:18:53.296291 kubelet[2746]: E0302 13:18:53.296255 2746 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 13:18:53.296457 kubelet[2746]: I0302 13:18:53.296423 2746 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 13:18:53.296485 kubelet[2746]: I0302 13:18:53.296450 2746 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 13:18:53.299265 kubelet[2746]: I0302 13:18:53.298269 2746 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 13:18:53.299467 kubelet[2746]: E0302 13:18:53.299387 2746 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 13:18:53.355360 kubelet[2746]: I0302 13:18:53.355295 2746 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:53.355585 kubelet[2746]: I0302 13:18:53.355489 2746 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.355675 kubelet[2746]: I0302 13:18:53.355374 2746 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:53.366524 kubelet[2746]: E0302 13:18:53.366437 2746 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:53.403316 kubelet[2746]: I0302 13:18:53.403156 2746 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Mar 2 13:18:53.415729 kubelet[2746]: I0302 13:18:53.415692 2746 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Mar 2 13:18:53.415804 kubelet[2746]: I0302 13:18:53.415786 2746 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Mar 2 13:18:53.436163 kubelet[2746]: I0302 13:18:53.436106 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:53.436163 kubelet[2746]: I0302 13:18:53.436145 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:53.436163 kubelet[2746]: I0302 13:18:53.436206 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.436336 kubelet[2746]: I0302 13:18:53.436223 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.436336 kubelet[2746]: I0302 13:18:53.436238 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.436336 kubelet[2746]: I0302 13:18:53.436251 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.436336 kubelet[2746]: I0302 13:18:53.436265 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8747e1f8a49a618fbc1324a8fe2d3754-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"8747e1f8a49a618fbc1324a8fe2d3754\") " pod="kube-system/kube-controller-manager-localhost" Mar 2 13:18:53.436336 kubelet[2746]: I0302 13:18:53.436279 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e944e4cb17af904786c3a2e01e298498-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"e944e4cb17af904786c3a2e01e298498\") " pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:53.436486 kubelet[2746]: I0302 13:18:53.436291 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2935ee7a8401bad51289f6726bd9fc6e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2935ee7a8401bad51289f6726bd9fc6e\") " pod="kube-system/kube-apiserver-localhost" Mar 2 13:18:53.605362 sudo[2785]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 2 13:18:53.605867 sudo[2785]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 2 13:18:53.664823 kubelet[2746]: E0302 13:18:53.664679 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:53.667152 kubelet[2746]: E0302 13:18:53.667076 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:53.667261 kubelet[2746]: E0302 13:18:53.667156 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:54.116300 sudo[2785]: pam_unix(sudo:session): session closed for user root Mar 2 13:18:54.213062 kubelet[2746]: I0302 13:18:54.212969 2746 apiserver.go:52] "Watching apiserver" Mar 2 13:18:54.234356 kubelet[2746]: I0302 13:18:54.234298 2746 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 2 13:18:54.269252 kubelet[2746]: E0302 13:18:54.269164 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:54.269430 kubelet[2746]: E0302 13:18:54.269402 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:54.269734 kubelet[2746]: I0302 13:18:54.269621 2746 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:54.279449 kubelet[2746]: E0302 13:18:54.279373 2746 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Mar 2 13:18:54.279640 kubelet[2746]: E0302 13:18:54.279590 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:54.298468 kubelet[2746]: I0302 13:18:54.298360 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.298348588 podStartE2EDuration="1.298348588s" podCreationTimestamp="2026-03-02 13:18:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:18:54.297035136 +0000 UTC m=+1.152146802" watchObservedRunningTime="2026-03-02 13:18:54.298348588 +0000 UTC m=+1.153460252" Mar 2 13:18:54.311358 kubelet[2746]: I0302 13:18:54.310944 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.310931658 podStartE2EDuration="3.310931658s" podCreationTimestamp="2026-03-02 13:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:18:54.310390278 +0000 UTC m=+1.165501963" watchObservedRunningTime="2026-03-02 13:18:54.310931658 +0000 UTC m=+1.166043324" Mar 2 13:18:54.352060 kubelet[2746]: I0302 13:18:54.352005 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.351986718 podStartE2EDuration="1.351986718s" podCreationTimestamp="2026-03-02 13:18:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:18:54.32305126 +0000 UTC m=+1.178162935" watchObservedRunningTime="2026-03-02 13:18:54.351986718 +0000 UTC m=+1.207098383" Mar 2 13:18:55.270831 kubelet[2746]: E0302 13:18:55.270603 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:55.272539 kubelet[2746]: E0302 13:18:55.272163 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:55.525040 sudo[1792]: pam_unix(sudo:session): session closed for user root Mar 2 13:18:55.528521 sshd[1785]: pam_unix(sshd:session): session closed for user core Mar 2 13:18:55.532710 systemd[1]: sshd@6-10.0.0.160:22-10.0.0.1:33570.service: Deactivated successfully. Mar 2 13:18:55.535743 systemd-logind[1577]: Session 7 logged out. Waiting for processes to exit. Mar 2 13:18:55.535998 systemd[1]: session-7.scope: Deactivated successfully. Mar 2 13:18:55.537211 systemd-logind[1577]: Removed session 7. Mar 2 13:18:57.329081 kubelet[2746]: I0302 13:18:57.329031 2746 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 2 13:18:57.329750 containerd[1604]: time="2026-03-02T13:18:57.329687630Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 2 13:18:57.330241 kubelet[2746]: I0302 13:18:57.330158 2746 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 2 13:18:58.252544 kubelet[2746]: E0302 13:18:58.252450 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.276220 kubelet[2746]: E0302 13:18:58.276052 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.474435 kubelet[2746]: I0302 13:18:58.473748 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-etc-cni-netd\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.474435 kubelet[2746]: I0302 13:18:58.473861 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ab2093cb-5670-4b75-883e-9e03aa6693fa-clustermesh-secrets\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.474435 kubelet[2746]: I0302 13:18:58.473895 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd2l4\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-kube-api-access-wd2l4\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.474435 kubelet[2746]: I0302 13:18:58.473923 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/9923dfeb-eab9-4698-b6a8-13ff8f93641c-kube-proxy\") pod \"kube-proxy-2w5sl\" (UID: \"9923dfeb-eab9-4698-b6a8-13ff8f93641c\") " pod="kube-system/kube-proxy-2w5sl" Mar 2 13:18:58.474435 kubelet[2746]: I0302 13:18:58.473941 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9923dfeb-eab9-4698-b6a8-13ff8f93641c-lib-modules\") pod \"kube-proxy-2w5sl\" (UID: \"9923dfeb-eab9-4698-b6a8-13ff8f93641c\") " pod="kube-system/kube-proxy-2w5sl" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.473967 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-cgroup\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.473987 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-lib-modules\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.474005 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-xtables-lock\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.474027 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9923dfeb-eab9-4698-b6a8-13ff8f93641c-xtables-lock\") pod \"kube-proxy-2w5sl\" (UID: \"9923dfeb-eab9-4698-b6a8-13ff8f93641c\") " pod="kube-system/kube-proxy-2w5sl" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.474045 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cni-path\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475016 kubelet[2746]: I0302 13:18:58.474070 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-run\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475145 kubelet[2746]: I0302 13:18:58.474096 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-hubble-tls\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475145 kubelet[2746]: I0302 13:18:58.474240 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-config-path\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475145 kubelet[2746]: I0302 13:18:58.474260 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-net\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475145 kubelet[2746]: I0302 13:18:58.474274 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-kernel\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475145 kubelet[2746]: I0302 13:18:58.474322 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw8gc\" (UniqueName: \"kubernetes.io/projected/9923dfeb-eab9-4698-b6a8-13ff8f93641c-kube-api-access-mw8gc\") pod \"kube-proxy-2w5sl\" (UID: \"9923dfeb-eab9-4698-b6a8-13ff8f93641c\") " pod="kube-system/kube-proxy-2w5sl" Mar 2 13:18:58.475298 kubelet[2746]: I0302 13:18:58.474337 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-bpf-maps\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.475298 kubelet[2746]: I0302 13:18:58.474350 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-hostproc\") pod \"cilium-28vz5\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " pod="kube-system/cilium-28vz5" Mar 2 13:18:58.575256 kubelet[2746]: I0302 13:18:58.574894 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5mkt\" (UniqueName: \"kubernetes.io/projected/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-kube-api-access-v5mkt\") pod \"cilium-operator-6c4d7847fc-grgmr\" (UID: \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\") " pod="kube-system/cilium-operator-6c4d7847fc-grgmr" Mar 2 13:18:58.575256 kubelet[2746]: I0302 13:18:58.574931 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-grgmr\" (UID: \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\") " pod="kube-system/cilium-operator-6c4d7847fc-grgmr" Mar 2 13:18:58.685041 kubelet[2746]: E0302 13:18:58.684965 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.686659 containerd[1604]: time="2026-03-02T13:18:58.686253986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2w5sl,Uid:9923dfeb-eab9-4698-b6a8-13ff8f93641c,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:58.690924 kubelet[2746]: E0302 13:18:58.689302 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.691032 containerd[1604]: time="2026-03-02T13:18:58.689840088Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-28vz5,Uid:ab2093cb-5670-4b75-883e-9e03aa6693fa,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:58.740382 containerd[1604]: time="2026-03-02T13:18:58.736494139Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:58.740382 containerd[1604]: time="2026-03-02T13:18:58.740271843Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:58.740382 containerd[1604]: time="2026-03-02T13:18:58.740283935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:58.740808 containerd[1604]: time="2026-03-02T13:18:58.740376448Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:58.761625 containerd[1604]: time="2026-03-02T13:18:58.761315703Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:58.761625 containerd[1604]: time="2026-03-02T13:18:58.761380924Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:58.761625 containerd[1604]: time="2026-03-02T13:18:58.761407594Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:58.761625 containerd[1604]: time="2026-03-02T13:18:58.761506198Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:58.829695 kubelet[2746]: E0302 13:18:58.829576 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.834234 containerd[1604]: time="2026-03-02T13:18:58.833477104Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-grgmr,Uid:e6953e94-ca7c-4e4a-91a9-af1c54d6459d,Namespace:kube-system,Attempt:0,}" Mar 2 13:18:58.834234 containerd[1604]: time="2026-03-02T13:18:58.833887683Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2w5sl,Uid:9923dfeb-eab9-4698-b6a8-13ff8f93641c,Namespace:kube-system,Attempt:0,} returns sandbox id \"2ef08d85845d347fbb8728aa0bab57ee9e5cb75208efd7e2b27ab4a245bf7001\"" Mar 2 13:18:58.835488 kubelet[2746]: E0302 13:18:58.835433 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.842590 containerd[1604]: time="2026-03-02T13:18:58.842530681Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-28vz5,Uid:ab2093cb-5670-4b75-883e-9e03aa6693fa,Namespace:kube-system,Attempt:0,} returns sandbox id \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\"" Mar 2 13:18:58.843549 kubelet[2746]: E0302 13:18:58.843513 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:58.844741 containerd[1604]: time="2026-03-02T13:18:58.844693856Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 2 13:18:58.896347 containerd[1604]: time="2026-03-02T13:18:58.896299300Z" level=info msg="CreateContainer within sandbox \"2ef08d85845d347fbb8728aa0bab57ee9e5cb75208efd7e2b27ab4a245bf7001\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 2 13:18:58.918513 containerd[1604]: time="2026-03-02T13:18:58.918445624Z" level=info msg="CreateContainer within sandbox \"2ef08d85845d347fbb8728aa0bab57ee9e5cb75208efd7e2b27ab4a245bf7001\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"460cd241c232c458ba6b37516beb0ca1e7b4e921063a54d131f170af0a188d26\"" Mar 2 13:18:58.919387 containerd[1604]: time="2026-03-02T13:18:58.919322520Z" level=info msg="StartContainer for \"460cd241c232c458ba6b37516beb0ca1e7b4e921063a54d131f170af0a188d26\"" Mar 2 13:18:58.932962 containerd[1604]: time="2026-03-02T13:18:58.932848408Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:18:58.933243 containerd[1604]: time="2026-03-02T13:18:58.932950609Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:18:58.933243 containerd[1604]: time="2026-03-02T13:18:58.932964405Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:58.933243 containerd[1604]: time="2026-03-02T13:18:58.933071114Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:18:59.006354 containerd[1604]: time="2026-03-02T13:18:59.004265798Z" level=info msg="StartContainer for \"460cd241c232c458ba6b37516beb0ca1e7b4e921063a54d131f170af0a188d26\" returns successfully" Mar 2 13:18:59.006354 containerd[1604]: time="2026-03-02T13:18:59.005516377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-grgmr,Uid:e6953e94-ca7c-4e4a-91a9-af1c54d6459d,Namespace:kube-system,Attempt:0,} returns sandbox id \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\"" Mar 2 13:18:59.010243 kubelet[2746]: E0302 13:18:59.009128 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:59.279031 kubelet[2746]: E0302 13:18:59.278985 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:18:59.294453 kubelet[2746]: I0302 13:18:59.294399 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-2w5sl" podStartSLOduration=1.29438465 podStartE2EDuration="1.29438465s" podCreationTimestamp="2026-03-02 13:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:18:59.293662845 +0000 UTC m=+6.148774531" watchObservedRunningTime="2026-03-02 13:18:59.29438465 +0000 UTC m=+6.149496315" Mar 2 13:19:00.856211 kubelet[2746]: E0302 13:19:00.854545 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:01.289040 kubelet[2746]: E0302 13:19:01.288937 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:02.519255 kubelet[2746]: E0302 13:19:02.518486 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:04.761335 kubelet[2746]: E0302 13:19:04.756813 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:06.193815 kubelet[2746]: E0302 13:19:06.193742 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:07.195215 kubelet[2746]: E0302 13:19:07.195109 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:07.792596 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3229076718.mount: Deactivated successfully. Mar 2 13:19:09.744116 containerd[1604]: time="2026-03-02T13:19:09.744029936Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:19:09.745536 containerd[1604]: time="2026-03-02T13:19:09.745406439Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Mar 2 13:19:09.746490 containerd[1604]: time="2026-03-02T13:19:09.746438585Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:19:09.748584 containerd[1604]: time="2026-03-02T13:19:09.748546720Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 10.903791077s" Mar 2 13:19:09.748633 containerd[1604]: time="2026-03-02T13:19:09.748590000Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Mar 2 13:19:09.754519 containerd[1604]: time="2026-03-02T13:19:09.754480149Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 2 13:19:09.758287 containerd[1604]: time="2026-03-02T13:19:09.758240487Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 2 13:19:09.772846 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3790620103.mount: Deactivated successfully. Mar 2 13:19:09.776197 containerd[1604]: time="2026-03-02T13:19:09.776089640Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\"" Mar 2 13:19:09.777046 containerd[1604]: time="2026-03-02T13:19:09.776898913Z" level=info msg="StartContainer for \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\"" Mar 2 13:19:09.877478 containerd[1604]: time="2026-03-02T13:19:09.877360337Z" level=info msg="StartContainer for \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\" returns successfully" Mar 2 13:19:10.033770 containerd[1604]: time="2026-03-02T13:19:10.033535220Z" level=info msg="shim disconnected" id=6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0 namespace=k8s.io Mar 2 13:19:10.033770 containerd[1604]: time="2026-03-02T13:19:10.033717339Z" level=warning msg="cleaning up after shim disconnected" id=6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0 namespace=k8s.io Mar 2 13:19:10.033770 containerd[1604]: time="2026-03-02T13:19:10.033741054Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:19:10.204424 kubelet[2746]: E0302 13:19:10.204361 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:10.212135 containerd[1604]: time="2026-03-02T13:19:10.211980344Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 2 13:19:10.235674 containerd[1604]: time="2026-03-02T13:19:10.235622419Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\"" Mar 2 13:19:10.237731 containerd[1604]: time="2026-03-02T13:19:10.237517130Z" level=info msg="StartContainer for \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\"" Mar 2 13:19:10.315365 containerd[1604]: time="2026-03-02T13:19:10.315162353Z" level=info msg="StartContainer for \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\" returns successfully" Mar 2 13:19:10.330355 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 13:19:10.330662 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:19:10.330735 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:19:10.341494 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 13:19:10.368228 containerd[1604]: time="2026-03-02T13:19:10.368068115Z" level=info msg="shim disconnected" id=b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b namespace=k8s.io Mar 2 13:19:10.368228 containerd[1604]: time="2026-03-02T13:19:10.368147774Z" level=warning msg="cleaning up after shim disconnected" id=b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b namespace=k8s.io Mar 2 13:19:10.368228 containerd[1604]: time="2026-03-02T13:19:10.368157692Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:19:10.370889 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 13:19:10.770651 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0-rootfs.mount: Deactivated successfully. Mar 2 13:19:11.208113 kubelet[2746]: E0302 13:19:11.208080 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:11.212585 containerd[1604]: time="2026-03-02T13:19:11.212505783Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 2 13:19:11.232459 containerd[1604]: time="2026-03-02T13:19:11.232386421Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\"" Mar 2 13:19:11.233980 containerd[1604]: time="2026-03-02T13:19:11.233239616Z" level=info msg="StartContainer for \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\"" Mar 2 13:19:11.329634 containerd[1604]: time="2026-03-02T13:19:11.329559103Z" level=info msg="StartContainer for \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\" returns successfully" Mar 2 13:19:11.370945 containerd[1604]: time="2026-03-02T13:19:11.370841314Z" level=info msg="shim disconnected" id=35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6 namespace=k8s.io Mar 2 13:19:11.370945 containerd[1604]: time="2026-03-02T13:19:11.370925853Z" level=warning msg="cleaning up after shim disconnected" id=35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6 namespace=k8s.io Mar 2 13:19:11.370945 containerd[1604]: time="2026-03-02T13:19:11.370938706Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:19:11.770058 systemd[1]: run-containerd-runc-k8s.io-35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6-runc.pGurj0.mount: Deactivated successfully. Mar 2 13:19:11.770331 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6-rootfs.mount: Deactivated successfully. Mar 2 13:19:12.213226 kubelet[2746]: E0302 13:19:12.213137 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:12.218919 containerd[1604]: time="2026-03-02T13:19:12.218867897Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 2 13:19:12.237881 containerd[1604]: time="2026-03-02T13:19:12.237753772Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\"" Mar 2 13:19:12.239051 containerd[1604]: time="2026-03-02T13:19:12.238942662Z" level=info msg="StartContainer for \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\"" Mar 2 13:19:12.324243 containerd[1604]: time="2026-03-02T13:19:12.322657160Z" level=info msg="StartContainer for \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\" returns successfully" Mar 2 13:19:12.367085 containerd[1604]: time="2026-03-02T13:19:12.367005279Z" level=info msg="shim disconnected" id=a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418 namespace=k8s.io Mar 2 13:19:12.367085 containerd[1604]: time="2026-03-02T13:19:12.367078635Z" level=warning msg="cleaning up after shim disconnected" id=a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418 namespace=k8s.io Mar 2 13:19:12.367085 containerd[1604]: time="2026-03-02T13:19:12.367088785Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:19:12.771531 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418-rootfs.mount: Deactivated successfully. Mar 2 13:19:12.965566 containerd[1604]: time="2026-03-02T13:19:12.965479253Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:19:12.966367 containerd[1604]: time="2026-03-02T13:19:12.966326366Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Mar 2 13:19:12.967527 containerd[1604]: time="2026-03-02T13:19:12.967497531Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 13:19:12.969760 containerd[1604]: time="2026-03-02T13:19:12.969714751Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.215199487s" Mar 2 13:19:12.969760 containerd[1604]: time="2026-03-02T13:19:12.969754385Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Mar 2 13:19:12.974541 containerd[1604]: time="2026-03-02T13:19:12.974490178Z" level=info msg="CreateContainer within sandbox \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 2 13:19:12.988051 containerd[1604]: time="2026-03-02T13:19:12.987988390Z" level=info msg="CreateContainer within sandbox \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\"" Mar 2 13:19:12.988582 containerd[1604]: time="2026-03-02T13:19:12.988543047Z" level=info msg="StartContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\"" Mar 2 13:19:13.117725 containerd[1604]: time="2026-03-02T13:19:13.117550346Z" level=info msg="StartContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" returns successfully" Mar 2 13:19:13.233632 kubelet[2746]: E0302 13:19:13.233565 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:13.249983 kubelet[2746]: E0302 13:19:13.247325 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:13.254349 containerd[1604]: time="2026-03-02T13:19:13.254291601Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 2 13:19:13.276287 kubelet[2746]: I0302 13:19:13.276140 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-grgmr" podStartSLOduration=1.315990424 podStartE2EDuration="15.276120955s" podCreationTimestamp="2026-03-02 13:18:58 +0000 UTC" firstStartedPulling="2026-03-02 13:18:59.010303523 +0000 UTC m=+5.865415188" lastFinishedPulling="2026-03-02 13:19:12.970434055 +0000 UTC m=+19.825545719" observedRunningTime="2026-03-02 13:19:13.273732348 +0000 UTC m=+20.128844073" watchObservedRunningTime="2026-03-02 13:19:13.276120955 +0000 UTC m=+20.131232621" Mar 2 13:19:13.302060 containerd[1604]: time="2026-03-02T13:19:13.301986548Z" level=info msg="CreateContainer within sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\"" Mar 2 13:19:13.311551 containerd[1604]: time="2026-03-02T13:19:13.311498629Z" level=info msg="StartContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\"" Mar 2 13:19:13.417290 containerd[1604]: time="2026-03-02T13:19:13.417106887Z" level=info msg="StartContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" returns successfully" Mar 2 13:19:13.641394 kubelet[2746]: I0302 13:19:13.640621 2746 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Mar 2 13:19:13.777686 kubelet[2746]: I0302 13:19:13.777555 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lksnl\" (UniqueName: \"kubernetes.io/projected/2a5ccb95-fe38-475a-9921-a7cf6e289400-kube-api-access-lksnl\") pod \"coredns-674b8bbfcf-kvlf5\" (UID: \"2a5ccb95-fe38-475a-9921-a7cf6e289400\") " pod="kube-system/coredns-674b8bbfcf-kvlf5" Mar 2 13:19:13.778126 kubelet[2746]: I0302 13:19:13.778108 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a5ccb95-fe38-475a-9921-a7cf6e289400-config-volume\") pod \"coredns-674b8bbfcf-kvlf5\" (UID: \"2a5ccb95-fe38-475a-9921-a7cf6e289400\") " pod="kube-system/coredns-674b8bbfcf-kvlf5" Mar 2 13:19:13.778437 kubelet[2746]: I0302 13:19:13.778384 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6fmk\" (UniqueName: \"kubernetes.io/projected/5d8e6175-0857-440d-822b-eaaf28757b1a-kube-api-access-p6fmk\") pod \"coredns-674b8bbfcf-mrscv\" (UID: \"5d8e6175-0857-440d-822b-eaaf28757b1a\") " pod="kube-system/coredns-674b8bbfcf-mrscv" Mar 2 13:19:13.778521 kubelet[2746]: I0302 13:19:13.778508 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5d8e6175-0857-440d-822b-eaaf28757b1a-config-volume\") pod \"coredns-674b8bbfcf-mrscv\" (UID: \"5d8e6175-0857-440d-822b-eaaf28757b1a\") " pod="kube-system/coredns-674b8bbfcf-mrscv" Mar 2 13:19:14.038353 kubelet[2746]: E0302 13:19:14.038236 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:14.042088 kubelet[2746]: E0302 13:19:14.039066 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:14.043327 containerd[1604]: time="2026-03-02T13:19:14.039421741Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mrscv,Uid:5d8e6175-0857-440d-822b-eaaf28757b1a,Namespace:kube-system,Attempt:0,}" Mar 2 13:19:14.043327 containerd[1604]: time="2026-03-02T13:19:14.040521260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kvlf5,Uid:2a5ccb95-fe38-475a-9921-a7cf6e289400,Namespace:kube-system,Attempt:0,}" Mar 2 13:19:14.262094 kubelet[2746]: E0302 13:19:14.262040 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:14.266625 kubelet[2746]: E0302 13:19:14.266570 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:14.296159 kubelet[2746]: I0302 13:19:14.295725 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-28vz5" podStartSLOduration=5.385761029 podStartE2EDuration="16.295701653s" podCreationTimestamp="2026-03-02 13:18:58 +0000 UTC" firstStartedPulling="2026-03-02 13:18:58.844218264 +0000 UTC m=+5.699329929" lastFinishedPulling="2026-03-02 13:19:09.754158887 +0000 UTC m=+16.609270553" observedRunningTime="2026-03-02 13:19:14.295681408 +0000 UTC m=+21.150793073" watchObservedRunningTime="2026-03-02 13:19:14.295701653 +0000 UTC m=+21.150813328" Mar 2 13:19:15.263661 kubelet[2746]: E0302 13:19:15.263506 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:15.887401 systemd-networkd[1252]: cilium_host: Link UP Mar 2 13:19:15.887685 systemd-networkd[1252]: cilium_net: Link UP Mar 2 13:19:15.888498 systemd-networkd[1252]: cilium_net: Gained carrier Mar 2 13:19:15.888872 systemd-networkd[1252]: cilium_host: Gained carrier Mar 2 13:19:15.889134 systemd-networkd[1252]: cilium_net: Gained IPv6LL Mar 2 13:19:15.889698 systemd-networkd[1252]: cilium_host: Gained IPv6LL Mar 2 13:19:16.026692 systemd-networkd[1252]: cilium_vxlan: Link UP Mar 2 13:19:16.026703 systemd-networkd[1252]: cilium_vxlan: Gained carrier Mar 2 13:19:16.264962 kubelet[2746]: E0302 13:19:16.264894 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:16.332294 kernel: NET: Registered PF_ALG protocol family Mar 2 13:19:17.219393 systemd-networkd[1252]: lxc_health: Link UP Mar 2 13:19:17.232064 systemd-networkd[1252]: lxc_health: Gained carrier Mar 2 13:19:17.683154 systemd-networkd[1252]: lxc3806c2770428: Link UP Mar 2 13:19:17.693251 kernel: eth0: renamed from tmpbbac9 Mar 2 13:19:17.701111 systemd-networkd[1252]: lxc3806c2770428: Gained carrier Mar 2 13:19:17.704364 systemd-networkd[1252]: lxc921c49d29f19: Link UP Mar 2 13:19:17.716254 kernel: eth0: renamed from tmpd34e9 Mar 2 13:19:17.722332 systemd-networkd[1252]: lxc921c49d29f19: Gained carrier Mar 2 13:19:17.937351 systemd-networkd[1252]: cilium_vxlan: Gained IPv6LL Mar 2 13:19:18.692259 kubelet[2746]: E0302 13:19:18.691996 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:18.768468 systemd-networkd[1252]: lxc_health: Gained IPv6LL Mar 2 13:19:19.024497 systemd-networkd[1252]: lxc921c49d29f19: Gained IPv6LL Mar 2 13:19:19.272254 kubelet[2746]: E0302 13:19:19.272109 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:19.600548 systemd-networkd[1252]: lxc3806c2770428: Gained IPv6LL Mar 2 13:19:20.274938 kubelet[2746]: E0302 13:19:20.274628 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:21.985247 containerd[1604]: time="2026-03-02T13:19:21.982745446Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:19:21.985247 containerd[1604]: time="2026-03-02T13:19:21.982865770Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:19:21.985247 containerd[1604]: time="2026-03-02T13:19:21.982896538Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:19:21.985247 containerd[1604]: time="2026-03-02T13:19:21.983025990Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:19:22.025359 containerd[1604]: time="2026-03-02T13:19:22.025053886Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:19:22.025359 containerd[1604]: time="2026-03-02T13:19:22.025131421Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:19:22.025359 containerd[1604]: time="2026-03-02T13:19:22.025149816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:19:22.026634 systemd-resolved[1469]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 13:19:22.028678 containerd[1604]: time="2026-03-02T13:19:22.027288997Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:19:22.067768 systemd-resolved[1469]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 2 13:19:22.086702 containerd[1604]: time="2026-03-02T13:19:22.086669713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-mrscv,Uid:5d8e6175-0857-440d-822b-eaaf28757b1a,Namespace:kube-system,Attempt:0,} returns sandbox id \"bbac90772c9c3a91f09712a5c36b061f57434c783470e74a53fbc62439f5a0dd\"" Mar 2 13:19:22.088755 kubelet[2746]: E0302 13:19:22.088733 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:22.100401 containerd[1604]: time="2026-03-02T13:19:22.100273416Z" level=info msg="CreateContainer within sandbox \"bbac90772c9c3a91f09712a5c36b061f57434c783470e74a53fbc62439f5a0dd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 13:19:22.111338 containerd[1604]: time="2026-03-02T13:19:22.111283280Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-kvlf5,Uid:2a5ccb95-fe38-475a-9921-a7cf6e289400,Namespace:kube-system,Attempt:0,} returns sandbox id \"d34e92b2f55669ab1d006d6a20d3b68e8eb7f2b30593f21b87d37fa635ffd367\"" Mar 2 13:19:22.112484 kubelet[2746]: E0302 13:19:22.112412 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:22.121777 containerd[1604]: time="2026-03-02T13:19:22.121706040Z" level=info msg="CreateContainer within sandbox \"d34e92b2f55669ab1d006d6a20d3b68e8eb7f2b30593f21b87d37fa635ffd367\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 13:19:22.139546 containerd[1604]: time="2026-03-02T13:19:22.139428918Z" level=info msg="CreateContainer within sandbox \"bbac90772c9c3a91f09712a5c36b061f57434c783470e74a53fbc62439f5a0dd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"2895e3251147fa6a1e3a5a359e53cc30ca8bfd9657f959737b6daf9d16cb1d51\"" Mar 2 13:19:22.142381 containerd[1604]: time="2026-03-02T13:19:22.142292720Z" level=info msg="StartContainer for \"2895e3251147fa6a1e3a5a359e53cc30ca8bfd9657f959737b6daf9d16cb1d51\"" Mar 2 13:19:22.148836 containerd[1604]: time="2026-03-02T13:19:22.148732234Z" level=info msg="CreateContainer within sandbox \"d34e92b2f55669ab1d006d6a20d3b68e8eb7f2b30593f21b87d37fa635ffd367\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a3beb325088fd48382e50234d6e5d0bdec07d16a87b9bc541ab593266fd88b3a\"" Mar 2 13:19:22.150532 containerd[1604]: time="2026-03-02T13:19:22.150509667Z" level=info msg="StartContainer for \"a3beb325088fd48382e50234d6e5d0bdec07d16a87b9bc541ab593266fd88b3a\"" Mar 2 13:19:22.216257 containerd[1604]: time="2026-03-02T13:19:22.216126748Z" level=info msg="StartContainer for \"2895e3251147fa6a1e3a5a359e53cc30ca8bfd9657f959737b6daf9d16cb1d51\" returns successfully" Mar 2 13:19:22.237131 containerd[1604]: time="2026-03-02T13:19:22.235868068Z" level=info msg="StartContainer for \"a3beb325088fd48382e50234d6e5d0bdec07d16a87b9bc541ab593266fd88b3a\" returns successfully" Mar 2 13:19:22.287140 kubelet[2746]: E0302 13:19:22.287025 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:22.293615 kubelet[2746]: E0302 13:19:22.291516 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:22.343460 kubelet[2746]: I0302 13:19:22.343342 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-kvlf5" podStartSLOduration=24.343319804 podStartE2EDuration="24.343319804s" podCreationTimestamp="2026-03-02 13:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:19:22.315508524 +0000 UTC m=+29.170620199" watchObservedRunningTime="2026-03-02 13:19:22.343319804 +0000 UTC m=+29.198431469" Mar 2 13:19:22.995121 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4056305481.mount: Deactivated successfully. Mar 2 13:19:23.295530 kubelet[2746]: E0302 13:19:23.294921 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:23.296764 kubelet[2746]: E0302 13:19:23.296120 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:23.315906 kubelet[2746]: I0302 13:19:23.315672 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-mrscv" podStartSLOduration=25.315647095 podStartE2EDuration="25.315647095s" podCreationTimestamp="2026-03-02 13:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:19:22.345484206 +0000 UTC m=+29.200595891" watchObservedRunningTime="2026-03-02 13:19:23.315647095 +0000 UTC m=+30.170758760" Mar 2 13:19:24.297444 kubelet[2746]: E0302 13:19:24.297106 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:24.298055 kubelet[2746]: E0302 13:19:24.297769 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:19:46.219583 systemd[1]: Started sshd@7-10.0.0.160:22-10.0.0.1:54546.service - OpenSSH per-connection server daemon (10.0.0.1:54546). Mar 2 13:19:46.260969 sshd[4151]: Accepted publickey for core from 10.0.0.1 port 54546 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:19:46.263472 sshd[4151]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:19:46.268453 systemd-logind[1577]: New session 8 of user core. Mar 2 13:19:46.278509 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 2 13:19:46.447718 sshd[4151]: pam_unix(sshd:session): session closed for user core Mar 2 13:19:46.452108 systemd[1]: sshd@7-10.0.0.160:22-10.0.0.1:54546.service: Deactivated successfully. Mar 2 13:19:46.455006 systemd-logind[1577]: Session 8 logged out. Waiting for processes to exit. Mar 2 13:19:46.455080 systemd[1]: session-8.scope: Deactivated successfully. Mar 2 13:19:46.456709 systemd-logind[1577]: Removed session 8. Mar 2 13:19:51.466482 systemd[1]: Started sshd@8-10.0.0.160:22-10.0.0.1:54554.service - OpenSSH per-connection server daemon (10.0.0.1:54554). Mar 2 13:19:51.500842 sshd[4168]: Accepted publickey for core from 10.0.0.1 port 54554 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:19:51.502688 sshd[4168]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:19:51.509067 systemd-logind[1577]: New session 9 of user core. Mar 2 13:19:51.522471 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 2 13:19:51.666610 sshd[4168]: pam_unix(sshd:session): session closed for user core Mar 2 13:19:51.671505 systemd[1]: sshd@8-10.0.0.160:22-10.0.0.1:54554.service: Deactivated successfully. Mar 2 13:19:51.674878 systemd-logind[1577]: Session 9 logged out. Waiting for processes to exit. Mar 2 13:19:51.674967 systemd[1]: session-9.scope: Deactivated successfully. Mar 2 13:19:51.677114 systemd-logind[1577]: Removed session 9. Mar 2 13:19:56.679528 systemd[1]: Started sshd@9-10.0.0.160:22-10.0.0.1:53642.service - OpenSSH per-connection server daemon (10.0.0.1:53642). Mar 2 13:19:56.717704 sshd[4187]: Accepted publickey for core from 10.0.0.1 port 53642 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:19:56.719600 sshd[4187]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:19:56.725366 systemd-logind[1577]: New session 10 of user core. Mar 2 13:19:56.743543 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 2 13:19:56.884222 sshd[4187]: pam_unix(sshd:session): session closed for user core Mar 2 13:19:56.889448 systemd[1]: sshd@9-10.0.0.160:22-10.0.0.1:53642.service: Deactivated successfully. Mar 2 13:19:56.893252 systemd-logind[1577]: Session 10 logged out. Waiting for processes to exit. Mar 2 13:19:56.894161 systemd[1]: session-10.scope: Deactivated successfully. Mar 2 13:19:56.896052 systemd-logind[1577]: Removed session 10. Mar 2 13:20:01.902617 systemd[1]: Started sshd@10-10.0.0.160:22-10.0.0.1:53656.service - OpenSSH per-connection server daemon (10.0.0.1:53656). Mar 2 13:20:01.957128 sshd[4206]: Accepted publickey for core from 10.0.0.1 port 53656 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:01.959109 sshd[4206]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:01.966067 systemd-logind[1577]: New session 11 of user core. Mar 2 13:20:01.978731 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 2 13:20:02.123626 sshd[4206]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:02.129077 systemd[1]: sshd@10-10.0.0.160:22-10.0.0.1:53656.service: Deactivated successfully. Mar 2 13:20:02.132753 systemd[1]: session-11.scope: Deactivated successfully. Mar 2 13:20:02.132774 systemd-logind[1577]: Session 11 logged out. Waiting for processes to exit. Mar 2 13:20:02.139051 systemd-logind[1577]: Removed session 11. Mar 2 13:20:07.142515 systemd[1]: Started sshd@11-10.0.0.160:22-10.0.0.1:39666.service - OpenSSH per-connection server daemon (10.0.0.1:39666). Mar 2 13:20:07.182245 sshd[4223]: Accepted publickey for core from 10.0.0.1 port 39666 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:07.184717 sshd[4223]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:07.192949 systemd-logind[1577]: New session 12 of user core. Mar 2 13:20:07.199586 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 2 13:20:07.351710 sshd[4223]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:07.356900 systemd[1]: sshd@11-10.0.0.160:22-10.0.0.1:39666.service: Deactivated successfully. Mar 2 13:20:07.360227 systemd-logind[1577]: Session 12 logged out. Waiting for processes to exit. Mar 2 13:20:07.360323 systemd[1]: session-12.scope: Deactivated successfully. Mar 2 13:20:07.362802 systemd-logind[1577]: Removed session 12. Mar 2 13:20:12.365462 systemd[1]: Started sshd@12-10.0.0.160:22-10.0.0.1:49214.service - OpenSSH per-connection server daemon (10.0.0.1:49214). Mar 2 13:20:12.399664 sshd[4240]: Accepted publickey for core from 10.0.0.1 port 49214 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:12.401893 sshd[4240]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:12.408421 systemd-logind[1577]: New session 13 of user core. Mar 2 13:20:12.419627 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 2 13:20:12.563608 sshd[4240]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:12.567625 systemd[1]: sshd@12-10.0.0.160:22-10.0.0.1:49214.service: Deactivated successfully. Mar 2 13:20:12.572114 systemd-logind[1577]: Session 13 logged out. Waiting for processes to exit. Mar 2 13:20:12.572482 systemd[1]: session-13.scope: Deactivated successfully. Mar 2 13:20:12.573940 systemd-logind[1577]: Removed session 13. Mar 2 13:20:17.578955 systemd[1]: Started sshd@13-10.0.0.160:22-10.0.0.1:49218.service - OpenSSH per-connection server daemon (10.0.0.1:49218). Mar 2 13:20:17.625536 sshd[4257]: Accepted publickey for core from 10.0.0.1 port 49218 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:17.638717 sshd[4257]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:17.651925 systemd-logind[1577]: New session 14 of user core. Mar 2 13:20:17.667702 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 2 13:20:17.851746 sshd[4257]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:17.856107 systemd[1]: sshd@13-10.0.0.160:22-10.0.0.1:49218.service: Deactivated successfully. Mar 2 13:20:17.858668 systemd[1]: session-14.scope: Deactivated successfully. Mar 2 13:20:17.858970 systemd-logind[1577]: Session 14 logged out. Waiting for processes to exit. Mar 2 13:20:17.860877 systemd-logind[1577]: Removed session 14. Mar 2 13:20:22.255445 kubelet[2746]: E0302 13:20:22.255325 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:22.862442 systemd[1]: Started sshd@14-10.0.0.160:22-10.0.0.1:54204.service - OpenSSH per-connection server daemon (10.0.0.1:54204). Mar 2 13:20:22.898339 sshd[4273]: Accepted publickey for core from 10.0.0.1 port 54204 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:22.900159 sshd[4273]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:22.905918 systemd-logind[1577]: New session 15 of user core. Mar 2 13:20:22.909523 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 2 13:20:23.050300 sshd[4273]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:23.054059 systemd[1]: sshd@14-10.0.0.160:22-10.0.0.1:54204.service: Deactivated successfully. Mar 2 13:20:23.058795 systemd[1]: session-15.scope: Deactivated successfully. Mar 2 13:20:23.059251 systemd-logind[1577]: Session 15 logged out. Waiting for processes to exit. Mar 2 13:20:23.061379 systemd-logind[1577]: Removed session 15. Mar 2 13:20:25.666460 kubelet[2746]: E0302 13:20:25.666124 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:28.071500 systemd[1]: Started sshd@15-10.0.0.160:22-10.0.0.1:54208.service - OpenSSH per-connection server daemon (10.0.0.1:54208). Mar 2 13:20:28.129491 sshd[4289]: Accepted publickey for core from 10.0.0.1 port 54208 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:28.131325 sshd[4289]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:28.165222 systemd-logind[1577]: New session 16 of user core. Mar 2 13:20:28.172556 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 2 13:20:28.349439 sshd[4289]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:28.354488 systemd[1]: sshd@15-10.0.0.160:22-10.0.0.1:54208.service: Deactivated successfully. Mar 2 13:20:28.358720 systemd[1]: session-16.scope: Deactivated successfully. Mar 2 13:20:28.361985 systemd-logind[1577]: Session 16 logged out. Waiting for processes to exit. Mar 2 13:20:28.363979 systemd-logind[1577]: Removed session 16. Mar 2 13:20:29.258975 kubelet[2746]: E0302 13:20:29.257692 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:33.362503 systemd[1]: Started sshd@16-10.0.0.160:22-10.0.0.1:47570.service - OpenSSH per-connection server daemon (10.0.0.1:47570). Mar 2 13:20:33.403762 sshd[4307]: Accepted publickey for core from 10.0.0.1 port 47570 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:33.405957 sshd[4307]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:33.414293 systemd-logind[1577]: New session 17 of user core. Mar 2 13:20:33.419560 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 2 13:20:33.575702 sshd[4307]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:33.589606 systemd[1]: Started sshd@17-10.0.0.160:22-10.0.0.1:47584.service - OpenSSH per-connection server daemon (10.0.0.1:47584). Mar 2 13:20:33.590731 systemd[1]: sshd@16-10.0.0.160:22-10.0.0.1:47570.service: Deactivated successfully. Mar 2 13:20:33.596884 systemd[1]: session-17.scope: Deactivated successfully. Mar 2 13:20:33.599746 systemd-logind[1577]: Session 17 logged out. Waiting for processes to exit. Mar 2 13:20:33.602347 systemd-logind[1577]: Removed session 17. Mar 2 13:20:33.630422 sshd[4320]: Accepted publickey for core from 10.0.0.1 port 47584 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:33.632698 sshd[4320]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:33.640101 systemd-logind[1577]: New session 18 of user core. Mar 2 13:20:33.650587 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 2 13:20:33.948868 sshd[4320]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:33.958858 systemd[1]: Started sshd@18-10.0.0.160:22-10.0.0.1:47592.service - OpenSSH per-connection server daemon (10.0.0.1:47592). Mar 2 13:20:33.959637 systemd[1]: sshd@17-10.0.0.160:22-10.0.0.1:47584.service: Deactivated successfully. Mar 2 13:20:33.970252 systemd-logind[1577]: Session 18 logged out. Waiting for processes to exit. Mar 2 13:20:33.971371 systemd[1]: session-18.scope: Deactivated successfully. Mar 2 13:20:33.976734 systemd-logind[1577]: Removed session 18. Mar 2 13:20:34.026074 sshd[4333]: Accepted publickey for core from 10.0.0.1 port 47592 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:34.028714 sshd[4333]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:34.059667 systemd-logind[1577]: New session 19 of user core. Mar 2 13:20:34.078670 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 2 13:20:34.270620 sshd[4333]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:34.277328 systemd[1]: sshd@18-10.0.0.160:22-10.0.0.1:47592.service: Deactivated successfully. Mar 2 13:20:34.280567 systemd-logind[1577]: Session 19 logged out. Waiting for processes to exit. Mar 2 13:20:34.280641 systemd[1]: session-19.scope: Deactivated successfully. Mar 2 13:20:34.282789 systemd-logind[1577]: Removed session 19. Mar 2 13:20:35.255421 kubelet[2746]: E0302 13:20:35.255261 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:39.282526 systemd[1]: Started sshd@19-10.0.0.160:22-10.0.0.1:47602.service - OpenSSH per-connection server daemon (10.0.0.1:47602). Mar 2 13:20:39.321145 sshd[4351]: Accepted publickey for core from 10.0.0.1 port 47602 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:39.323573 sshd[4351]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:39.330348 systemd-logind[1577]: New session 20 of user core. Mar 2 13:20:39.342703 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 2 13:20:39.493793 sshd[4351]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:39.497934 systemd[1]: sshd@19-10.0.0.160:22-10.0.0.1:47602.service: Deactivated successfully. Mar 2 13:20:39.502613 systemd-logind[1577]: Session 20 logged out. Waiting for processes to exit. Mar 2 13:20:39.502947 systemd[1]: session-20.scope: Deactivated successfully. Mar 2 13:20:39.505114 systemd-logind[1577]: Removed session 20. Mar 2 13:20:40.256329 kubelet[2746]: E0302 13:20:40.256154 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:42.256034 kubelet[2746]: E0302 13:20:42.255943 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:44.508435 systemd[1]: Started sshd@20-10.0.0.160:22-10.0.0.1:50428.service - OpenSSH per-connection server daemon (10.0.0.1:50428). Mar 2 13:20:44.544540 sshd[4367]: Accepted publickey for core from 10.0.0.1 port 50428 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:44.546540 sshd[4367]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:44.552525 systemd-logind[1577]: New session 21 of user core. Mar 2 13:20:44.559618 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 2 13:20:44.682371 sshd[4367]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:44.686345 systemd[1]: sshd@20-10.0.0.160:22-10.0.0.1:50428.service: Deactivated successfully. Mar 2 13:20:44.689748 systemd-logind[1577]: Session 21 logged out. Waiting for processes to exit. Mar 2 13:20:44.689750 systemd[1]: session-21.scope: Deactivated successfully. Mar 2 13:20:44.691240 systemd-logind[1577]: Removed session 21. Mar 2 13:20:46.255966 kubelet[2746]: E0302 13:20:46.255879 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:49.706543 systemd[1]: Started sshd@21-10.0.0.160:22-10.0.0.1:50432.service - OpenSSH per-connection server daemon (10.0.0.1:50432). Mar 2 13:20:49.742545 sshd[4384]: Accepted publickey for core from 10.0.0.1 port 50432 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:49.744287 sshd[4384]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:49.749442 systemd-logind[1577]: New session 22 of user core. Mar 2 13:20:49.763594 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 2 13:20:49.892519 sshd[4384]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:49.907432 systemd[1]: Started sshd@22-10.0.0.160:22-10.0.0.1:50448.service - OpenSSH per-connection server daemon (10.0.0.1:50448). Mar 2 13:20:49.908233 systemd[1]: sshd@21-10.0.0.160:22-10.0.0.1:50432.service: Deactivated successfully. Mar 2 13:20:49.911679 systemd-logind[1577]: Session 22 logged out. Waiting for processes to exit. Mar 2 13:20:49.912589 systemd[1]: session-22.scope: Deactivated successfully. Mar 2 13:20:49.914265 systemd-logind[1577]: Removed session 22. Mar 2 13:20:49.942586 sshd[4396]: Accepted publickey for core from 10.0.0.1 port 50448 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:49.944213 sshd[4396]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:49.949450 systemd-logind[1577]: New session 23 of user core. Mar 2 13:20:49.959556 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 2 13:20:50.230408 sshd[4396]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:50.238553 systemd[1]: Started sshd@23-10.0.0.160:22-10.0.0.1:50464.service - OpenSSH per-connection server daemon (10.0.0.1:50464). Mar 2 13:20:50.239341 systemd[1]: sshd@22-10.0.0.160:22-10.0.0.1:50448.service: Deactivated successfully. Mar 2 13:20:50.243386 systemd-logind[1577]: Session 23 logged out. Waiting for processes to exit. Mar 2 13:20:50.244689 systemd[1]: session-23.scope: Deactivated successfully. Mar 2 13:20:50.247646 systemd-logind[1577]: Removed session 23. Mar 2 13:20:50.280090 sshd[4410]: Accepted publickey for core from 10.0.0.1 port 50464 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:50.282331 sshd[4410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:50.288073 systemd-logind[1577]: New session 24 of user core. Mar 2 13:20:50.293503 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 2 13:20:51.122475 sshd[4410]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:51.137600 systemd[1]: Started sshd@24-10.0.0.160:22-10.0.0.1:50480.service - OpenSSH per-connection server daemon (10.0.0.1:50480). Mar 2 13:20:51.152936 systemd[1]: sshd@23-10.0.0.160:22-10.0.0.1:50464.service: Deactivated successfully. Mar 2 13:20:51.167825 systemd[1]: session-24.scope: Deactivated successfully. Mar 2 13:20:51.169039 systemd-logind[1577]: Session 24 logged out. Waiting for processes to exit. Mar 2 13:20:51.170643 systemd-logind[1577]: Removed session 24. Mar 2 13:20:51.198679 sshd[4430]: Accepted publickey for core from 10.0.0.1 port 50480 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:51.201056 sshd[4430]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:51.209003 systemd-logind[1577]: New session 25 of user core. Mar 2 13:20:51.224098 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 2 13:20:51.583448 sshd[4430]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:51.591669 systemd[1]: Started sshd@25-10.0.0.160:22-10.0.0.1:50486.service - OpenSSH per-connection server daemon (10.0.0.1:50486). Mar 2 13:20:51.592515 systemd[1]: sshd@24-10.0.0.160:22-10.0.0.1:50480.service: Deactivated successfully. Mar 2 13:20:51.601091 systemd[1]: session-25.scope: Deactivated successfully. Mar 2 13:20:51.602341 systemd-logind[1577]: Session 25 logged out. Waiting for processes to exit. Mar 2 13:20:51.605582 systemd-logind[1577]: Removed session 25. Mar 2 13:20:51.668123 sshd[4445]: Accepted publickey for core from 10.0.0.1 port 50486 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:51.670305 sshd[4445]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:51.677266 systemd-logind[1577]: New session 26 of user core. Mar 2 13:20:51.684719 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 2 13:20:51.822662 sshd[4445]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:51.828676 systemd[1]: sshd@25-10.0.0.160:22-10.0.0.1:50486.service: Deactivated successfully. Mar 2 13:20:51.831766 systemd[1]: session-26.scope: Deactivated successfully. Mar 2 13:20:51.831809 systemd-logind[1577]: Session 26 logged out. Waiting for processes to exit. Mar 2 13:20:51.834602 systemd-logind[1577]: Removed session 26. Mar 2 13:20:53.257870 kubelet[2746]: E0302 13:20:53.257749 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:20:56.834012 systemd[1]: Started sshd@26-10.0.0.160:22-10.0.0.1:45900.service - OpenSSH per-connection server daemon (10.0.0.1:45900). Mar 2 13:20:56.882227 sshd[4466]: Accepted publickey for core from 10.0.0.1 port 45900 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:20:56.884444 sshd[4466]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:20:56.890379 systemd-logind[1577]: New session 27 of user core. Mar 2 13:20:56.900741 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 2 13:20:57.033934 sshd[4466]: pam_unix(sshd:session): session closed for user core Mar 2 13:20:57.039556 systemd[1]: sshd@26-10.0.0.160:22-10.0.0.1:45900.service: Deactivated successfully. Mar 2 13:20:57.042708 systemd[1]: session-27.scope: Deactivated successfully. Mar 2 13:20:57.042759 systemd-logind[1577]: Session 27 logged out. Waiting for processes to exit. Mar 2 13:20:57.045056 systemd-logind[1577]: Removed session 27. Mar 2 13:21:02.052568 systemd[1]: Started sshd@27-10.0.0.160:22-10.0.0.1:59428.service - OpenSSH per-connection server daemon (10.0.0.1:59428). Mar 2 13:21:02.087359 sshd[4487]: Accepted publickey for core from 10.0.0.1 port 59428 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:02.089422 sshd[4487]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:02.095318 systemd-logind[1577]: New session 28 of user core. Mar 2 13:21:02.107609 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 2 13:21:02.232469 sshd[4487]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:02.237629 systemd[1]: sshd@27-10.0.0.160:22-10.0.0.1:59428.service: Deactivated successfully. Mar 2 13:21:02.240812 systemd[1]: session-28.scope: Deactivated successfully. Mar 2 13:21:02.241906 systemd-logind[1577]: Session 28 logged out. Waiting for processes to exit. Mar 2 13:21:02.243583 systemd-logind[1577]: Removed session 28. Mar 2 13:21:07.243625 systemd[1]: Started sshd@28-10.0.0.160:22-10.0.0.1:59434.service - OpenSSH per-connection server daemon (10.0.0.1:59434). Mar 2 13:21:07.281475 sshd[4503]: Accepted publickey for core from 10.0.0.1 port 59434 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:07.283567 sshd[4503]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:07.289659 systemd-logind[1577]: New session 29 of user core. Mar 2 13:21:07.296588 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 2 13:21:07.416615 sshd[4503]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:07.421782 systemd[1]: sshd@28-10.0.0.160:22-10.0.0.1:59434.service: Deactivated successfully. Mar 2 13:21:07.424832 systemd-logind[1577]: Session 29 logged out. Waiting for processes to exit. Mar 2 13:21:07.425018 systemd[1]: session-29.scope: Deactivated successfully. Mar 2 13:21:07.427038 systemd-logind[1577]: Removed session 29. Mar 2 13:21:12.435452 systemd[1]: Started sshd@29-10.0.0.160:22-10.0.0.1:60842.service - OpenSSH per-connection server daemon (10.0.0.1:60842). Mar 2 13:21:12.467958 sshd[4520]: Accepted publickey for core from 10.0.0.1 port 60842 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:12.469798 sshd[4520]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:12.475573 systemd-logind[1577]: New session 30 of user core. Mar 2 13:21:12.485571 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 2 13:21:12.614289 sshd[4520]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:12.621657 systemd[1]: Started sshd@30-10.0.0.160:22-10.0.0.1:60858.service - OpenSSH per-connection server daemon (10.0.0.1:60858). Mar 2 13:21:12.622277 systemd[1]: sshd@29-10.0.0.160:22-10.0.0.1:60842.service: Deactivated successfully. Mar 2 13:21:12.626651 systemd-logind[1577]: Session 30 logged out. Waiting for processes to exit. Mar 2 13:21:12.626783 systemd[1]: session-30.scope: Deactivated successfully. Mar 2 13:21:12.628421 systemd-logind[1577]: Removed session 30. Mar 2 13:21:12.657052 sshd[4532]: Accepted publickey for core from 10.0.0.1 port 60858 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:12.658793 sshd[4532]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:12.664926 systemd-logind[1577]: New session 31 of user core. Mar 2 13:21:12.674482 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 2 13:21:14.075460 containerd[1604]: time="2026-03-02T13:21:14.075386736Z" level=info msg="StopContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" with timeout 30 (s)" Mar 2 13:21:14.076130 containerd[1604]: time="2026-03-02T13:21:14.076069549Z" level=info msg="Stop container \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" with signal terminated" Mar 2 13:21:14.128219 containerd[1604]: time="2026-03-02T13:21:14.128048721Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 13:21:14.134311 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727-rootfs.mount: Deactivated successfully. Mar 2 13:21:14.137264 containerd[1604]: time="2026-03-02T13:21:14.137152037Z" level=info msg="StopContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" with timeout 2 (s)" Mar 2 13:21:14.137667 containerd[1604]: time="2026-03-02T13:21:14.137626313Z" level=info msg="Stop container \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" with signal terminated" Mar 2 13:21:14.147620 systemd-networkd[1252]: lxc_health: Link DOWN Mar 2 13:21:14.147638 systemd-networkd[1252]: lxc_health: Lost carrier Mar 2 13:21:14.159481 containerd[1604]: time="2026-03-02T13:21:14.159385371Z" level=info msg="shim disconnected" id=d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727 namespace=k8s.io Mar 2 13:21:14.159481 containerd[1604]: time="2026-03-02T13:21:14.159463196Z" level=warning msg="cleaning up after shim disconnected" id=d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727 namespace=k8s.io Mar 2 13:21:14.159481 containerd[1604]: time="2026-03-02T13:21:14.159480879Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:14.199430 containerd[1604]: time="2026-03-02T13:21:14.198557897Z" level=info msg="StopContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" returns successfully" Mar 2 13:21:14.205463 containerd[1604]: time="2026-03-02T13:21:14.205405314Z" level=info msg="StopPodSandbox for \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\"" Mar 2 13:21:14.205603 containerd[1604]: time="2026-03-02T13:21:14.205479462Z" level=info msg="Container to stop \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.208734 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af-rootfs.mount: Deactivated successfully. Mar 2 13:21:14.209002 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e-shm.mount: Deactivated successfully. Mar 2 13:21:14.222618 containerd[1604]: time="2026-03-02T13:21:14.222417021Z" level=info msg="shim disconnected" id=731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af namespace=k8s.io Mar 2 13:21:14.222618 containerd[1604]: time="2026-03-02T13:21:14.222467806Z" level=warning msg="cleaning up after shim disconnected" id=731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af namespace=k8s.io Mar 2 13:21:14.222618 containerd[1604]: time="2026-03-02T13:21:14.222483706Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:14.249065 containerd[1604]: time="2026-03-02T13:21:14.249000675Z" level=info msg="StopContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" returns successfully" Mar 2 13:21:14.250492 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e-rootfs.mount: Deactivated successfully. Mar 2 13:21:14.250647 containerd[1604]: time="2026-03-02T13:21:14.250543346Z" level=info msg="StopPodSandbox for \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\"" Mar 2 13:21:14.250647 containerd[1604]: time="2026-03-02T13:21:14.250587728Z" level=info msg="Container to stop \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.250647 containerd[1604]: time="2026-03-02T13:21:14.250608477Z" level=info msg="Container to stop \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.250647 containerd[1604]: time="2026-03-02T13:21:14.250621652Z" level=info msg="Container to stop \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.250647 containerd[1604]: time="2026-03-02T13:21:14.250638293Z" level=info msg="Container to stop \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.250893 containerd[1604]: time="2026-03-02T13:21:14.250652510Z" level=info msg="Container to stop \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 13:21:14.253514 containerd[1604]: time="2026-03-02T13:21:14.253404891Z" level=info msg="shim disconnected" id=902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e namespace=k8s.io Mar 2 13:21:14.253514 containerd[1604]: time="2026-03-02T13:21:14.253460295Z" level=warning msg="cleaning up after shim disconnected" id=902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e namespace=k8s.io Mar 2 13:21:14.253514 containerd[1604]: time="2026-03-02T13:21:14.253476705Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:14.274236 containerd[1604]: time="2026-03-02T13:21:14.274137888Z" level=info msg="TearDown network for sandbox \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\" successfully" Mar 2 13:21:14.274508 containerd[1604]: time="2026-03-02T13:21:14.274444069Z" level=info msg="StopPodSandbox for \"902713313e3db9ebe81142a9e0cea0cb799081a1f424cb097dad2f94c8358b0e\" returns successfully" Mar 2 13:21:14.288351 containerd[1604]: time="2026-03-02T13:21:14.288271305Z" level=info msg="shim disconnected" id=087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803 namespace=k8s.io Mar 2 13:21:14.288625 containerd[1604]: time="2026-03-02T13:21:14.288356474Z" level=warning msg="cleaning up after shim disconnected" id=087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803 namespace=k8s.io Mar 2 13:21:14.288625 containerd[1604]: time="2026-03-02T13:21:14.288372794Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:14.313036 containerd[1604]: time="2026-03-02T13:21:14.312953363Z" level=info msg="TearDown network for sandbox \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" successfully" Mar 2 13:21:14.313036 containerd[1604]: time="2026-03-02T13:21:14.312998727Z" level=info msg="StopPodSandbox for \"087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803\" returns successfully" Mar 2 13:21:14.322535 kubelet[2746]: I0302 13:21:14.322466 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-cilium-config-path\") pod \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\" (UID: \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\") " Mar 2 13:21:14.322984 kubelet[2746]: I0302 13:21:14.322538 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5mkt\" (UniqueName: \"kubernetes.io/projected/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-kube-api-access-v5mkt\") pod \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\" (UID: \"e6953e94-ca7c-4e4a-91a9-af1c54d6459d\") " Mar 2 13:21:14.329480 kubelet[2746]: I0302 13:21:14.328059 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "e6953e94-ca7c-4e4a-91a9-af1c54d6459d" (UID: "e6953e94-ca7c-4e4a-91a9-af1c54d6459d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 13:21:14.330514 kubelet[2746]: I0302 13:21:14.330444 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-kube-api-access-v5mkt" (OuterVolumeSpecName: "kube-api-access-v5mkt") pod "e6953e94-ca7c-4e4a-91a9-af1c54d6459d" (UID: "e6953e94-ca7c-4e4a-91a9-af1c54d6459d"). InnerVolumeSpecName "kube-api-access-v5mkt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 13:21:14.423129 kubelet[2746]: I0302 13:21:14.423064 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-xtables-lock\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423129 kubelet[2746]: I0302 13:21:14.423125 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ab2093cb-5670-4b75-883e-9e03aa6693fa-clustermesh-secrets\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423129 kubelet[2746]: I0302 13:21:14.423142 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-etc-cni-netd\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423129 kubelet[2746]: I0302 13:21:14.423138 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.423129 kubelet[2746]: I0302 13:21:14.423158 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd2l4\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-kube-api-access-wd2l4\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423211 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cni-path\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423227 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-kernel\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423233 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423242 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-hubble-tls\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423253 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-bpf-maps\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.423942 kubelet[2746]: I0302 13:21:14.423268 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-config-path\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423282 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-hostproc\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423295 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-cgroup\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423330 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-run\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423346 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-lib-modules\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423358 2746 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-net\") pod \"ab2093cb-5670-4b75-883e-9e03aa6693fa\" (UID: \"ab2093cb-5670-4b75-883e-9e03aa6693fa\") " Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423389 2746 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-xtables-lock\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.424074 kubelet[2746]: I0302 13:21:14.423398 2746 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.424270 kubelet[2746]: I0302 13:21:14.423408 2746 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.424270 kubelet[2746]: I0302 13:21:14.423417 2746 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-v5mkt\" (UniqueName: \"kubernetes.io/projected/e6953e94-ca7c-4e4a-91a9-af1c54d6459d-kube-api-access-v5mkt\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.424270 kubelet[2746]: I0302 13:21:14.423441 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.424270 kubelet[2746]: I0302 13:21:14.423511 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.424270 kubelet[2746]: I0302 13:21:14.423529 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.424376 kubelet[2746]: I0302 13:21:14.423546 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-hostproc" (OuterVolumeSpecName: "hostproc") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.424376 kubelet[2746]: I0302 13:21:14.423565 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cni-path" (OuterVolumeSpecName: "cni-path") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.424376 kubelet[2746]: I0302 13:21:14.423572 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.426236 kubelet[2746]: I0302 13:21:14.424974 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.426236 kubelet[2746]: I0302 13:21:14.425002 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 13:21:14.426796 kubelet[2746]: I0302 13:21:14.426775 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2093cb-5670-4b75-883e-9e03aa6693fa-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 13:21:14.427254 kubelet[2746]: I0302 13:21:14.426977 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 13:21:14.427997 kubelet[2746]: I0302 13:21:14.427965 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-kube-api-access-wd2l4" (OuterVolumeSpecName: "kube-api-access-wd2l4") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "kube-api-access-wd2l4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 13:21:14.428065 kubelet[2746]: I0302 13:21:14.428052 2746 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "ab2093cb-5670-4b75-883e-9e03aa6693fa" (UID: "ab2093cb-5670-4b75-883e-9e03aa6693fa"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 13:21:14.524733 kubelet[2746]: I0302 13:21:14.524654 2746 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-lib-modules\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.524733 kubelet[2746]: I0302 13:21:14.524706 2746 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.524733 kubelet[2746]: I0302 13:21:14.524722 2746 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ab2093cb-5670-4b75-883e-9e03aa6693fa-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.524733 kubelet[2746]: I0302 13:21:14.524734 2746 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wd2l4\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-kube-api-access-wd2l4\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.524733 kubelet[2746]: I0302 13:21:14.524748 2746 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cni-path\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524759 2746 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524771 2746 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ab2093cb-5670-4b75-883e-9e03aa6693fa-hubble-tls\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524781 2746 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-bpf-maps\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524792 2746 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524803 2746 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-hostproc\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524814 2746 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.525057 kubelet[2746]: I0302 13:21:14.524824 2746 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ab2093cb-5670-4b75-883e-9e03aa6693fa-cilium-run\") on node \"localhost\" DevicePath \"\"" Mar 2 13:21:14.863292 kubelet[2746]: I0302 13:21:14.863209 2746 scope.go:117] "RemoveContainer" containerID="731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af" Mar 2 13:21:14.866334 containerd[1604]: time="2026-03-02T13:21:14.866256432Z" level=info msg="RemoveContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\"" Mar 2 13:21:14.874780 containerd[1604]: time="2026-03-02T13:21:14.874657339Z" level=info msg="RemoveContainer for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" returns successfully" Mar 2 13:21:14.877619 kubelet[2746]: I0302 13:21:14.877581 2746 scope.go:117] "RemoveContainer" containerID="a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418" Mar 2 13:21:14.883549 containerd[1604]: time="2026-03-02T13:21:14.882118500Z" level=info msg="RemoveContainer for \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\"" Mar 2 13:21:14.886638 containerd[1604]: time="2026-03-02T13:21:14.886543528Z" level=info msg="RemoveContainer for \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\" returns successfully" Mar 2 13:21:14.886940 kubelet[2746]: I0302 13:21:14.886902 2746 scope.go:117] "RemoveContainer" containerID="35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6" Mar 2 13:21:14.888297 containerd[1604]: time="2026-03-02T13:21:14.888275762Z" level=info msg="RemoveContainer for \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\"" Mar 2 13:21:14.893072 containerd[1604]: time="2026-03-02T13:21:14.893008564Z" level=info msg="RemoveContainer for \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\" returns successfully" Mar 2 13:21:14.893343 kubelet[2746]: I0302 13:21:14.893278 2746 scope.go:117] "RemoveContainer" containerID="b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b" Mar 2 13:21:14.895357 containerd[1604]: time="2026-03-02T13:21:14.895280913Z" level=info msg="RemoveContainer for \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\"" Mar 2 13:21:14.899432 containerd[1604]: time="2026-03-02T13:21:14.899349674Z" level=info msg="RemoveContainer for \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\" returns successfully" Mar 2 13:21:14.899703 kubelet[2746]: I0302 13:21:14.899665 2746 scope.go:117] "RemoveContainer" containerID="6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0" Mar 2 13:21:14.901988 containerd[1604]: time="2026-03-02T13:21:14.901887151Z" level=info msg="RemoveContainer for \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\"" Mar 2 13:21:14.906682 containerd[1604]: time="2026-03-02T13:21:14.906582404Z" level=info msg="RemoveContainer for \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\" returns successfully" Mar 2 13:21:14.907071 kubelet[2746]: I0302 13:21:14.907023 2746 scope.go:117] "RemoveContainer" containerID="731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af" Mar 2 13:21:14.907635 containerd[1604]: time="2026-03-02T13:21:14.907558217Z" level=error msg="ContainerStatus for \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\": not found" Mar 2 13:21:14.916311 kubelet[2746]: E0302 13:21:14.916234 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\": not found" containerID="731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af" Mar 2 13:21:14.916407 kubelet[2746]: I0302 13:21:14.916297 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af"} err="failed to get container status \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\": rpc error: code = NotFound desc = an error occurred when try to find container \"731be32709eadba7c45ffedbdeffc948b955a78153883450f80af19711b524af\": not found" Mar 2 13:21:14.916407 kubelet[2746]: I0302 13:21:14.916343 2746 scope.go:117] "RemoveContainer" containerID="a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418" Mar 2 13:21:14.916764 containerd[1604]: time="2026-03-02T13:21:14.916692641Z" level=error msg="ContainerStatus for \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\": not found" Mar 2 13:21:14.917130 kubelet[2746]: E0302 13:21:14.916975 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\": not found" containerID="a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418" Mar 2 13:21:14.917130 kubelet[2746]: I0302 13:21:14.917008 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418"} err="failed to get container status \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\": rpc error: code = NotFound desc = an error occurred when try to find container \"a0570e59317c53bbbf541c2e062ce231cf3705e66ba567b60d73dc4980513418\": not found" Mar 2 13:21:14.917130 kubelet[2746]: I0302 13:21:14.917031 2746 scope.go:117] "RemoveContainer" containerID="35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6" Mar 2 13:21:14.917383 containerd[1604]: time="2026-03-02T13:21:14.917323249Z" level=error msg="ContainerStatus for \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\": not found" Mar 2 13:21:14.917516 kubelet[2746]: E0302 13:21:14.917470 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\": not found" containerID="35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6" Mar 2 13:21:14.917516 kubelet[2746]: I0302 13:21:14.917492 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6"} err="failed to get container status \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\": rpc error: code = NotFound desc = an error occurred when try to find container \"35fe92d6445f8ba36bb5b1d7cc4501d3a82407977e43ea60234f9ea3f2037dd6\": not found" Mar 2 13:21:14.917516 kubelet[2746]: I0302 13:21:14.917510 2746 scope.go:117] "RemoveContainer" containerID="b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b" Mar 2 13:21:14.917785 containerd[1604]: time="2026-03-02T13:21:14.917725076Z" level=error msg="ContainerStatus for \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\": not found" Mar 2 13:21:14.917906 kubelet[2746]: E0302 13:21:14.917828 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\": not found" containerID="b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b" Mar 2 13:21:14.917906 kubelet[2746]: I0302 13:21:14.917857 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b"} err="failed to get container status \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\": rpc error: code = NotFound desc = an error occurred when try to find container \"b4a2afb34e06b9e1f3fe2737e9e83cc12cca37489591a1cbe65e13eb36375a8b\": not found" Mar 2 13:21:14.917988 kubelet[2746]: I0302 13:21:14.917914 2746 scope.go:117] "RemoveContainer" containerID="6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0" Mar 2 13:21:14.918265 containerd[1604]: time="2026-03-02T13:21:14.918156362Z" level=error msg="ContainerStatus for \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\": not found" Mar 2 13:21:14.918484 kubelet[2746]: E0302 13:21:14.918449 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\": not found" containerID="6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0" Mar 2 13:21:14.918580 kubelet[2746]: I0302 13:21:14.918490 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0"} err="failed to get container status \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\": rpc error: code = NotFound desc = an error occurred when try to find container \"6d8271c3fdadc753e83adb3f55b99f60e1b045f2c7d4da2adb95de7285fd7cf0\": not found" Mar 2 13:21:14.918580 kubelet[2746]: I0302 13:21:14.918509 2746 scope.go:117] "RemoveContainer" containerID="d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727" Mar 2 13:21:14.919791 containerd[1604]: time="2026-03-02T13:21:14.919738135Z" level=info msg="RemoveContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\"" Mar 2 13:21:14.923764 containerd[1604]: time="2026-03-02T13:21:14.923677891Z" level=info msg="RemoveContainer for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" returns successfully" Mar 2 13:21:14.923896 kubelet[2746]: I0302 13:21:14.923834 2746 scope.go:117] "RemoveContainer" containerID="d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727" Mar 2 13:21:14.924160 containerd[1604]: time="2026-03-02T13:21:14.924071484Z" level=error msg="ContainerStatus for \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\": not found" Mar 2 13:21:14.924322 kubelet[2746]: E0302 13:21:14.924270 2746 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\": not found" containerID="d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727" Mar 2 13:21:14.924322 kubelet[2746]: I0302 13:21:14.924290 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727"} err="failed to get container status \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\": rpc error: code = NotFound desc = an error occurred when try to find container \"d7073a8cd2f5a6b967908006674b6df47dfe9d17e04fc57ad39c1436770c4727\": not found" Mar 2 13:21:15.099398 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803-rootfs.mount: Deactivated successfully. Mar 2 13:21:15.099589 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-087ed67617e011c88aa5a7c81e32977ce5c861dee8ad5d2287e10d5c29e45803-shm.mount: Deactivated successfully. Mar 2 13:21:15.099725 systemd[1]: var-lib-kubelet-pods-e6953e94\x2dca7c\x2d4e4a\x2d91a9\x2daf1c54d6459d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dv5mkt.mount: Deactivated successfully. Mar 2 13:21:15.099913 systemd[1]: var-lib-kubelet-pods-ab2093cb\x2d5670\x2d4b75\x2d883e\x2d9e03aa6693fa-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dwd2l4.mount: Deactivated successfully. Mar 2 13:21:15.100122 systemd[1]: var-lib-kubelet-pods-ab2093cb\x2d5670\x2d4b75\x2d883e\x2d9e03aa6693fa-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 2 13:21:15.100320 systemd[1]: var-lib-kubelet-pods-ab2093cb\x2d5670\x2d4b75\x2d883e\x2d9e03aa6693fa-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 2 13:21:15.258262 kubelet[2746]: I0302 13:21:15.258149 2746 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab2093cb-5670-4b75-883e-9e03aa6693fa" path="/var/lib/kubelet/pods/ab2093cb-5670-4b75-883e-9e03aa6693fa/volumes" Mar 2 13:21:15.259507 kubelet[2746]: I0302 13:21:15.259430 2746 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6953e94-ca7c-4e4a-91a9-af1c54d6459d" path="/var/lib/kubelet/pods/e6953e94-ca7c-4e4a-91a9-af1c54d6459d/volumes" Mar 2 13:21:15.991451 sshd[4532]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:16.000496 systemd[1]: Started sshd@31-10.0.0.160:22-10.0.0.1:60866.service - OpenSSH per-connection server daemon (10.0.0.1:60866). Mar 2 13:21:16.000986 systemd[1]: sshd@30-10.0.0.160:22-10.0.0.1:60858.service: Deactivated successfully. Mar 2 13:21:16.005127 systemd-logind[1577]: Session 31 logged out. Waiting for processes to exit. Mar 2 13:21:16.006334 systemd[1]: session-31.scope: Deactivated successfully. Mar 2 13:21:16.007529 systemd-logind[1577]: Removed session 31. Mar 2 13:21:16.032469 sshd[4701]: Accepted publickey for core from 10.0.0.1 port 60866 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:16.034058 sshd[4701]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:16.038995 systemd-logind[1577]: New session 32 of user core. Mar 2 13:21:16.056636 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 2 13:21:16.640253 sshd[4701]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:16.654585 systemd[1]: Started sshd@32-10.0.0.160:22-10.0.0.1:60876.service - OpenSSH per-connection server daemon (10.0.0.1:60876). Mar 2 13:21:16.658450 systemd[1]: sshd@31-10.0.0.160:22-10.0.0.1:60866.service: Deactivated successfully. Mar 2 13:21:16.666961 systemd[1]: session-32.scope: Deactivated successfully. Mar 2 13:21:16.670006 systemd-logind[1577]: Session 32 logged out. Waiting for processes to exit. Mar 2 13:21:16.675777 systemd-logind[1577]: Removed session 32. Mar 2 13:21:16.712932 sshd[4715]: Accepted publickey for core from 10.0.0.1 port 60876 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:16.715118 sshd[4715]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:16.722411 systemd-logind[1577]: New session 33 of user core. Mar 2 13:21:16.727702 systemd[1]: Started session-33.scope - Session 33 of User core. Mar 2 13:21:16.739757 kubelet[2746]: I0302 13:21:16.739679 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-cilium-run\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.739757 kubelet[2746]: I0302 13:21:16.739753 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-bpf-maps\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739820 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-xtables-lock\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739850 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7736fa06-1ab2-4121-8083-ef7b4ba60886-hubble-tls\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739917 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-cilium-cgroup\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739944 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-cni-path\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739967 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-etc-cni-netd\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740384 kubelet[2746]: I0302 13:21:16.739991 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7736fa06-1ab2-4121-8083-ef7b4ba60886-clustermesh-secrets\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740552 kubelet[2746]: I0302 13:21:16.740030 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/7736fa06-1ab2-4121-8083-ef7b4ba60886-cilium-ipsec-secrets\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740552 kubelet[2746]: I0302 13:21:16.740053 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-host-proc-sys-net\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740552 kubelet[2746]: I0302 13:21:16.740138 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8cc\" (UniqueName: \"kubernetes.io/projected/7736fa06-1ab2-4121-8083-ef7b4ba60886-kube-api-access-jc8cc\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740552 kubelet[2746]: I0302 13:21:16.740226 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-lib-modules\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740552 kubelet[2746]: I0302 13:21:16.740265 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-hostproc\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740699 kubelet[2746]: I0302 13:21:16.740288 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7736fa06-1ab2-4121-8083-ef7b4ba60886-host-proc-sys-kernel\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.740699 kubelet[2746]: I0302 13:21:16.740511 2746 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7736fa06-1ab2-4121-8083-ef7b4ba60886-cilium-config-path\") pod \"cilium-zj9pq\" (UID: \"7736fa06-1ab2-4121-8083-ef7b4ba60886\") " pod="kube-system/cilium-zj9pq" Mar 2 13:21:16.786119 sshd[4715]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:16.801453 systemd[1]: Started sshd@33-10.0.0.160:22-10.0.0.1:60882.service - OpenSSH per-connection server daemon (10.0.0.1:60882). Mar 2 13:21:16.802013 systemd[1]: sshd@32-10.0.0.160:22-10.0.0.1:60876.service: Deactivated successfully. Mar 2 13:21:16.805976 systemd[1]: session-33.scope: Deactivated successfully. Mar 2 13:21:16.806071 systemd-logind[1577]: Session 33 logged out. Waiting for processes to exit. Mar 2 13:21:16.808369 systemd-logind[1577]: Removed session 33. Mar 2 13:21:16.840287 sshd[4724]: Accepted publickey for core from 10.0.0.1 port 60882 ssh2: RSA SHA256:I7frh5Ho+GNZYlhwMF3Kg7xi/C+xdSmVTMEFrO7Zj60 Mar 2 13:21:16.842666 sshd[4724]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 13:21:16.867266 systemd-logind[1577]: New session 34 of user core. Mar 2 13:21:16.878641 systemd[1]: Started session-34.scope - Session 34 of User core. Mar 2 13:21:16.996283 kubelet[2746]: E0302 13:21:16.996030 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:16.997906 containerd[1604]: time="2026-03-02T13:21:16.997828385Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zj9pq,Uid:7736fa06-1ab2-4121-8083-ef7b4ba60886,Namespace:kube-system,Attempt:0,}" Mar 2 13:21:17.041956 containerd[1604]: time="2026-03-02T13:21:17.041641980Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 2 13:21:17.043115 containerd[1604]: time="2026-03-02T13:21:17.043041022Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 2 13:21:17.043233 containerd[1604]: time="2026-03-02T13:21:17.043107596Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:21:17.043401 containerd[1604]: time="2026-03-02T13:21:17.043328839Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 2 13:21:17.097517 containerd[1604]: time="2026-03-02T13:21:17.097461779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zj9pq,Uid:7736fa06-1ab2-4121-8083-ef7b4ba60886,Namespace:kube-system,Attempt:0,} returns sandbox id \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\"" Mar 2 13:21:17.098396 kubelet[2746]: E0302 13:21:17.098324 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:17.104568 containerd[1604]: time="2026-03-02T13:21:17.104538340Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 2 13:21:17.121258 containerd[1604]: time="2026-03-02T13:21:17.121126567Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5d07a8d9f23f84eb312e4eafd134079a604dbbc113f3d921f45a2e8cebfb146a\"" Mar 2 13:21:17.121979 containerd[1604]: time="2026-03-02T13:21:17.121932390Z" level=info msg="StartContainer for \"5d07a8d9f23f84eb312e4eafd134079a604dbbc113f3d921f45a2e8cebfb146a\"" Mar 2 13:21:17.262380 containerd[1604]: time="2026-03-02T13:21:17.261719625Z" level=info msg="StartContainer for \"5d07a8d9f23f84eb312e4eafd134079a604dbbc113f3d921f45a2e8cebfb146a\" returns successfully" Mar 2 13:21:17.394573 containerd[1604]: time="2026-03-02T13:21:17.394512255Z" level=info msg="shim disconnected" id=5d07a8d9f23f84eb312e4eafd134079a604dbbc113f3d921f45a2e8cebfb146a namespace=k8s.io Mar 2 13:21:17.394955 containerd[1604]: time="2026-03-02T13:21:17.394795624Z" level=warning msg="cleaning up after shim disconnected" id=5d07a8d9f23f84eb312e4eafd134079a604dbbc113f3d921f45a2e8cebfb146a namespace=k8s.io Mar 2 13:21:17.394955 containerd[1604]: time="2026-03-02T13:21:17.394837623Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:17.875577 kubelet[2746]: E0302 13:21:17.875504 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:17.882479 containerd[1604]: time="2026-03-02T13:21:17.882420919Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 2 13:21:17.905111 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3757492977.mount: Deactivated successfully. Mar 2 13:21:17.908377 containerd[1604]: time="2026-03-02T13:21:17.908268087Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0\"" Mar 2 13:21:17.909396 containerd[1604]: time="2026-03-02T13:21:17.909321015Z" level=info msg="StartContainer for \"e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0\"" Mar 2 13:21:17.981158 containerd[1604]: time="2026-03-02T13:21:17.981094296Z" level=info msg="StartContainer for \"e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0\" returns successfully" Mar 2 13:21:18.027590 containerd[1604]: time="2026-03-02T13:21:18.027449825Z" level=info msg="shim disconnected" id=e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0 namespace=k8s.io Mar 2 13:21:18.027590 containerd[1604]: time="2026-03-02T13:21:18.027527200Z" level=warning msg="cleaning up after shim disconnected" id=e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0 namespace=k8s.io Mar 2 13:21:18.027590 containerd[1604]: time="2026-03-02T13:21:18.027539383Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:18.464767 kubelet[2746]: E0302 13:21:18.464695 2746 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 2 13:21:18.849073 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e2169b5bf0fb045e297ddd2e79ac907f1035e22794b7f049b927bee0a9b325b0-rootfs.mount: Deactivated successfully. Mar 2 13:21:18.878935 kubelet[2746]: E0302 13:21:18.878823 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:18.887728 containerd[1604]: time="2026-03-02T13:21:18.887631064Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 2 13:21:18.905434 containerd[1604]: time="2026-03-02T13:21:18.905380660Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a\"" Mar 2 13:21:18.906229 containerd[1604]: time="2026-03-02T13:21:18.906122328Z" level=info msg="StartContainer for \"e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a\"" Mar 2 13:21:18.980731 containerd[1604]: time="2026-03-02T13:21:18.980638103Z" level=info msg="StartContainer for \"e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a\" returns successfully" Mar 2 13:21:19.009012 containerd[1604]: time="2026-03-02T13:21:19.008928737Z" level=info msg="shim disconnected" id=e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a namespace=k8s.io Mar 2 13:21:19.009012 containerd[1604]: time="2026-03-02T13:21:19.008984561Z" level=warning msg="cleaning up after shim disconnected" id=e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a namespace=k8s.io Mar 2 13:21:19.009012 containerd[1604]: time="2026-03-02T13:21:19.008993538Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:19.849742 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e9b57bbd38994c9dc0530c431ccfebf0d0558fb2a3a0e000de09763d7956f98a-rootfs.mount: Deactivated successfully. Mar 2 13:21:19.884944 kubelet[2746]: E0302 13:21:19.884779 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:19.891599 containerd[1604]: time="2026-03-02T13:21:19.891395825Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 2 13:21:19.908008 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount197859613.mount: Deactivated successfully. Mar 2 13:21:19.908162 containerd[1604]: time="2026-03-02T13:21:19.908098610Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79\"" Mar 2 13:21:19.909282 containerd[1604]: time="2026-03-02T13:21:19.909113936Z" level=info msg="StartContainer for \"a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79\"" Mar 2 13:21:19.987810 containerd[1604]: time="2026-03-02T13:21:19.987686180Z" level=info msg="StartContainer for \"a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79\" returns successfully" Mar 2 13:21:20.017435 containerd[1604]: time="2026-03-02T13:21:20.017377525Z" level=info msg="shim disconnected" id=a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79 namespace=k8s.io Mar 2 13:21:20.017435 containerd[1604]: time="2026-03-02T13:21:20.017434231Z" level=warning msg="cleaning up after shim disconnected" id=a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79 namespace=k8s.io Mar 2 13:21:20.017696 containerd[1604]: time="2026-03-02T13:21:20.017443628Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 13:21:20.850162 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a36cfb1d37da40409cefda6e38a9e17bc56e1eac90affc1033de278264bf4b79-rootfs.mount: Deactivated successfully. Mar 2 13:21:20.890583 kubelet[2746]: E0302 13:21:20.890522 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:20.896400 containerd[1604]: time="2026-03-02T13:21:20.896327176Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 2 13:21:20.912412 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3012300233.mount: Deactivated successfully. Mar 2 13:21:20.918252 containerd[1604]: time="2026-03-02T13:21:20.918133934Z" level=info msg="CreateContainer within sandbox \"0b68d8e189377ca57c1ffb4226d70940a393a853b9477cc398ad508449d7d23d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"522b5974539864cf126a1cd851560f9b13922b416ccc0c18bdcb6f1da594a0fd\"" Mar 2 13:21:20.919142 containerd[1604]: time="2026-03-02T13:21:20.919100310Z" level=info msg="StartContainer for \"522b5974539864cf126a1cd851560f9b13922b416ccc0c18bdcb6f1da594a0fd\"" Mar 2 13:21:20.998470 containerd[1604]: time="2026-03-02T13:21:20.998370682Z" level=info msg="StartContainer for \"522b5974539864cf126a1cd851560f9b13922b416ccc0c18bdcb6f1da594a0fd\" returns successfully" Mar 2 13:21:21.523273 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Mar 2 13:21:21.897123 kubelet[2746]: E0302 13:21:21.896926 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:21.922037 kubelet[2746]: I0302 13:21:21.921858 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-zj9pq" podStartSLOduration=5.921835136 podStartE2EDuration="5.921835136s" podCreationTimestamp="2026-03-02 13:21:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 13:21:21.92082673 +0000 UTC m=+148.775938406" watchObservedRunningTime="2026-03-02 13:21:21.921835136 +0000 UTC m=+148.776946821" Mar 2 13:21:22.998476 kubelet[2746]: E0302 13:21:22.998377 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:23.278011 systemd[1]: run-containerd-runc-k8s.io-522b5974539864cf126a1cd851560f9b13922b416ccc0c18bdcb6f1da594a0fd-runc.HuWTJr.mount: Deactivated successfully. Mar 2 13:21:25.271577 systemd-networkd[1252]: lxc_health: Link UP Mar 2 13:21:25.282280 systemd-networkd[1252]: lxc_health: Gained carrier Mar 2 13:21:25.541322 kubelet[2746]: E0302 13:21:25.539861 2746 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:36948->127.0.0.1:41159: write tcp 127.0.0.1:36948->127.0.0.1:41159: write: broken pipe Mar 2 13:21:26.320482 systemd-networkd[1252]: lxc_health: Gained IPv6LL Mar 2 13:21:26.999086 kubelet[2746]: E0302 13:21:26.998623 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:27.916663 kubelet[2746]: E0302 13:21:27.915963 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:28.918263 kubelet[2746]: E0302 13:21:28.918137 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 2 13:21:30.149310 systemd[1]: run-containerd-runc-k8s.io-522b5974539864cf126a1cd851560f9b13922b416ccc0c18bdcb6f1da594a0fd-runc.4ZBZxA.mount: Deactivated successfully. Mar 2 13:21:32.492760 sshd[4724]: pam_unix(sshd:session): session closed for user core Mar 2 13:21:32.496956 systemd[1]: sshd@33-10.0.0.160:22-10.0.0.1:60882.service: Deactivated successfully. Mar 2 13:21:32.500289 systemd-logind[1577]: Session 34 logged out. Waiting for processes to exit. Mar 2 13:21:32.501357 systemd[1]: session-34.scope: Deactivated successfully. Mar 2 13:21:32.502393 systemd-logind[1577]: Removed session 34.