May 8 03:06:10.871188 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] May 8 03:06:10.871209 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT Thu May 8 02:00:10 -00 2025 May 8 03:06:10.871218 kernel: KASLR enabled May 8 03:06:10.871224 kernel: efi: EFI v2.7 by EDK II May 8 03:06:10.871229 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 May 8 03:06:10.871234 kernel: random: crng init done May 8 03:06:10.871241 kernel: secureboot: Secure boot disabled May 8 03:06:10.871247 kernel: ACPI: Early table checksum verification disabled May 8 03:06:10.871252 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) May 8 03:06:10.871259 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) May 8 03:06:10.871265 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871271 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871276 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871282 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871289 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871296 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871302 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871308 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871314 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 8 03:06:10.871320 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 May 8 03:06:10.871326 kernel: NUMA: Failed to initialise from firmware May 8 03:06:10.871333 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] May 8 03:06:10.871339 kernel: NUMA: NODE_DATA [mem 0xdc957e00-0xdc95efff] May 8 03:06:10.871345 kernel: Zone ranges: May 8 03:06:10.871351 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] May 8 03:06:10.871358 kernel: DMA32 empty May 8 03:06:10.871379 kernel: Normal empty May 8 03:06:10.871385 kernel: Device empty May 8 03:06:10.871391 kernel: Movable zone start for each node May 8 03:06:10.871398 kernel: Early memory node ranges May 8 03:06:10.871404 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] May 8 03:06:10.871410 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] May 8 03:06:10.871416 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] May 8 03:06:10.871422 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] May 8 03:06:10.871428 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] May 8 03:06:10.871434 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] May 8 03:06:10.871440 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] May 8 03:06:10.871446 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] May 8 03:06:10.871454 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] May 8 03:06:10.871460 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] May 8 03:06:10.871468 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] May 8 03:06:10.871475 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] May 8 03:06:10.871481 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] May 8 03:06:10.871489 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] May 8 03:06:10.871496 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges May 8 03:06:10.871502 kernel: psci: probing for conduit method from ACPI. May 8 03:06:10.871508 kernel: psci: PSCIv1.1 detected in firmware. May 8 03:06:10.871515 kernel: psci: Using standard PSCI v0.2 function IDs May 8 03:06:10.871521 kernel: psci: Trusted OS migration not required May 8 03:06:10.871527 kernel: psci: SMC Calling Convention v1.1 May 8 03:06:10.871534 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) May 8 03:06:10.871540 kernel: percpu: Embedded 31 pages/cpu s87080 r8192 d31704 u126976 May 8 03:06:10.871546 kernel: pcpu-alloc: s87080 r8192 d31704 u126976 alloc=31*4096 May 8 03:06:10.871553 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 May 8 03:06:10.871560 kernel: Detected PIPT I-cache on CPU0 May 8 03:06:10.871567 kernel: CPU features: detected: GIC system register CPU interface May 8 03:06:10.871573 kernel: CPU features: detected: Hardware dirty bit management May 8 03:06:10.871579 kernel: CPU features: detected: Spectre-v4 May 8 03:06:10.871586 kernel: CPU features: detected: Spectre-BHB May 8 03:06:10.871592 kernel: CPU features: kernel page table isolation forced ON by KASLR May 8 03:06:10.871598 kernel: CPU features: detected: Kernel page table isolation (KPTI) May 8 03:06:10.871605 kernel: CPU features: detected: ARM erratum 1418040 May 8 03:06:10.871611 kernel: CPU features: detected: SSBS not fully self-synchronizing May 8 03:06:10.871617 kernel: alternatives: applying boot alternatives May 8 03:06:10.871625 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=b77ba75748ce73d160d869c11543eefaa814f1163d4a6d367aa8765f0183246a May 8 03:06:10.871632 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 8 03:06:10.871639 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 8 03:06:10.871645 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 8 03:06:10.871652 kernel: Fallback order for Node 0: 0 May 8 03:06:10.871658 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 May 8 03:06:10.871665 kernel: Policy zone: DMA May 8 03:06:10.871671 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 8 03:06:10.871677 kernel: software IO TLB: area num 4. May 8 03:06:10.871683 kernel: software IO TLB: mapped [mem 0x00000000d5000000-0x00000000d9000000] (64MB) May 8 03:06:10.871690 kernel: Memory: 2386508K/2572288K available (10432K kernel code, 2202K rwdata, 8168K rodata, 39040K init, 993K bss, 185780K reserved, 0K cma-reserved) May 8 03:06:10.871697 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 May 8 03:06:10.871704 kernel: rcu: Preemptible hierarchical RCU implementation. May 8 03:06:10.871711 kernel: rcu: RCU event tracing is enabled. May 8 03:06:10.871718 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. May 8 03:06:10.871724 kernel: Trampoline variant of Tasks RCU enabled. May 8 03:06:10.871730 kernel: Tracing variant of Tasks RCU enabled. May 8 03:06:10.871737 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 8 03:06:10.871743 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 May 8 03:06:10.871750 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 May 8 03:06:10.871756 kernel: GICv3: 256 SPIs implemented May 8 03:06:10.871762 kernel: GICv3: 0 Extended SPIs implemented May 8 03:06:10.871769 kernel: Root IRQ handler: gic_handle_irq May 8 03:06:10.871775 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI May 8 03:06:10.871783 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 May 8 03:06:10.871797 kernel: ITS [mem 0x08080000-0x0809ffff] May 8 03:06:10.871804 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400d0000 (indirect, esz 8, psz 64K, shr 1) May 8 03:06:10.871810 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400e0000 (flat, esz 8, psz 64K, shr 1) May 8 03:06:10.871817 kernel: GICv3: using LPI property table @0x00000000400f0000 May 8 03:06:10.871823 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000 May 8 03:06:10.871830 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 8 03:06:10.871836 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 8 03:06:10.871876 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). May 8 03:06:10.871885 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns May 8 03:06:10.871891 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns May 8 03:06:10.871901 kernel: arm-pv: using stolen time PV May 8 03:06:10.871907 kernel: Console: colour dummy device 80x25 May 8 03:06:10.871914 kernel: ACPI: Core revision 20230628 May 8 03:06:10.871921 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) May 8 03:06:10.871928 kernel: pid_max: default: 32768 minimum: 301 May 8 03:06:10.871934 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 8 03:06:10.871940 kernel: landlock: Up and running. May 8 03:06:10.871947 kernel: SELinux: Initializing. May 8 03:06:10.871953 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 03:06:10.871961 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 03:06:10.871968 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 8 03:06:10.871975 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 8 03:06:10.871981 kernel: rcu: Hierarchical SRCU implementation. May 8 03:06:10.871988 kernel: rcu: Max phase no-delay instances is 400. May 8 03:06:10.871994 kernel: Platform MSI: ITS@0x8080000 domain created May 8 03:06:10.872000 kernel: PCI/MSI: ITS@0x8080000 domain created May 8 03:06:10.872007 kernel: Remapping and enabling EFI services. May 8 03:06:10.872013 kernel: smp: Bringing up secondary CPUs ... May 8 03:06:10.872022 kernel: Detected PIPT I-cache on CPU1 May 8 03:06:10.872033 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 May 8 03:06:10.872040 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000 May 8 03:06:10.872048 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 8 03:06:10.872055 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] May 8 03:06:10.872062 kernel: Detected PIPT I-cache on CPU2 May 8 03:06:10.872069 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 May 8 03:06:10.872076 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000 May 8 03:06:10.872084 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 8 03:06:10.872090 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] May 8 03:06:10.872097 kernel: Detected PIPT I-cache on CPU3 May 8 03:06:10.872104 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 May 8 03:06:10.872111 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000 May 8 03:06:10.872118 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 8 03:06:10.872124 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] May 8 03:06:10.872131 kernel: smp: Brought up 1 node, 4 CPUs May 8 03:06:10.872138 kernel: SMP: Total of 4 processors activated. May 8 03:06:10.872145 kernel: CPU features: detected: 32-bit EL0 Support May 8 03:06:10.872153 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence May 8 03:06:10.872160 kernel: CPU features: detected: Common not Private translations May 8 03:06:10.872167 kernel: CPU features: detected: CRC32 instructions May 8 03:06:10.872174 kernel: CPU features: detected: Enhanced Virtualization Traps May 8 03:06:10.872181 kernel: CPU features: detected: RCpc load-acquire (LDAPR) May 8 03:06:10.872187 kernel: CPU features: detected: LSE atomic instructions May 8 03:06:10.872194 kernel: CPU features: detected: Privileged Access Never May 8 03:06:10.872201 kernel: CPU features: detected: RAS Extension Support May 8 03:06:10.872208 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) May 8 03:06:10.872216 kernel: CPU: All CPU(s) started at EL1 May 8 03:06:10.872223 kernel: alternatives: applying system-wide alternatives May 8 03:06:10.872230 kernel: devtmpfs: initialized May 8 03:06:10.872237 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 8 03:06:10.872243 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) May 8 03:06:10.872251 kernel: pinctrl core: initialized pinctrl subsystem May 8 03:06:10.872257 kernel: SMBIOS 3.0.0 present. May 8 03:06:10.872264 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 May 8 03:06:10.872271 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 8 03:06:10.872279 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations May 8 03:06:10.872286 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations May 8 03:06:10.872293 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations May 8 03:06:10.872300 kernel: audit: initializing netlink subsys (disabled) May 8 03:06:10.872307 kernel: audit: type=2000 audit(0.018:1): state=initialized audit_enabled=0 res=1 May 8 03:06:10.872314 kernel: thermal_sys: Registered thermal governor 'step_wise' May 8 03:06:10.872321 kernel: cpuidle: using governor menu May 8 03:06:10.872327 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. May 8 03:06:10.872334 kernel: ASID allocator initialised with 32768 entries May 8 03:06:10.872343 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 8 03:06:10.872350 kernel: Serial: AMBA PL011 UART driver May 8 03:06:10.872357 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL May 8 03:06:10.872363 kernel: Modules: 0 pages in range for non-PLT usage May 8 03:06:10.872370 kernel: Modules: 509024 pages in range for PLT usage May 8 03:06:10.872377 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 8 03:06:10.872384 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page May 8 03:06:10.872391 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages May 8 03:06:10.872398 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page May 8 03:06:10.872406 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 8 03:06:10.872413 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page May 8 03:06:10.872420 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages May 8 03:06:10.872427 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page May 8 03:06:10.872434 kernel: ACPI: Added _OSI(Module Device) May 8 03:06:10.872440 kernel: ACPI: Added _OSI(Processor Device) May 8 03:06:10.872447 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 8 03:06:10.872454 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 8 03:06:10.872461 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 8 03:06:10.872469 kernel: ACPI: Interpreter enabled May 8 03:06:10.872476 kernel: ACPI: Using GIC for interrupt routing May 8 03:06:10.872483 kernel: ACPI: MCFG table detected, 1 entries May 8 03:06:10.872490 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA May 8 03:06:10.872497 kernel: printk: console [ttyAMA0] enabled May 8 03:06:10.872503 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 8 03:06:10.872645 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 8 03:06:10.872716 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] May 8 03:06:10.872792 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] May 8 03:06:10.872879 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 May 8 03:06:10.872945 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] May 8 03:06:10.872955 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] May 8 03:06:10.872962 kernel: PCI host bridge to bus 0000:00 May 8 03:06:10.873033 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] May 8 03:06:10.873094 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] May 8 03:06:10.873155 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] May 8 03:06:10.873210 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 8 03:06:10.873287 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 May 8 03:06:10.873365 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 May 8 03:06:10.873430 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] May 8 03:06:10.873494 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] May 8 03:06:10.873558 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] May 8 03:06:10.873625 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] May 8 03:06:10.873690 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] May 8 03:06:10.873755 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] May 8 03:06:10.873825 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] May 8 03:06:10.873939 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] May 8 03:06:10.873998 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] May 8 03:06:10.874007 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 May 8 03:06:10.874017 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 May 8 03:06:10.874025 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 May 8 03:06:10.874031 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 May 8 03:06:10.874038 kernel: iommu: Default domain type: Translated May 8 03:06:10.874045 kernel: iommu: DMA domain TLB invalidation policy: strict mode May 8 03:06:10.874052 kernel: efivars: Registered efivars operations May 8 03:06:10.874059 kernel: vgaarb: loaded May 8 03:06:10.874066 kernel: clocksource: Switched to clocksource arch_sys_counter May 8 03:06:10.874073 kernel: VFS: Disk quotas dquot_6.6.0 May 8 03:06:10.874081 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 8 03:06:10.874088 kernel: pnp: PnP ACPI init May 8 03:06:10.874157 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved May 8 03:06:10.874167 kernel: pnp: PnP ACPI: found 1 devices May 8 03:06:10.874174 kernel: NET: Registered PF_INET protocol family May 8 03:06:10.874181 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 8 03:06:10.874188 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 8 03:06:10.874195 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 8 03:06:10.874204 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 8 03:06:10.874211 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 8 03:06:10.874218 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 8 03:06:10.874225 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 03:06:10.874231 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 03:06:10.874238 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 8 03:06:10.874245 kernel: PCI: CLS 0 bytes, default 64 May 8 03:06:10.874252 kernel: kvm [1]: HYP mode not available May 8 03:06:10.874259 kernel: Initialise system trusted keyrings May 8 03:06:10.874267 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 8 03:06:10.874274 kernel: Key type asymmetric registered May 8 03:06:10.874281 kernel: Asymmetric key parser 'x509' registered May 8 03:06:10.874288 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 8 03:06:10.874295 kernel: io scheduler mq-deadline registered May 8 03:06:10.874302 kernel: io scheduler kyber registered May 8 03:06:10.874308 kernel: io scheduler bfq registered May 8 03:06:10.874315 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 May 8 03:06:10.874322 kernel: ACPI: button: Power Button [PWRB] May 8 03:06:10.874331 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 May 8 03:06:10.874395 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) May 8 03:06:10.874405 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 8 03:06:10.874411 kernel: thunder_xcv, ver 1.0 May 8 03:06:10.874418 kernel: thunder_bgx, ver 1.0 May 8 03:06:10.874425 kernel: nicpf, ver 1.0 May 8 03:06:10.874432 kernel: nicvf, ver 1.0 May 8 03:06:10.874502 kernel: rtc-efi rtc-efi.0: registered as rtc0 May 8 03:06:10.874563 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-05-08T03:06:10 UTC (1746673570) May 8 03:06:10.874574 kernel: hid: raw HID events driver (C) Jiri Kosina May 8 03:06:10.874581 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available May 8 03:06:10.874588 kernel: watchdog: Delayed init of the lockup detector failed: -19 May 8 03:06:10.874594 kernel: watchdog: Hard watchdog permanently disabled May 8 03:06:10.874601 kernel: NET: Registered PF_INET6 protocol family May 8 03:06:10.874608 kernel: Segment Routing with IPv6 May 8 03:06:10.874615 kernel: In-situ OAM (IOAM) with IPv6 May 8 03:06:10.874622 kernel: NET: Registered PF_PACKET protocol family May 8 03:06:10.874630 kernel: Key type dns_resolver registered May 8 03:06:10.874637 kernel: registered taskstats version 1 May 8 03:06:10.874644 kernel: Loading compiled-in X.509 certificates May 8 03:06:10.874651 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: 0483a4f2a10389c228a3c9adee09b6a8520e21d0' May 8 03:06:10.874657 kernel: Key type .fscrypt registered May 8 03:06:10.874664 kernel: Key type fscrypt-provisioning registered May 8 03:06:10.874671 kernel: ima: No TPM chip found, activating TPM-bypass! May 8 03:06:10.874677 kernel: ima: Allocated hash algorithm: sha1 May 8 03:06:10.874684 kernel: ima: No architecture policies found May 8 03:06:10.874693 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) May 8 03:06:10.874699 kernel: clk: Disabling unused clocks May 8 03:06:10.874706 kernel: Warning: unable to open an initial console. May 8 03:06:10.874713 kernel: Freeing unused kernel memory: 39040K May 8 03:06:10.874720 kernel: Run /init as init process May 8 03:06:10.874727 kernel: with arguments: May 8 03:06:10.874734 kernel: /init May 8 03:06:10.874740 kernel: with environment: May 8 03:06:10.874747 kernel: HOME=/ May 8 03:06:10.874755 kernel: TERM=linux May 8 03:06:10.874762 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 8 03:06:10.874769 systemd[1]: Successfully made /usr/ read-only. May 8 03:06:10.874779 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 8 03:06:10.874796 systemd[1]: Detected virtualization kvm. May 8 03:06:10.874803 systemd[1]: Detected architecture arm64. May 8 03:06:10.874810 systemd[1]: Running in initrd. May 8 03:06:10.874820 systemd[1]: No hostname configured, using default hostname. May 8 03:06:10.874827 systemd[1]: Hostname set to . May 8 03:06:10.874835 systemd[1]: Initializing machine ID from VM UUID. May 8 03:06:10.874849 systemd[1]: Queued start job for default target initrd.target. May 8 03:06:10.874857 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 03:06:10.874864 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 03:06:10.874872 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 8 03:06:10.874880 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 8 03:06:10.874889 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 8 03:06:10.874897 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 8 03:06:10.874905 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 8 03:06:10.874913 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 8 03:06:10.874920 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 03:06:10.874928 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 8 03:06:10.874935 systemd[1]: Reached target paths.target - Path Units. May 8 03:06:10.874944 systemd[1]: Reached target slices.target - Slice Units. May 8 03:06:10.874951 systemd[1]: Reached target swap.target - Swaps. May 8 03:06:10.874958 systemd[1]: Reached target timers.target - Timer Units. May 8 03:06:10.874966 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 8 03:06:10.874973 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 8 03:06:10.874981 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 8 03:06:10.874988 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 8 03:06:10.874995 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 8 03:06:10.875003 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 8 03:06:10.875011 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 8 03:06:10.875019 systemd[1]: Reached target sockets.target - Socket Units. May 8 03:06:10.875026 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 8 03:06:10.875033 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 8 03:06:10.875041 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 8 03:06:10.875049 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 8 03:06:10.875056 systemd[1]: Starting systemd-fsck-usr.service... May 8 03:06:10.875063 systemd[1]: Starting systemd-journald.service - Journal Service... May 8 03:06:10.875072 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 8 03:06:10.875080 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 03:06:10.875087 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 8 03:06:10.875095 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 8 03:06:10.875102 systemd[1]: Finished systemd-fsck-usr.service. May 8 03:06:10.875111 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 8 03:06:10.875136 systemd-journald[237]: Collecting audit messages is disabled. May 8 03:06:10.875155 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 03:06:10.875165 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 8 03:06:10.875173 systemd-journald[237]: Journal started May 8 03:06:10.875190 systemd-journald[237]: Runtime Journal (/run/log/journal/ceb03f368d2748288151915c69eca4c7) is 5.9M, max 47.3M, 41.4M free. May 8 03:06:10.868589 systemd-modules-load[239]: Inserted module 'overlay' May 8 03:06:10.879576 systemd[1]: Started systemd-journald.service - Journal Service. May 8 03:06:10.881872 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 8 03:06:10.882047 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 8 03:06:10.884755 systemd-modules-load[239]: Inserted module 'br_netfilter' May 8 03:06:10.885658 kernel: Bridge firewalling registered May 8 03:06:10.885001 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 8 03:06:10.893614 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 8 03:06:10.895194 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 8 03:06:10.897985 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 03:06:10.900337 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 03:06:10.901234 systemd-tmpfiles[262]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 8 03:06:10.903899 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 8 03:06:10.911024 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 03:06:10.915860 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 8 03:06:10.919169 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 03:06:10.924927 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 8 03:06:10.935482 dracut-cmdline[281]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=b77ba75748ce73d160d869c11543eefaa814f1163d4a6d367aa8765f0183246a May 8 03:06:10.955929 systemd-resolved[285]: Positive Trust Anchors: May 8 03:06:10.955946 systemd-resolved[285]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 03:06:10.955977 systemd-resolved[285]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 8 03:06:10.960754 systemd-resolved[285]: Defaulting to hostname 'linux'. May 8 03:06:10.963739 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 8 03:06:10.964953 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 8 03:06:11.007870 kernel: SCSI subsystem initialized May 8 03:06:11.011864 kernel: Loading iSCSI transport class v2.0-870. May 8 03:06:11.020863 kernel: iscsi: registered transport (tcp) May 8 03:06:11.033899 kernel: iscsi: registered transport (qla4xxx) May 8 03:06:11.033948 kernel: QLogic iSCSI HBA Driver May 8 03:06:11.051698 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 8 03:06:11.067899 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 8 03:06:11.069999 systemd[1]: Reached target network-pre.target - Preparation for Network. May 8 03:06:11.115619 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 8 03:06:11.117953 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 8 03:06:11.181872 kernel: raid6: neonx8 gen() 15789 MB/s May 8 03:06:11.198870 kernel: raid6: neonx4 gen() 15818 MB/s May 8 03:06:11.215864 kernel: raid6: neonx2 gen() 13195 MB/s May 8 03:06:11.232867 kernel: raid6: neonx1 gen() 10527 MB/s May 8 03:06:11.249859 kernel: raid6: int64x8 gen() 6785 MB/s May 8 03:06:11.266856 kernel: raid6: int64x4 gen() 7344 MB/s May 8 03:06:11.283856 kernel: raid6: int64x2 gen() 6111 MB/s May 8 03:06:11.300868 kernel: raid6: int64x1 gen() 5058 MB/s May 8 03:06:11.300890 kernel: raid6: using algorithm neonx4 gen() 15818 MB/s May 8 03:06:11.317862 kernel: raid6: .... xor() 12424 MB/s, rmw enabled May 8 03:06:11.317880 kernel: raid6: using neon recovery algorithm May 8 03:06:11.323184 kernel: xor: measuring software checksum speed May 8 03:06:11.323201 kernel: 8regs : 21579 MB/sec May 8 03:06:11.323857 kernel: 32regs : 21699 MB/sec May 8 03:06:11.323868 kernel: arm64_neon : 25191 MB/sec May 8 03:06:11.324862 kernel: xor: using function: arm64_neon (25191 MB/sec) May 8 03:06:11.374865 kernel: Btrfs loaded, zoned=no, fsverity=no May 8 03:06:11.381624 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 8 03:06:11.384122 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 03:06:11.411100 systemd-udevd[493]: Using default interface naming scheme 'v255'. May 8 03:06:11.415299 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 03:06:11.418296 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 8 03:06:11.440342 dracut-pre-trigger[501]: rd.md=0: removing MD RAID activation May 8 03:06:11.463608 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 8 03:06:11.465981 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 8 03:06:11.524415 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 8 03:06:11.529073 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 8 03:06:11.572950 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues May 8 03:06:11.578408 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) May 8 03:06:11.578519 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 8 03:06:11.578537 kernel: GPT:9289727 != 19775487 May 8 03:06:11.578546 kernel: GPT:Alternate GPT header not at the end of the disk. May 8 03:06:11.578555 kernel: GPT:9289727 != 19775487 May 8 03:06:11.578563 kernel: GPT: Use GNU Parted to correct GPT errors. May 8 03:06:11.578571 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 03:06:11.583990 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 03:06:11.584113 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 03:06:11.586627 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 8 03:06:11.588879 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 03:06:11.604901 kernel: BTRFS: device fsid 3a1347d7-efc3-4e1d-b50f-33279a0b8e11 devid 1 transid 39 /dev/vda3 scanned by (udev-worker) (537) May 8 03:06:11.607885 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (546) May 8 03:06:11.615810 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 8 03:06:11.617000 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 03:06:11.624252 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 8 03:06:11.637171 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 8 03:06:11.643131 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 8 03:06:11.644039 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 8 03:06:11.652279 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 8 03:06:11.653212 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 8 03:06:11.654969 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 03:06:11.656693 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 8 03:06:11.659131 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 8 03:06:11.660646 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 8 03:06:11.675357 disk-uuid[587]: Primary Header is updated. May 8 03:06:11.675357 disk-uuid[587]: Secondary Entries is updated. May 8 03:06:11.675357 disk-uuid[587]: Secondary Header is updated. May 8 03:06:11.679445 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 8 03:06:11.684074 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 03:06:12.693866 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 03:06:12.694710 disk-uuid[592]: The operation has completed successfully. May 8 03:06:12.723432 systemd[1]: disk-uuid.service: Deactivated successfully. May 8 03:06:12.723548 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 8 03:06:12.747719 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 8 03:06:12.768378 sh[606]: Success May 8 03:06:12.781484 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 8 03:06:12.781536 kernel: device-mapper: uevent: version 1.0.3 May 8 03:06:12.781564 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 8 03:06:12.789891 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" May 8 03:06:12.815209 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 8 03:06:12.817953 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 8 03:06:12.832040 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 8 03:06:12.841499 kernel: BTRFS info (device dm-0): first mount of filesystem 3a1347d7-efc3-4e1d-b50f-33279a0b8e11 May 8 03:06:12.841534 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm May 8 03:06:12.841545 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 8 03:06:12.841555 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 8 03:06:12.842853 kernel: BTRFS info (device dm-0): using free space tree May 8 03:06:12.846561 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 8 03:06:12.847677 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 8 03:06:12.848822 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 8 03:06:12.849549 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 8 03:06:12.850982 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 8 03:06:12.877172 kernel: BTRFS info (device vda6): first mount of filesystem aa4c3b42-75ac-4f57-aade-913808f7268e May 8 03:06:12.877210 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm May 8 03:06:12.877220 kernel: BTRFS info (device vda6): using free space tree May 8 03:06:12.879862 kernel: BTRFS info (device vda6): auto enabling async discard May 8 03:06:12.882873 kernel: BTRFS info (device vda6): last unmount of filesystem aa4c3b42-75ac-4f57-aade-913808f7268e May 8 03:06:12.887391 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 8 03:06:12.890003 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 8 03:06:12.960924 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 8 03:06:12.965543 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 8 03:06:13.018695 systemd-networkd[793]: lo: Link UP May 8 03:06:13.018711 systemd-networkd[793]: lo: Gained carrier May 8 03:06:13.019533 systemd-networkd[793]: Enumeration completed May 8 03:06:13.020022 systemd-networkd[793]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 03:06:13.020026 systemd-networkd[793]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 03:06:13.020743 systemd-networkd[793]: eth0: Link UP May 8 03:06:13.020746 systemd-networkd[793]: eth0: Gained carrier May 8 03:06:13.020754 systemd-networkd[793]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 03:06:13.022889 systemd[1]: Started systemd-networkd.service - Network Configuration. May 8 03:06:13.028663 ignition[698]: Ignition 2.21.0 May 8 03:06:13.024103 systemd[1]: Reached target network.target - Network. May 8 03:06:13.028669 ignition[698]: Stage: fetch-offline May 8 03:06:13.028705 ignition[698]: no configs at "/usr/lib/ignition/base.d" May 8 03:06:13.028712 ignition[698]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:13.029052 ignition[698]: parsed url from cmdline: "" May 8 03:06:13.029058 ignition[698]: no config URL provided May 8 03:06:13.029063 ignition[698]: reading system config file "/usr/lib/ignition/user.ign" May 8 03:06:13.029071 ignition[698]: no config at "/usr/lib/ignition/user.ign" May 8 03:06:13.029213 ignition[698]: op(1): [started] loading QEMU firmware config module May 8 03:06:13.029217 ignition[698]: op(1): executing: "modprobe" "qemu_fw_cfg" May 8 03:06:13.037967 ignition[698]: op(1): [finished] loading QEMU firmware config module May 8 03:06:13.043909 systemd-networkd[793]: eth0: DHCPv4 address 10.0.0.7/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 8 03:06:13.076689 ignition[698]: parsing config with SHA512: fae2d209d64464d7ad3ac8174c51c72e6261e11419f5615f2f4896c05b5e5bf4c5f0d45ed8b07abb7feae994fa5fd920978455cc991dc18d03d67ab06f49d62d May 8 03:06:13.080935 unknown[698]: fetched base config from "system" May 8 03:06:13.080946 unknown[698]: fetched user config from "qemu" May 8 03:06:13.081429 ignition[698]: fetch-offline: fetch-offline passed May 8 03:06:13.084892 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 8 03:06:13.081492 ignition[698]: Ignition finished successfully May 8 03:06:13.086214 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 8 03:06:13.086964 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 8 03:06:13.114002 ignition[807]: Ignition 2.21.0 May 8 03:06:13.114015 ignition[807]: Stage: kargs May 8 03:06:13.114132 ignition[807]: no configs at "/usr/lib/ignition/base.d" May 8 03:06:13.114141 ignition[807]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:13.117210 ignition[807]: kargs: kargs passed May 8 03:06:13.117281 ignition[807]: Ignition finished successfully May 8 03:06:13.119625 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 8 03:06:13.121536 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 8 03:06:13.143281 ignition[816]: Ignition 2.21.0 May 8 03:06:13.143301 ignition[816]: Stage: disks May 8 03:06:13.143442 ignition[816]: no configs at "/usr/lib/ignition/base.d" May 8 03:06:13.143451 ignition[816]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:13.144628 ignition[816]: disks: disks passed May 8 03:06:13.144677 ignition[816]: Ignition finished successfully May 8 03:06:13.146902 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 8 03:06:13.148394 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 8 03:06:13.149868 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 8 03:06:13.151840 systemd[1]: Reached target local-fs.target - Local File Systems. May 8 03:06:13.153706 systemd[1]: Reached target sysinit.target - System Initialization. May 8 03:06:13.155685 systemd[1]: Reached target basic.target - Basic System. May 8 03:06:13.158146 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 8 03:06:13.160121 systemd-resolved[285]: Detected conflict on linux IN A 10.0.0.7 May 8 03:06:13.160129 systemd-resolved[285]: Hostname conflict, changing published hostname from 'linux' to 'linux11'. May 8 03:06:13.183246 systemd-fsck[826]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 8 03:06:13.187048 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 8 03:06:13.189617 systemd[1]: Mounting sysroot.mount - /sysroot... May 8 03:06:13.245874 kernel: EXT4-fs (vda9): mounted filesystem 2ee8b9b8-c65d-4d4e-8e48-c1c49d82b981 r/w with ordered data mode. Quota mode: none. May 8 03:06:13.246109 systemd[1]: Mounted sysroot.mount - /sysroot. May 8 03:06:13.247149 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 8 03:06:13.250060 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 8 03:06:13.252016 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 8 03:06:13.252773 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 8 03:06:13.252819 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 8 03:06:13.252859 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 8 03:06:13.259027 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 8 03:06:13.261008 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 8 03:06:13.266716 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (834) May 8 03:06:13.266752 kernel: BTRFS info (device vda6): first mount of filesystem aa4c3b42-75ac-4f57-aade-913808f7268e May 8 03:06:13.266763 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm May 8 03:06:13.268098 kernel: BTRFS info (device vda6): using free space tree May 8 03:06:13.270868 kernel: BTRFS info (device vda6): auto enabling async discard May 8 03:06:13.270989 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 8 03:06:13.304711 initrd-setup-root[858]: cut: /sysroot/etc/passwd: No such file or directory May 8 03:06:13.308959 initrd-setup-root[865]: cut: /sysroot/etc/group: No such file or directory May 8 03:06:13.312871 initrd-setup-root[872]: cut: /sysroot/etc/shadow: No such file or directory May 8 03:06:13.315425 initrd-setup-root[879]: cut: /sysroot/etc/gshadow: No such file or directory May 8 03:06:13.380096 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 8 03:06:13.381999 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 8 03:06:13.383228 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 8 03:06:13.405866 kernel: BTRFS info (device vda6): last unmount of filesystem aa4c3b42-75ac-4f57-aade-913808f7268e May 8 03:06:13.420047 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 8 03:06:13.433717 ignition[948]: INFO : Ignition 2.21.0 May 8 03:06:13.433717 ignition[948]: INFO : Stage: mount May 8 03:06:13.435740 ignition[948]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 03:06:13.435740 ignition[948]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:13.438496 ignition[948]: INFO : mount: mount passed May 8 03:06:13.438496 ignition[948]: INFO : Ignition finished successfully May 8 03:06:13.439004 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 8 03:06:13.440943 systemd[1]: Starting ignition-files.service - Ignition (files)... May 8 03:06:13.980180 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 8 03:06:13.981698 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 8 03:06:14.000428 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (960) May 8 03:06:14.000474 kernel: BTRFS info (device vda6): first mount of filesystem aa4c3b42-75ac-4f57-aade-913808f7268e May 8 03:06:14.000484 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm May 8 03:06:14.001174 kernel: BTRFS info (device vda6): using free space tree May 8 03:06:14.003855 kernel: BTRFS info (device vda6): auto enabling async discard May 8 03:06:14.004834 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 8 03:06:14.030603 ignition[977]: INFO : Ignition 2.21.0 May 8 03:06:14.030603 ignition[977]: INFO : Stage: files May 8 03:06:14.032298 ignition[977]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 03:06:14.032298 ignition[977]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:14.032298 ignition[977]: DEBUG : files: compiled without relabeling support, skipping May 8 03:06:14.035705 ignition[977]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 8 03:06:14.035705 ignition[977]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 8 03:06:14.038418 ignition[977]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 8 03:06:14.038418 ignition[977]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 8 03:06:14.038418 ignition[977]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 8 03:06:14.038418 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 8 03:06:14.038418 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 May 8 03:06:14.036313 unknown[977]: wrote ssh authorized keys file for user: core May 8 03:06:14.087298 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 8 03:06:14.293108 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 8 03:06:14.293108 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 8 03:06:14.297079 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 May 8 03:06:14.357963 systemd-networkd[793]: eth0: Gained IPv6LL May 8 03:06:15.333215 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 8 03:06:15.585479 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 8 03:06:15.585479 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 8 03:06:15.589005 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 May 8 03:06:16.118172 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 8 03:06:17.109454 ignition[977]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 8 03:06:17.109454 ignition[977]: INFO : files: op(c): [started] processing unit "prepare-helm.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" May 8 03:06:17.113299 ignition[977]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" May 8 03:06:17.130115 ignition[977]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" May 8 03:06:17.133187 ignition[977]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 8 03:06:17.134691 ignition[977]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" May 8 03:06:17.134691 ignition[977]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" May 8 03:06:17.134691 ignition[977]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" May 8 03:06:17.134691 ignition[977]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" May 8 03:06:17.134691 ignition[977]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" May 8 03:06:17.134691 ignition[977]: INFO : files: files passed May 8 03:06:17.134691 ignition[977]: INFO : Ignition finished successfully May 8 03:06:17.136523 systemd[1]: Finished ignition-files.service - Ignition (files). May 8 03:06:17.139190 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 8 03:06:17.140953 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 8 03:06:17.157758 systemd[1]: ignition-quench.service: Deactivated successfully. May 8 03:06:17.157876 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 8 03:06:17.161396 initrd-setup-root-after-ignition[1005]: grep: /sysroot/oem/oem-release: No such file or directory May 8 03:06:17.162750 initrd-setup-root-after-ignition[1007]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 8 03:06:17.162750 initrd-setup-root-after-ignition[1007]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 8 03:06:17.165629 initrd-setup-root-after-ignition[1012]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 8 03:06:17.164956 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 8 03:06:17.167067 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 8 03:06:17.169650 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 8 03:06:17.204954 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 8 03:06:17.205056 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 8 03:06:17.207002 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 8 03:06:17.208640 systemd[1]: Reached target initrd.target - Initrd Default Target. May 8 03:06:17.210316 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 8 03:06:17.210978 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 8 03:06:17.224438 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 8 03:06:17.226508 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 8 03:06:17.243172 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 8 03:06:17.244099 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 03:06:17.245817 systemd[1]: Stopped target timers.target - Timer Units. May 8 03:06:17.247411 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 8 03:06:17.247521 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 8 03:06:17.249694 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 8 03:06:17.250527 systemd[1]: Stopped target basic.target - Basic System. May 8 03:06:17.252130 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 8 03:06:17.253683 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 8 03:06:17.255181 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 8 03:06:17.256756 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 8 03:06:17.258527 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 8 03:06:17.260101 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 8 03:06:17.261813 systemd[1]: Stopped target sysinit.target - System Initialization. May 8 03:06:17.263354 systemd[1]: Stopped target local-fs.target - Local File Systems. May 8 03:06:17.265008 systemd[1]: Stopped target swap.target - Swaps. May 8 03:06:17.266334 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 8 03:06:17.266444 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 8 03:06:17.268468 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 8 03:06:17.269334 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 03:06:17.270865 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 8 03:06:17.271922 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 03:06:17.273612 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 8 03:06:17.273715 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 8 03:06:17.276265 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 8 03:06:17.276373 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 8 03:06:17.278411 systemd[1]: Stopped target paths.target - Path Units. May 8 03:06:17.279694 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 8 03:06:17.282898 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 03:06:17.284395 systemd[1]: Stopped target slices.target - Slice Units. May 8 03:06:17.285866 systemd[1]: Stopped target sockets.target - Socket Units. May 8 03:06:17.287726 systemd[1]: iscsid.socket: Deactivated successfully. May 8 03:06:17.287817 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 8 03:06:17.289198 systemd[1]: iscsiuio.socket: Deactivated successfully. May 8 03:06:17.289269 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 8 03:06:17.290727 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 8 03:06:17.290857 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 8 03:06:17.292523 systemd[1]: ignition-files.service: Deactivated successfully. May 8 03:06:17.292621 systemd[1]: Stopped ignition-files.service - Ignition (files). May 8 03:06:17.294887 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 8 03:06:17.296569 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 8 03:06:17.296700 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 8 03:06:17.309408 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 8 03:06:17.310251 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 8 03:06:17.310371 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 8 03:06:17.312153 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 8 03:06:17.312250 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 8 03:06:17.318747 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 8 03:06:17.318855 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 8 03:06:17.322897 ignition[1033]: INFO : Ignition 2.21.0 May 8 03:06:17.322897 ignition[1033]: INFO : Stage: umount May 8 03:06:17.322897 ignition[1033]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 03:06:17.322897 ignition[1033]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 03:06:17.326714 ignition[1033]: INFO : umount: umount passed May 8 03:06:17.326714 ignition[1033]: INFO : Ignition finished successfully May 8 03:06:17.324071 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 8 03:06:17.326250 systemd[1]: ignition-mount.service: Deactivated successfully. May 8 03:06:17.326910 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 8 03:06:17.328510 systemd[1]: sysroot-boot.service: Deactivated successfully. May 8 03:06:17.328592 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 8 03:06:17.330790 systemd[1]: Stopped target network.target - Network. May 8 03:06:17.332323 systemd[1]: ignition-disks.service: Deactivated successfully. May 8 03:06:17.332389 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 8 03:06:17.333963 systemd[1]: ignition-kargs.service: Deactivated successfully. May 8 03:06:17.334017 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 8 03:06:17.335804 systemd[1]: ignition-setup.service: Deactivated successfully. May 8 03:06:17.335866 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 8 03:06:17.337411 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 8 03:06:17.337458 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 8 03:06:17.339034 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 8 03:06:17.339086 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 8 03:06:17.341015 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 8 03:06:17.342710 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 8 03:06:17.346221 systemd[1]: systemd-resolved.service: Deactivated successfully. May 8 03:06:17.346330 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 8 03:06:17.349328 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 8 03:06:17.349622 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 8 03:06:17.349658 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 03:06:17.352889 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 8 03:06:17.354963 systemd[1]: systemd-networkd.service: Deactivated successfully. May 8 03:06:17.355079 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 8 03:06:17.359786 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 8 03:06:17.359972 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 8 03:06:17.361642 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 8 03:06:17.361678 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 8 03:06:17.364568 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 8 03:06:17.366069 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 8 03:06:17.366126 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 8 03:06:17.367805 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 8 03:06:17.367905 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 8 03:06:17.370805 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 8 03:06:17.370909 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 8 03:06:17.372715 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 03:06:17.375657 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 8 03:06:17.391619 systemd[1]: systemd-udevd.service: Deactivated successfully. May 8 03:06:17.392638 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 03:06:17.394349 systemd[1]: network-cleanup.service: Deactivated successfully. May 8 03:06:17.394439 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 8 03:06:17.397115 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 8 03:06:17.397176 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 8 03:06:17.398635 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 8 03:06:17.398669 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 8 03:06:17.400519 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 8 03:06:17.400573 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 8 03:06:17.403803 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 8 03:06:17.403862 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 8 03:06:17.406550 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 8 03:06:17.406598 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 8 03:06:17.410187 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 8 03:06:17.411208 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 8 03:06:17.411281 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 8 03:06:17.414080 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 8 03:06:17.414121 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 03:06:17.417409 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 03:06:17.417453 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 03:06:17.426917 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 8 03:06:17.427012 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 8 03:06:17.429169 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 8 03:06:17.431643 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 8 03:06:17.453829 systemd[1]: Switching root. May 8 03:06:17.481163 systemd-journald[237]: Journal stopped May 8 03:06:18.241896 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). May 8 03:06:18.241956 kernel: SELinux: policy capability network_peer_controls=1 May 8 03:06:18.241969 kernel: SELinux: policy capability open_perms=1 May 8 03:06:18.241978 kernel: SELinux: policy capability extended_socket_class=1 May 8 03:06:18.241994 kernel: SELinux: policy capability always_check_network=0 May 8 03:06:18.242005 kernel: SELinux: policy capability cgroup_seclabel=1 May 8 03:06:18.242015 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 8 03:06:18.242024 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 8 03:06:18.242033 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 8 03:06:18.242042 kernel: audit: type=1403 audit(1746673577.650:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 8 03:06:18.242055 systemd[1]: Successfully loaded SELinux policy in 39.148ms. May 8 03:06:18.242076 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.159ms. May 8 03:06:18.242087 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 8 03:06:18.242098 systemd[1]: Detected virtualization kvm. May 8 03:06:18.242108 systemd[1]: Detected architecture arm64. May 8 03:06:18.242119 systemd[1]: Detected first boot. May 8 03:06:18.242129 systemd[1]: Initializing machine ID from VM UUID. May 8 03:06:18.242140 zram_generator::config[1078]: No configuration found. May 8 03:06:18.242154 kernel: NET: Registered PF_VSOCK protocol family May 8 03:06:18.242164 systemd[1]: Populated /etc with preset unit settings. May 8 03:06:18.242175 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 8 03:06:18.242185 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 8 03:06:18.242195 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 8 03:06:18.242206 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 8 03:06:18.242217 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 8 03:06:18.242227 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 8 03:06:18.242239 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 8 03:06:18.242249 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 8 03:06:18.242259 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 8 03:06:18.242269 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 8 03:06:18.242279 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 8 03:06:18.242290 systemd[1]: Created slice user.slice - User and Session Slice. May 8 03:06:18.242302 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 03:06:18.242312 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 03:06:18.242322 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 8 03:06:18.242332 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 8 03:06:18.242343 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 8 03:06:18.242353 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 8 03:06:18.242363 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... May 8 03:06:18.242373 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 03:06:18.242383 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 8 03:06:18.242395 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 8 03:06:18.242406 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 8 03:06:18.242415 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 8 03:06:18.242425 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 8 03:06:18.242436 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 03:06:18.242446 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 8 03:06:18.242457 systemd[1]: Reached target slices.target - Slice Units. May 8 03:06:18.242467 systemd[1]: Reached target swap.target - Swaps. May 8 03:06:18.242478 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 8 03:06:18.242489 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 8 03:06:18.242499 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 8 03:06:18.242509 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 8 03:06:18.242520 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 8 03:06:18.242530 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 8 03:06:18.242540 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 8 03:06:18.242550 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 8 03:06:18.242560 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 8 03:06:18.242572 systemd[1]: Mounting media.mount - External Media Directory... May 8 03:06:18.242582 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 8 03:06:18.242592 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 8 03:06:18.242602 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 8 03:06:18.242613 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 8 03:06:18.242623 systemd[1]: Reached target machines.target - Containers. May 8 03:06:18.242633 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 8 03:06:18.242643 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 03:06:18.242655 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 8 03:06:18.242665 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 8 03:06:18.242676 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 03:06:18.242687 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 8 03:06:18.242696 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 03:06:18.242707 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 8 03:06:18.242716 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 03:06:18.242727 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 8 03:06:18.242737 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 8 03:06:18.242749 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 8 03:06:18.242759 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 8 03:06:18.242774 systemd[1]: Stopped systemd-fsck-usr.service. May 8 03:06:18.242787 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 03:06:18.242797 systemd[1]: Starting systemd-journald.service - Journal Service... May 8 03:06:18.242807 kernel: fuse: init (API version 7.39) May 8 03:06:18.242817 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 8 03:06:18.242827 kernel: loop: module loaded May 8 03:06:18.242837 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 8 03:06:18.242856 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 8 03:06:18.242867 kernel: ACPI: bus type drm_connector registered May 8 03:06:18.242877 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 8 03:06:18.242888 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 8 03:06:18.242899 systemd[1]: verity-setup.service: Deactivated successfully. May 8 03:06:18.242910 systemd[1]: Stopped verity-setup.service. May 8 03:06:18.242921 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 8 03:06:18.242931 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 8 03:06:18.242941 systemd[1]: Mounted media.mount - External Media Directory. May 8 03:06:18.242951 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 8 03:06:18.242961 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 8 03:06:18.242991 systemd-journald[1147]: Collecting audit messages is disabled. May 8 03:06:18.243013 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 8 03:06:18.243024 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 8 03:06:18.243035 systemd-journald[1147]: Journal started May 8 03:06:18.243056 systemd-journald[1147]: Runtime Journal (/run/log/journal/ceb03f368d2748288151915c69eca4c7) is 5.9M, max 47.3M, 41.4M free. May 8 03:06:18.035822 systemd[1]: Queued start job for default target multi-user.target. May 8 03:06:18.048759 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 8 03:06:18.049141 systemd[1]: systemd-journald.service: Deactivated successfully. May 8 03:06:18.245206 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 8 03:06:18.246993 systemd[1]: Started systemd-journald.service - Journal Service. May 8 03:06:18.247716 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 8 03:06:18.247938 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 8 03:06:18.249313 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 03:06:18.249476 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 03:06:18.250570 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 03:06:18.250738 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 8 03:06:18.251972 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 03:06:18.252141 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 03:06:18.253253 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 8 03:06:18.253412 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 8 03:06:18.254531 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 03:06:18.254690 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 03:06:18.255796 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 8 03:06:18.256977 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 8 03:06:18.258130 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 8 03:06:18.259502 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 8 03:06:18.272530 systemd[1]: Reached target network-pre.target - Preparation for Network. May 8 03:06:18.274798 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 8 03:06:18.276713 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 8 03:06:18.277642 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 8 03:06:18.277677 systemd[1]: Reached target local-fs.target - Local File Systems. May 8 03:06:18.279438 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 8 03:06:18.287657 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 8 03:06:18.288887 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 03:06:18.289973 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 8 03:06:18.292037 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 8 03:06:18.293345 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 03:06:18.296986 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 8 03:06:18.298109 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 8 03:06:18.299233 systemd-journald[1147]: Time spent on flushing to /var/log/journal/ceb03f368d2748288151915c69eca4c7 is 20.831ms for 879 entries. May 8 03:06:18.299233 systemd-journald[1147]: System Journal (/var/log/journal/ceb03f368d2748288151915c69eca4c7) is 8M, max 195.6M, 187.6M free. May 8 03:06:18.338573 systemd-journald[1147]: Received client request to flush runtime journal. May 8 03:06:18.338926 kernel: loop0: detected capacity change from 0 to 138376 May 8 03:06:18.299282 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 03:06:18.302300 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 8 03:06:18.305536 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 8 03:06:18.311294 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 8 03:06:18.315138 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 8 03:06:18.321105 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 8 03:06:18.329907 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 8 03:06:18.331370 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 8 03:06:18.334316 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 8 03:06:18.336266 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 03:06:18.345264 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 8 03:06:18.346070 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 8 03:06:18.362633 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 8 03:06:18.366163 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 8 03:06:18.371908 kernel: loop1: detected capacity change from 0 to 194096 May 8 03:06:18.377235 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 8 03:06:18.403999 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. May 8 03:06:18.404017 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. May 8 03:06:18.412474 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 03:06:18.430867 kernel: loop2: detected capacity change from 0 to 107312 May 8 03:06:18.479893 kernel: loop3: detected capacity change from 0 to 138376 May 8 03:06:18.488451 kernel: loop4: detected capacity change from 0 to 194096 May 8 03:06:18.494891 kernel: loop5: detected capacity change from 0 to 107312 May 8 03:06:18.499607 (sd-merge)[1218]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. May 8 03:06:18.500079 (sd-merge)[1218]: Merged extensions into '/usr'. May 8 03:06:18.503461 systemd[1]: Reload requested from client PID 1194 ('systemd-sysext') (unit systemd-sysext.service)... May 8 03:06:18.503481 systemd[1]: Reloading... May 8 03:06:18.562882 zram_generator::config[1247]: No configuration found. May 8 03:06:18.609940 ldconfig[1189]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 8 03:06:18.637535 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 03:06:18.699554 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 8 03:06:18.699694 systemd[1]: Reloading finished in 195 ms. May 8 03:06:18.719535 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 8 03:06:18.721129 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 8 03:06:18.738098 systemd[1]: Starting ensure-sysext.service... May 8 03:06:18.739935 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 8 03:06:18.750534 systemd[1]: Reload requested from client PID 1278 ('systemctl') (unit ensure-sysext.service)... May 8 03:06:18.750549 systemd[1]: Reloading... May 8 03:06:18.794945 zram_generator::config[1307]: No configuration found. May 8 03:06:18.800706 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 8 03:06:18.800740 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 8 03:06:18.801074 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 8 03:06:18.801262 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 8 03:06:18.801959 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 8 03:06:18.802179 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. May 8 03:06:18.802228 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. May 8 03:06:18.805064 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. May 8 03:06:18.805078 systemd-tmpfiles[1279]: Skipping /boot May 8 03:06:18.814632 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. May 8 03:06:18.814650 systemd-tmpfiles[1279]: Skipping /boot May 8 03:06:18.891598 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 03:06:18.953805 systemd[1]: Reloading finished in 202 ms. May 8 03:06:18.964357 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 8 03:06:18.980143 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 03:06:18.990141 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 8 03:06:18.993521 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 8 03:06:19.000858 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 8 03:06:19.004199 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 8 03:06:19.006539 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 03:06:19.012413 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 8 03:06:19.016023 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 03:06:19.017270 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 03:06:19.019679 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 03:06:19.028914 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 03:06:19.030093 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 03:06:19.030221 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 03:06:19.031988 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 03:06:19.032906 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 03:06:19.034306 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 03:06:19.034452 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 03:06:19.037328 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 03:06:19.037477 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 03:06:19.042480 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 8 03:06:19.047018 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 03:06:19.048494 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 03:06:19.052155 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 03:06:19.062716 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 03:06:19.063935 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 03:06:19.064108 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 03:06:19.068117 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 8 03:06:19.068722 systemd-udevd[1347]: Using default interface naming scheme 'v255'. May 8 03:06:19.071963 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 8 03:06:19.076306 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 8 03:06:19.079887 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 03:06:19.080047 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 03:06:19.081640 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 03:06:19.081815 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 03:06:19.083439 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 03:06:19.083586 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 03:06:19.088119 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 8 03:06:19.090474 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 03:06:19.090980 augenrules[1380]: No rules May 8 03:06:19.098346 systemd[1]: audit-rules.service: Deactivated successfully. May 8 03:06:19.099861 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 8 03:06:19.101436 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 8 03:06:19.113280 systemd[1]: Finished ensure-sysext.service. May 8 03:06:19.116693 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 03:06:19.118943 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 03:06:19.121156 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 8 03:06:19.122748 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 03:06:19.125855 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 03:06:19.126753 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 03:06:19.126812 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 03:06:19.139116 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 8 03:06:19.146243 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 8 03:06:19.147096 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 03:06:19.147733 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 03:06:19.147938 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 03:06:19.149247 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 03:06:19.149399 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 8 03:06:19.150442 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 03:06:19.150600 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 03:06:19.151951 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 03:06:19.152101 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 03:06:19.162400 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 03:06:19.162461 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 8 03:06:19.173524 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 8 03:06:19.184867 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1401) May 8 03:06:19.186141 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. May 8 03:06:19.267371 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 8 03:06:19.268986 systemd[1]: Reached target time-set.target - System Time Set. May 8 03:06:19.274897 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 8 03:06:19.276689 systemd-networkd[1424]: lo: Link UP May 8 03:06:19.276706 systemd-networkd[1424]: lo: Gained carrier May 8 03:06:19.277833 systemd-networkd[1424]: Enumeration completed May 8 03:06:19.278737 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 8 03:06:19.279358 systemd-resolved[1346]: Positive Trust Anchors: May 8 03:06:19.279373 systemd-resolved[1346]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 03:06:19.279405 systemd-resolved[1346]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 8 03:06:19.280109 systemd[1]: Started systemd-networkd.service - Network Configuration. May 8 03:06:19.281389 systemd-networkd[1424]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 03:06:19.281393 systemd-networkd[1424]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 03:06:19.283625 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 8 03:06:19.284293 systemd-networkd[1424]: eth0: Link UP May 8 03:06:19.284300 systemd-networkd[1424]: eth0: Gained carrier May 8 03:06:19.284316 systemd-networkd[1424]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 03:06:19.289445 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 8 03:06:19.293589 systemd-resolved[1346]: Defaulting to hostname 'linux'. May 8 03:06:19.296605 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 8 03:06:19.297827 systemd[1]: Reached target network.target - Network. May 8 03:06:19.298820 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 8 03:06:19.299410 systemd-networkd[1424]: eth0: DHCPv4 address 10.0.0.7/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 8 03:06:19.299950 systemd-timesyncd[1425]: Network configuration changed, trying to establish connection. May 8 03:06:19.300529 systemd[1]: Reached target sysinit.target - System Initialization. May 8 03:06:18.882958 systemd-timesyncd[1425]: Contacted time server 10.0.0.1:123 (10.0.0.1). May 8 03:06:18.890998 systemd-journald[1147]: Time jumped backwards, rotating. May 8 03:06:18.883008 systemd-timesyncd[1425]: Initial clock synchronization to Thu 2025-05-08 03:06:18.882876 UTC. May 8 03:06:18.883368 systemd-resolved[1346]: Clock change detected. Flushing caches. May 8 03:06:18.884376 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 8 03:06:18.885878 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 8 03:06:18.887738 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 8 03:06:18.889112 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 8 03:06:18.890822 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 8 03:06:18.892046 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 8 03:06:18.892081 systemd[1]: Reached target paths.target - Path Units. May 8 03:06:18.892945 systemd[1]: Reached target timers.target - Timer Units. May 8 03:06:18.894844 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 8 03:06:18.897353 systemd[1]: Starting docker.socket - Docker Socket for the API... May 8 03:06:18.901287 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 8 03:06:18.903006 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 8 03:06:18.904264 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 8 03:06:18.913815 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 8 03:06:18.914994 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 8 03:06:18.916922 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 8 03:06:18.918176 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 8 03:06:18.919264 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 8 03:06:18.921107 systemd[1]: Reached target sockets.target - Socket Units. May 8 03:06:18.922900 systemd[1]: Reached target basic.target - Basic System. May 8 03:06:18.923634 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 8 03:06:18.923674 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 8 03:06:18.927914 systemd[1]: Starting containerd.service - containerd container runtime... May 8 03:06:18.929636 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 8 03:06:18.931340 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 8 03:06:18.934919 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 8 03:06:18.937968 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 8 03:06:18.938767 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 8 03:06:18.940979 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 8 03:06:18.945325 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 8 03:06:18.948695 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 8 03:06:18.953852 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 8 03:06:18.958297 jq[1471]: false May 8 03:06:18.959412 systemd[1]: Starting systemd-logind.service - User Login Management... May 8 03:06:18.961061 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 8 03:06:18.961476 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 8 03:06:18.962109 systemd[1]: Starting update-engine.service - Update Engine... May 8 03:06:18.966689 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 8 03:06:18.971854 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 8 03:06:18.973841 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 8 03:06:18.974065 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 8 03:06:18.974909 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 8 03:06:18.975105 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 8 03:06:18.980396 systemd[1]: motdgen.service: Deactivated successfully. May 8 03:06:18.980585 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 8 03:06:18.981046 jq[1485]: true May 8 03:06:18.982287 extend-filesystems[1472]: Found loop3 May 8 03:06:18.992362 extend-filesystems[1472]: Found loop4 May 8 03:06:18.992362 extend-filesystems[1472]: Found loop5 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda May 8 03:06:18.992362 extend-filesystems[1472]: Found vda1 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda2 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda3 May 8 03:06:18.992362 extend-filesystems[1472]: Found usr May 8 03:06:18.992362 extend-filesystems[1472]: Found vda4 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda6 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda7 May 8 03:06:18.992362 extend-filesystems[1472]: Found vda9 May 8 03:06:18.992362 extend-filesystems[1472]: Checking size of /dev/vda9 May 8 03:06:19.010869 jq[1495]: true May 8 03:06:19.014760 (ntainerd)[1496]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 8 03:06:19.020900 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 03:06:19.027294 extend-filesystems[1472]: Resized partition /dev/vda9 May 8 03:06:19.029898 extend-filesystems[1510]: resize2fs 1.47.2 (1-Jan-2025) May 8 03:06:19.036816 tar[1490]: linux-arm64/helm May 8 03:06:19.043752 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks May 8 03:06:19.047578 dbus-daemon[1469]: [system] SELinux support is enabled May 8 03:06:19.051374 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 8 03:06:19.056740 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 8 03:06:19.056769 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 8 03:06:19.059451 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 8 03:06:19.059475 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 8 03:06:19.062738 kernel: EXT4-fs (vda9): resized filesystem to 1864699 May 8 03:06:19.074216 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1400) May 8 03:06:19.074253 update_engine[1483]: I20250508 03:06:19.066371 1483 main.cc:92] Flatcar Update Engine starting May 8 03:06:19.075674 extend-filesystems[1510]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 8 03:06:19.075674 extend-filesystems[1510]: old_desc_blocks = 1, new_desc_blocks = 1 May 8 03:06:19.075674 extend-filesystems[1510]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. May 8 03:06:19.087751 extend-filesystems[1472]: Resized filesystem in /dev/vda9 May 8 03:06:19.091801 update_engine[1483]: I20250508 03:06:19.082628 1483 update_check_scheduler.cc:74] Next update check in 5m51s May 8 03:06:19.078310 systemd[1]: extend-filesystems.service: Deactivated successfully. May 8 03:06:19.078497 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 8 03:06:19.110589 systemd[1]: Started update-engine.service - Update Engine. May 8 03:06:19.115464 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 03:06:19.118638 systemd-logind[1481]: Watching system buttons on /dev/input/event0 (Power Button) May 8 03:06:19.119967 systemd-logind[1481]: New seat seat0. May 8 03:06:19.128221 systemd[1]: Started systemd-logind.service - User Login Management. May 8 03:06:19.136007 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 8 03:06:19.158278 bash[1529]: Updated "/home/core/.ssh/authorized_keys" May 8 03:06:19.160229 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 8 03:06:19.164065 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. May 8 03:06:19.198804 locksmithd[1533]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 8 03:06:19.252106 sshd_keygen[1491]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 8 03:06:19.271081 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 8 03:06:19.276820 systemd[1]: Starting issuegen.service - Generate /run/issue... May 8 03:06:19.284991 containerd[1496]: time="2025-05-08T03:06:19Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 8 03:06:19.287118 containerd[1496]: time="2025-05-08T03:06:19.286144259Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 8 03:06:19.293404 systemd[1]: issuegen.service: Deactivated successfully. May 8 03:06:19.293632 systemd[1]: Finished issuegen.service - Generate /run/issue. May 8 03:06:19.296636 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 8 03:06:19.298420 containerd[1496]: time="2025-05-08T03:06:19.298385939Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="9.12µs" May 8 03:06:19.298752 containerd[1496]: time="2025-05-08T03:06:19.298483179Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 8 03:06:19.298752 containerd[1496]: time="2025-05-08T03:06:19.298507659Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 8 03:06:19.298752 containerd[1496]: time="2025-05-08T03:06:19.298655819Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 8 03:06:19.298752 containerd[1496]: time="2025-05-08T03:06:19.298673339Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 8 03:06:19.298752 containerd[1496]: time="2025-05-08T03:06:19.298697659Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 8 03:06:19.298927 containerd[1496]: time="2025-05-08T03:06:19.298906379Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 8 03:06:19.298973 containerd[1496]: time="2025-05-08T03:06:19.298961819Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 8 03:06:19.299273 containerd[1496]: time="2025-05-08T03:06:19.299250219Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 8 03:06:19.299344 containerd[1496]: time="2025-05-08T03:06:19.299330219Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 8 03:06:19.299391 containerd[1496]: time="2025-05-08T03:06:19.299379579Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 8 03:06:19.299431 containerd[1496]: time="2025-05-08T03:06:19.299420779Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 8 03:06:19.299561 containerd[1496]: time="2025-05-08T03:06:19.299543699Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 8 03:06:19.299863 containerd[1496]: time="2025-05-08T03:06:19.299840259Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 8 03:06:19.299967 containerd[1496]: time="2025-05-08T03:06:19.299951299Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 8 03:06:19.300022 containerd[1496]: time="2025-05-08T03:06:19.300008779Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 8 03:06:19.300597 containerd[1496]: time="2025-05-08T03:06:19.300565899Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 8 03:06:19.300992 containerd[1496]: time="2025-05-08T03:06:19.300957739Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 8 03:06:19.301075 containerd[1496]: time="2025-05-08T03:06:19.301058939Z" level=info msg="metadata content store policy set" policy=shared May 8 03:06:19.304944 containerd[1496]: time="2025-05-08T03:06:19.304906179Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 8 03:06:19.305007 containerd[1496]: time="2025-05-08T03:06:19.304961619Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 8 03:06:19.305007 containerd[1496]: time="2025-05-08T03:06:19.304976779Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 8 03:06:19.305007 containerd[1496]: time="2025-05-08T03:06:19.304989299Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 8 03:06:19.305007 containerd[1496]: time="2025-05-08T03:06:19.305001579Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305011979Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305022739Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305034499Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305047179Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305057299Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305066419Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 8 03:06:19.305089 containerd[1496]: time="2025-05-08T03:06:19.305078259Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 8 03:06:19.305220 containerd[1496]: time="2025-05-08T03:06:19.305199699Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 8 03:06:19.305246 containerd[1496]: time="2025-05-08T03:06:19.305228779Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 8 03:06:19.305270 containerd[1496]: time="2025-05-08T03:06:19.305244539Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 8 03:06:19.305270 containerd[1496]: time="2025-05-08T03:06:19.305255219Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 8 03:06:19.305270 containerd[1496]: time="2025-05-08T03:06:19.305265939Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 8 03:06:19.305320 containerd[1496]: time="2025-05-08T03:06:19.305275979Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 8 03:06:19.305320 containerd[1496]: time="2025-05-08T03:06:19.305286659Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 8 03:06:19.305320 containerd[1496]: time="2025-05-08T03:06:19.305296259Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 8 03:06:19.305320 containerd[1496]: time="2025-05-08T03:06:19.305307699Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 8 03:06:19.305320 containerd[1496]: time="2025-05-08T03:06:19.305318779Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 8 03:06:19.305557 containerd[1496]: time="2025-05-08T03:06:19.305339459Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 8 03:06:19.305557 containerd[1496]: time="2025-05-08T03:06:19.305522019Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 8 03:06:19.305557 containerd[1496]: time="2025-05-08T03:06:19.305537179Z" level=info msg="Start snapshots syncer" May 8 03:06:19.305604 containerd[1496]: time="2025-05-08T03:06:19.305557659Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 8 03:06:19.305853 containerd[1496]: time="2025-05-08T03:06:19.305797299Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 8 03:06:19.305853 containerd[1496]: time="2025-05-08T03:06:19.305848739Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.305921499Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306027379Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306048339Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306064859Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306074779Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306085579Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 8 03:06:19.306098 containerd[1496]: time="2025-05-08T03:06:19.306096339Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 8 03:06:19.306213 containerd[1496]: time="2025-05-08T03:06:19.306106619Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 8 03:06:19.306213 containerd[1496]: time="2025-05-08T03:06:19.306134859Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 8 03:06:19.306213 containerd[1496]: time="2025-05-08T03:06:19.306145859Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 8 03:06:19.306213 containerd[1496]: time="2025-05-08T03:06:19.306155339Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 8 03:06:19.306213 containerd[1496]: time="2025-05-08T03:06:19.306200379Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306215419Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306225859Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306234379Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306242059Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306254499Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 8 03:06:19.306313 containerd[1496]: time="2025-05-08T03:06:19.306264699Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 8 03:06:19.306405 containerd[1496]: time="2025-05-08T03:06:19.306347019Z" level=info msg="runtime interface created" May 8 03:06:19.306405 containerd[1496]: time="2025-05-08T03:06:19.306352579Z" level=info msg="created NRI interface" May 8 03:06:19.306405 containerd[1496]: time="2025-05-08T03:06:19.306360459Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 8 03:06:19.306405 containerd[1496]: time="2025-05-08T03:06:19.306372899Z" level=info msg="Connect containerd service" May 8 03:06:19.306479 containerd[1496]: time="2025-05-08T03:06:19.306406219Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 8 03:06:19.307066 containerd[1496]: time="2025-05-08T03:06:19.307039019Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 8 03:06:19.317219 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 8 03:06:19.321000 systemd[1]: Started getty@tty1.service - Getty on tty1. May 8 03:06:19.324738 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. May 8 03:06:19.326040 systemd[1]: Reached target getty.target - Login Prompts. May 8 03:06:19.413242 containerd[1496]: time="2025-05-08T03:06:19.413137979Z" level=info msg="Start subscribing containerd event" May 8 03:06:19.413371 containerd[1496]: time="2025-05-08T03:06:19.413357139Z" level=info msg="Start recovering state" May 8 03:06:19.413436 containerd[1496]: time="2025-05-08T03:06:19.413406819Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 8 03:06:19.413477 containerd[1496]: time="2025-05-08T03:06:19.413463299Z" level=info msg=serving... address=/run/containerd/containerd.sock May 8 03:06:19.413561 containerd[1496]: time="2025-05-08T03:06:19.413547659Z" level=info msg="Start event monitor" May 8 03:06:19.413617 containerd[1496]: time="2025-05-08T03:06:19.413606499Z" level=info msg="Start cni network conf syncer for default" May 8 03:06:19.413674 containerd[1496]: time="2025-05-08T03:06:19.413662779Z" level=info msg="Start streaming server" May 8 03:06:19.413758 containerd[1496]: time="2025-05-08T03:06:19.413745539Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 8 03:06:19.413802 containerd[1496]: time="2025-05-08T03:06:19.413791499Z" level=info msg="runtime interface starting up..." May 8 03:06:19.413843 containerd[1496]: time="2025-05-08T03:06:19.413833619Z" level=info msg="starting plugins..." May 8 03:06:19.413895 containerd[1496]: time="2025-05-08T03:06:19.413884779Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 8 03:06:19.414083 containerd[1496]: time="2025-05-08T03:06:19.414066579Z" level=info msg="containerd successfully booted in 0.129958s" May 8 03:06:19.414163 systemd[1]: Started containerd.service - containerd container runtime. May 8 03:06:19.431096 tar[1490]: linux-arm64/LICENSE May 8 03:06:19.431161 tar[1490]: linux-arm64/README.md May 8 03:06:19.451957 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 8 03:06:20.659826 systemd-networkd[1424]: eth0: Gained IPv6LL May 8 03:06:20.663811 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 8 03:06:20.666437 systemd[1]: Reached target network-online.target - Network is Online. May 8 03:06:20.669155 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... May 8 03:06:20.671686 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:20.680710 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 8 03:06:20.696351 systemd[1]: coreos-metadata.service: Deactivated successfully. May 8 03:06:20.696564 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. May 8 03:06:20.698229 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 8 03:06:20.703309 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 8 03:06:21.162257 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:21.163847 systemd[1]: Reached target multi-user.target - Multi-User System. May 8 03:06:21.164991 systemd[1]: Startup finished in 2.152s (kernel) + 6.949s (initrd) + 3.978s (userspace) = 13.080s. May 8 03:06:21.165920 (kubelet)[1602]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 03:06:21.656551 kubelet[1602]: E0508 03:06:21.656436 1602 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 03:06:21.658289 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 03:06:21.658417 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 03:06:21.658955 systemd[1]: kubelet.service: Consumed 826ms CPU time, 241.3M memory peak. May 8 03:06:23.580362 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 8 03:06:23.581816 systemd[1]: Started sshd@0-10.0.0.7:22-10.0.0.1:36938.service - OpenSSH per-connection server daemon (10.0.0.1:36938). May 8 03:06:23.679935 sshd[1617]: Accepted publickey for core from 10.0.0.1 port 36938 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:23.684615 sshd-session[1617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:23.700574 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 8 03:06:23.701501 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 8 03:06:23.707061 systemd-logind[1481]: New session 1 of user core. May 8 03:06:23.724293 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 8 03:06:23.727298 systemd[1]: Starting user@500.service - User Manager for UID 500... May 8 03:06:23.744985 (systemd)[1621]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 8 03:06:23.747655 systemd-logind[1481]: New session c1 of user core. May 8 03:06:23.856350 systemd[1621]: Queued start job for default target default.target. May 8 03:06:23.866575 systemd[1621]: Created slice app.slice - User Application Slice. May 8 03:06:23.866750 systemd[1621]: Reached target paths.target - Paths. May 8 03:06:23.866848 systemd[1621]: Reached target timers.target - Timers. May 8 03:06:23.868200 systemd[1621]: Starting dbus.socket - D-Bus User Message Bus Socket... May 8 03:06:23.877077 systemd[1621]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 8 03:06:23.877136 systemd[1621]: Reached target sockets.target - Sockets. May 8 03:06:23.877170 systemd[1621]: Reached target basic.target - Basic System. May 8 03:06:23.877200 systemd[1621]: Reached target default.target - Main User Target. May 8 03:06:23.877224 systemd[1621]: Startup finished in 123ms. May 8 03:06:23.877428 systemd[1]: Started user@500.service - User Manager for UID 500. May 8 03:06:23.878801 systemd[1]: Started session-1.scope - Session 1 of User core. May 8 03:06:23.943922 systemd[1]: Started sshd@1-10.0.0.7:22-10.0.0.1:36942.service - OpenSSH per-connection server daemon (10.0.0.1:36942). May 8 03:06:24.001348 sshd[1632]: Accepted publickey for core from 10.0.0.1 port 36942 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.002613 sshd-session[1632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.006846 systemd-logind[1481]: New session 2 of user core. May 8 03:06:24.017903 systemd[1]: Started session-2.scope - Session 2 of User core. May 8 03:06:24.068240 sshd[1634]: Connection closed by 10.0.0.1 port 36942 May 8 03:06:24.068527 sshd-session[1632]: pam_unix(sshd:session): session closed for user core May 8 03:06:24.080843 systemd[1]: sshd@1-10.0.0.7:22-10.0.0.1:36942.service: Deactivated successfully. May 8 03:06:24.082299 systemd[1]: session-2.scope: Deactivated successfully. May 8 03:06:24.083565 systemd-logind[1481]: Session 2 logged out. Waiting for processes to exit. May 8 03:06:24.084612 systemd[1]: Started sshd@2-10.0.0.7:22-10.0.0.1:36958.service - OpenSSH per-connection server daemon (10.0.0.1:36958). May 8 03:06:24.085463 systemd-logind[1481]: Removed session 2. May 8 03:06:24.136558 sshd[1639]: Accepted publickey for core from 10.0.0.1 port 36958 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.137669 sshd-session[1639]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.142119 systemd-logind[1481]: New session 3 of user core. May 8 03:06:24.153861 systemd[1]: Started session-3.scope - Session 3 of User core. May 8 03:06:24.201335 sshd[1642]: Connection closed by 10.0.0.1 port 36958 May 8 03:06:24.201645 sshd-session[1639]: pam_unix(sshd:session): session closed for user core May 8 03:06:24.218809 systemd[1]: sshd@2-10.0.0.7:22-10.0.0.1:36958.service: Deactivated successfully. May 8 03:06:24.220151 systemd[1]: session-3.scope: Deactivated successfully. May 8 03:06:24.220892 systemd-logind[1481]: Session 3 logged out. Waiting for processes to exit. May 8 03:06:24.222680 systemd[1]: Started sshd@3-10.0.0.7:22-10.0.0.1:36968.service - OpenSSH per-connection server daemon (10.0.0.1:36968). May 8 03:06:24.223534 systemd-logind[1481]: Removed session 3. May 8 03:06:24.272070 sshd[1647]: Accepted publickey for core from 10.0.0.1 port 36968 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.273246 sshd-session[1647]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.277306 systemd-logind[1481]: New session 4 of user core. May 8 03:06:24.288880 systemd[1]: Started session-4.scope - Session 4 of User core. May 8 03:06:24.339765 sshd[1650]: Connection closed by 10.0.0.1 port 36968 May 8 03:06:24.340075 sshd-session[1647]: pam_unix(sshd:session): session closed for user core May 8 03:06:24.350815 systemd[1]: sshd@3-10.0.0.7:22-10.0.0.1:36968.service: Deactivated successfully. May 8 03:06:24.352229 systemd[1]: session-4.scope: Deactivated successfully. May 8 03:06:24.353688 systemd-logind[1481]: Session 4 logged out. Waiting for processes to exit. May 8 03:06:24.354963 systemd[1]: Started sshd@4-10.0.0.7:22-10.0.0.1:36984.service - OpenSSH per-connection server daemon (10.0.0.1:36984). May 8 03:06:24.356786 systemd-logind[1481]: Removed session 4. May 8 03:06:24.397338 sshd[1655]: Accepted publickey for core from 10.0.0.1 port 36984 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.398267 sshd-session[1655]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.401999 systemd-logind[1481]: New session 5 of user core. May 8 03:06:24.411919 systemd[1]: Started session-5.scope - Session 5 of User core. May 8 03:06:24.470769 sudo[1659]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 8 03:06:24.471036 sudo[1659]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 03:06:24.485528 sudo[1659]: pam_unix(sudo:session): session closed for user root May 8 03:06:24.487254 sshd[1658]: Connection closed by 10.0.0.1 port 36984 May 8 03:06:24.487154 sshd-session[1655]: pam_unix(sshd:session): session closed for user core May 8 03:06:24.506797 systemd[1]: sshd@4-10.0.0.7:22-10.0.0.1:36984.service: Deactivated successfully. May 8 03:06:24.508059 systemd[1]: session-5.scope: Deactivated successfully. May 8 03:06:24.508766 systemd-logind[1481]: Session 5 logged out. Waiting for processes to exit. May 8 03:06:24.510329 systemd[1]: Started sshd@5-10.0.0.7:22-10.0.0.1:37000.service - OpenSSH per-connection server daemon (10.0.0.1:37000). May 8 03:06:24.511046 systemd-logind[1481]: Removed session 5. May 8 03:06:24.561248 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 37000 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.562476 sshd-session[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.566333 systemd-logind[1481]: New session 6 of user core. May 8 03:06:24.584877 systemd[1]: Started session-6.scope - Session 6 of User core. May 8 03:06:24.635006 sudo[1669]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 8 03:06:24.635261 sudo[1669]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 03:06:24.638277 sudo[1669]: pam_unix(sudo:session): session closed for user root May 8 03:06:24.642487 sudo[1668]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 8 03:06:24.642773 sudo[1668]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 03:06:24.650369 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 8 03:06:24.680247 augenrules[1691]: No rules May 8 03:06:24.680858 systemd[1]: audit-rules.service: Deactivated successfully. May 8 03:06:24.681037 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 8 03:06:24.682212 sudo[1668]: pam_unix(sudo:session): session closed for user root May 8 03:06:24.683894 sshd[1667]: Connection closed by 10.0.0.1 port 37000 May 8 03:06:24.683687 sshd-session[1664]: pam_unix(sshd:session): session closed for user core May 8 03:06:24.694269 systemd[1]: sshd@5-10.0.0.7:22-10.0.0.1:37000.service: Deactivated successfully. May 8 03:06:24.695566 systemd[1]: session-6.scope: Deactivated successfully. May 8 03:06:24.696296 systemd-logind[1481]: Session 6 logged out. Waiting for processes to exit. May 8 03:06:24.697813 systemd[1]: Started sshd@6-10.0.0.7:22-10.0.0.1:37014.service - OpenSSH per-connection server daemon (10.0.0.1:37014). May 8 03:06:24.698488 systemd-logind[1481]: Removed session 6. May 8 03:06:24.741098 sshd[1699]: Accepted publickey for core from 10.0.0.1 port 37014 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:06:24.742252 sshd-session[1699]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:06:24.745711 systemd-logind[1481]: New session 7 of user core. May 8 03:06:24.757922 systemd[1]: Started session-7.scope - Session 7 of User core. May 8 03:06:24.808673 sudo[1703]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 8 03:06:24.808963 sudo[1703]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 03:06:25.164548 systemd[1]: Starting docker.service - Docker Application Container Engine... May 8 03:06:25.177061 (dockerd)[1723]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 8 03:06:25.431957 dockerd[1723]: time="2025-05-08T03:06:25.431842739Z" level=info msg="Starting up" May 8 03:06:25.433372 dockerd[1723]: time="2025-05-08T03:06:25.433337619Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 8 03:06:25.604062 dockerd[1723]: time="2025-05-08T03:06:25.604020819Z" level=info msg="Loading containers: start." May 8 03:06:25.615748 kernel: Initializing XFRM netlink socket May 8 03:06:25.807693 systemd-networkd[1424]: docker0: Link UP May 8 03:06:25.811393 dockerd[1723]: time="2025-05-08T03:06:25.811283819Z" level=info msg="Loading containers: done." May 8 03:06:25.822294 dockerd[1723]: time="2025-05-08T03:06:25.822243379Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 8 03:06:25.822419 dockerd[1723]: time="2025-05-08T03:06:25.822323019Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 8 03:06:25.822443 dockerd[1723]: time="2025-05-08T03:06:25.822432899Z" level=info msg="Initializing buildkit" May 8 03:06:25.869444 dockerd[1723]: time="2025-05-08T03:06:25.869398059Z" level=info msg="Completed buildkit initialization" May 8 03:06:25.875463 dockerd[1723]: time="2025-05-08T03:06:25.875420579Z" level=info msg="Daemon has completed initialization" May 8 03:06:25.875640 dockerd[1723]: time="2025-05-08T03:06:25.875584179Z" level=info msg="API listen on /run/docker.sock" May 8 03:06:25.875743 systemd[1]: Started docker.service - Docker Application Container Engine. May 8 03:06:26.949525 containerd[1496]: time="2025-05-08T03:06:26.949481699Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 8 03:06:27.726259 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount213087018.mount: Deactivated successfully. May 8 03:06:29.410314 containerd[1496]: time="2025-05-08T03:06:29.410228939Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:29.411030 containerd[1496]: time="2025-05-08T03:06:29.410993939Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.12: active requests=0, bytes read=29794152" May 8 03:06:29.411664 containerd[1496]: time="2025-05-08T03:06:29.411633059Z" level=info msg="ImageCreate event name:\"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:29.414016 containerd[1496]: time="2025-05-08T03:06:29.413983139Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:29.415113 containerd[1496]: time="2025-05-08T03:06:29.415080659Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.12\" with image id \"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.12\", repo digest \"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\", size \"29790950\" in 2.46555792s" May 8 03:06:29.415144 containerd[1496]: time="2025-05-08T03:06:29.415120019Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\"" May 8 03:06:29.430364 containerd[1496]: time="2025-05-08T03:06:29.430334459Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 8 03:06:31.664169 containerd[1496]: time="2025-05-08T03:06:31.664093099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:31.664823 containerd[1496]: time="2025-05-08T03:06:31.664787659Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.12: active requests=0, bytes read=26855552" May 8 03:06:31.665696 containerd[1496]: time="2025-05-08T03:06:31.665663379Z" level=info msg="ImageCreate event name:\"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:31.668509 containerd[1496]: time="2025-05-08T03:06:31.668476899Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:31.670104 containerd[1496]: time="2025-05-08T03:06:31.670070779Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.12\" with image id \"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.12\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\", size \"28297111\" in 2.23949708s" May 8 03:06:31.670132 containerd[1496]: time="2025-05-08T03:06:31.670105219Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\"" May 8 03:06:31.684513 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 8 03:06:31.686164 containerd[1496]: time="2025-05-08T03:06:31.685988779Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 8 03:06:31.686569 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:31.800807 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:31.804565 (kubelet)[2028]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 03:06:31.848736 kubelet[2028]: E0508 03:06:31.848678 2028 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 03:06:31.851928 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 03:06:31.852062 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 03:06:31.852385 systemd[1]: kubelet.service: Consumed 138ms CPU time, 99.5M memory peak. May 8 03:06:33.153761 containerd[1496]: time="2025-05-08T03:06:33.153710299Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:33.154670 containerd[1496]: time="2025-05-08T03:06:33.154174899Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.12: active requests=0, bytes read=16263947" May 8 03:06:33.155126 containerd[1496]: time="2025-05-08T03:06:33.155080499Z" level=info msg="ImageCreate event name:\"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:33.157809 containerd[1496]: time="2025-05-08T03:06:33.157781899Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:33.158853 containerd[1496]: time="2025-05-08T03:06:33.158682299Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.12\" with image id \"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.12\", repo digest \"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\", size \"17705524\" in 1.47265032s" May 8 03:06:33.158853 containerd[1496]: time="2025-05-08T03:06:33.158732619Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\"" May 8 03:06:33.173038 containerd[1496]: time="2025-05-08T03:06:33.173009499Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 8 03:06:34.407964 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1745924248.mount: Deactivated successfully. May 8 03:06:34.620603 containerd[1496]: time="2025-05-08T03:06:34.620542699Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:34.621270 containerd[1496]: time="2025-05-08T03:06:34.621246699Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.12: active requests=0, bytes read=25775707" May 8 03:06:34.621883 containerd[1496]: time="2025-05-08T03:06:34.621855539Z" level=info msg="ImageCreate event name:\"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:34.623799 containerd[1496]: time="2025-05-08T03:06:34.623773659Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:34.624363 containerd[1496]: time="2025-05-08T03:06:34.624335779Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.12\" with image id \"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\", repo tag \"registry.k8s.io/kube-proxy:v1.30.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\", size \"25774724\" in 1.45129004s" May 8 03:06:34.624419 containerd[1496]: time="2025-05-08T03:06:34.624367899Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\"" May 8 03:06:34.639527 containerd[1496]: time="2025-05-08T03:06:34.639492379Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 8 03:06:35.244660 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3569036417.mount: Deactivated successfully. May 8 03:06:36.301179 containerd[1496]: time="2025-05-08T03:06:36.300790739Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.301179 containerd[1496]: time="2025-05-08T03:06:36.301172059Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383" May 8 03:06:36.302285 containerd[1496]: time="2025-05-08T03:06:36.302262579Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.305466 containerd[1496]: time="2025-05-08T03:06:36.305406659Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.306417 containerd[1496]: time="2025-05-08T03:06:36.306384859Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.66686164s" May 8 03:06:36.306478 containerd[1496]: time="2025-05-08T03:06:36.306423859Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" May 8 03:06:36.321796 containerd[1496]: time="2025-05-08T03:06:36.321766819Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 8 03:06:36.746562 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount126091692.mount: Deactivated successfully. May 8 03:06:36.751426 containerd[1496]: time="2025-05-08T03:06:36.751105899Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.752022 containerd[1496]: time="2025-05-08T03:06:36.752000699Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823" May 8 03:06:36.753944 containerd[1496]: time="2025-05-08T03:06:36.753124659Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.755170 containerd[1496]: time="2025-05-08T03:06:36.755131059Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:36.756438 containerd[1496]: time="2025-05-08T03:06:36.756414619Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 434.51612ms" May 8 03:06:36.756537 containerd[1496]: time="2025-05-08T03:06:36.756522459Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" May 8 03:06:36.771415 containerd[1496]: time="2025-05-08T03:06:36.771370499Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 8 03:06:37.312288 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1852745130.mount: Deactivated successfully. May 8 03:06:40.097776 containerd[1496]: time="2025-05-08T03:06:40.097697859Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:40.098788 containerd[1496]: time="2025-05-08T03:06:40.098714179Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191474" May 8 03:06:40.099367 containerd[1496]: time="2025-05-08T03:06:40.099341539Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:40.102407 containerd[1496]: time="2025-05-08T03:06:40.102361659Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:06:40.104000 containerd[1496]: time="2025-05-08T03:06:40.103969619Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 3.33255384s" May 8 03:06:40.104176 containerd[1496]: time="2025-05-08T03:06:40.104082459Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" May 8 03:06:41.934650 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 8 03:06:41.936622 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:42.040563 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:42.043903 (kubelet)[2280]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 03:06:42.083233 kubelet[2280]: E0508 03:06:42.083116 2280 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 03:06:42.085464 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 03:06:42.085628 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 03:06:42.086359 systemd[1]: kubelet.service: Consumed 132ms CPU time, 96.8M memory peak. May 8 03:06:46.143764 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:46.144205 systemd[1]: kubelet.service: Consumed 132ms CPU time, 96.8M memory peak. May 8 03:06:46.146055 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:46.168804 systemd[1]: Reload requested from client PID 2295 ('systemctl') (unit session-7.scope)... May 8 03:06:46.168822 systemd[1]: Reloading... May 8 03:06:46.234932 zram_generator::config[2338]: No configuration found. May 8 03:06:46.437956 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 03:06:46.522211 systemd[1]: Reloading finished in 353 ms. May 8 03:06:46.576769 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:46.580498 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:46.580894 systemd[1]: kubelet.service: Deactivated successfully. May 8 03:06:46.581076 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:46.581107 systemd[1]: kubelet.service: Consumed 84ms CPU time, 82.4M memory peak. May 8 03:06:46.582446 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:46.687338 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:46.691980 (kubelet)[2384]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 8 03:06:46.736115 kubelet[2384]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 03:06:46.736115 kubelet[2384]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 03:06:46.736115 kubelet[2384]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 03:06:46.736450 kubelet[2384]: I0508 03:06:46.736137 2384 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 03:06:47.595377 kubelet[2384]: I0508 03:06:47.595333 2384 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 03:06:47.595377 kubelet[2384]: I0508 03:06:47.595361 2384 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 03:06:47.595605 kubelet[2384]: I0508 03:06:47.595559 2384 server.go:927] "Client rotation is on, will bootstrap in background" May 8 03:06:47.619484 kubelet[2384]: I0508 03:06:47.619439 2384 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 03:06:47.619605 kubelet[2384]: E0508 03:06:47.619553 2384 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.7:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.629731 kubelet[2384]: I0508 03:06:47.629686 2384 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 03:06:47.631007 kubelet[2384]: I0508 03:06:47.630952 2384 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 03:06:47.631193 kubelet[2384]: I0508 03:06:47.631003 2384 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 03:06:47.631288 kubelet[2384]: I0508 03:06:47.631256 2384 topology_manager.go:138] "Creating topology manager with none policy" May 8 03:06:47.631288 kubelet[2384]: I0508 03:06:47.631267 2384 container_manager_linux.go:301] "Creating device plugin manager" May 8 03:06:47.631551 kubelet[2384]: I0508 03:06:47.631524 2384 state_mem.go:36] "Initialized new in-memory state store" May 8 03:06:47.635967 kubelet[2384]: I0508 03:06:47.635234 2384 kubelet.go:400] "Attempting to sync node with API server" May 8 03:06:47.635967 kubelet[2384]: I0508 03:06:47.635258 2384 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 03:06:47.635967 kubelet[2384]: I0508 03:06:47.635706 2384 kubelet.go:312] "Adding apiserver pod source" May 8 03:06:47.635967 kubelet[2384]: I0508 03:06:47.635870 2384 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 03:06:47.635967 kubelet[2384]: W0508 03:06:47.635926 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.7:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.635967 kubelet[2384]: E0508 03:06:47.635968 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.7:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.636338 kubelet[2384]: W0508 03:06:47.636305 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.7:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.636338 kubelet[2384]: E0508 03:06:47.636336 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.7:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.637112 2384 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.637458 2384 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 03:06:47.640861 kubelet[2384]: W0508 03:06:47.637558 2384 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.638359 2384 server.go:1264] "Started kubelet" May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.639352 2384 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.639483 2384 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.639834 2384 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 03:06:47.640861 kubelet[2384]: I0508 03:06:47.640359 2384 server.go:455] "Adding debug handlers to kubelet server" May 8 03:06:47.641051 kubelet[2384]: I0508 03:06:47.641037 2384 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 03:06:47.643289 kubelet[2384]: E0508 03:06:47.643085 2384 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.7:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.7:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183d6e6451d53593 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-08 03:06:47.638332819 +0000 UTC m=+0.943114721,LastTimestamp:2025-05-08 03:06:47.638332819 +0000 UTC m=+0.943114721,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 8 03:06:47.643397 kubelet[2384]: E0508 03:06:47.643319 2384 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 03:06:47.643608 kubelet[2384]: I0508 03:06:47.643454 2384 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 03:06:47.643608 kubelet[2384]: I0508 03:06:47.643555 2384 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 03:06:47.643704 kubelet[2384]: E0508 03:06:47.643661 2384 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.7:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.7:6443: connect: connection refused" interval="200ms" May 8 03:06:47.644070 kubelet[2384]: W0508 03:06:47.644022 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.7:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.644070 kubelet[2384]: E0508 03:06:47.644068 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.7:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.644530 kubelet[2384]: I0508 03:06:47.644502 2384 factory.go:221] Registration of the systemd container factory successfully May 8 03:06:47.644595 kubelet[2384]: I0508 03:06:47.644572 2384 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 03:06:47.644944 kubelet[2384]: I0508 03:06:47.644829 2384 reconciler.go:26] "Reconciler: start to sync state" May 8 03:06:47.645597 kubelet[2384]: E0508 03:06:47.645571 2384 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 8 03:06:47.646761 kubelet[2384]: I0508 03:06:47.645831 2384 factory.go:221] Registration of the containerd container factory successfully May 8 03:06:47.658237 kubelet[2384]: I0508 03:06:47.658191 2384 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 03:06:47.659499 kubelet[2384]: I0508 03:06:47.659459 2384 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 03:06:47.659499 kubelet[2384]: I0508 03:06:47.659499 2384 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 03:06:47.659593 kubelet[2384]: I0508 03:06:47.659516 2384 kubelet.go:2337] "Starting kubelet main sync loop" May 8 03:06:47.659593 kubelet[2384]: E0508 03:06:47.659557 2384 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 03:06:47.660555 kubelet[2384]: W0508 03:06:47.660504 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.7:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.660555 kubelet[2384]: E0508 03:06:47.660546 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.7:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:47.663249 kubelet[2384]: I0508 03:06:47.663142 2384 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 03:06:47.663249 kubelet[2384]: I0508 03:06:47.663156 2384 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 03:06:47.663249 kubelet[2384]: I0508 03:06:47.663173 2384 state_mem.go:36] "Initialized new in-memory state store" May 8 03:06:47.743954 kubelet[2384]: I0508 03:06:47.743837 2384 policy_none.go:49] "None policy: Start" May 8 03:06:47.745099 kubelet[2384]: I0508 03:06:47.745027 2384 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:47.745568 kubelet[2384]: I0508 03:06:47.745254 2384 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 03:06:47.745568 kubelet[2384]: I0508 03:06:47.745279 2384 state_mem.go:35] "Initializing new in-memory state store" May 8 03:06:47.745568 kubelet[2384]: E0508 03:06:47.745320 2384 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.7:6443/api/v1/nodes\": dial tcp 10.0.0.7:6443: connect: connection refused" node="localhost" May 8 03:06:47.750966 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 8 03:06:47.759927 kubelet[2384]: E0508 03:06:47.759884 2384 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 8 03:06:47.763669 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 8 03:06:47.766680 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 8 03:06:47.786321 kubelet[2384]: I0508 03:06:47.785852 2384 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 03:06:47.786321 kubelet[2384]: I0508 03:06:47.786075 2384 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 03:06:47.786321 kubelet[2384]: I0508 03:06:47.786177 2384 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 03:06:47.788409 kubelet[2384]: E0508 03:06:47.788383 2384 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 8 03:06:47.844328 kubelet[2384]: E0508 03:06:47.844269 2384 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.7:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.7:6443: connect: connection refused" interval="400ms" May 8 03:06:47.946700 kubelet[2384]: I0508 03:06:47.946648 2384 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:47.946986 kubelet[2384]: E0508 03:06:47.946962 2384 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.7:6443/api/v1/nodes\": dial tcp 10.0.0.7:6443: connect: connection refused" node="localhost" May 8 03:06:47.960251 kubelet[2384]: I0508 03:06:47.960183 2384 topology_manager.go:215] "Topology Admit Handler" podUID="f11c1a596cc34caf0a7a3c821540240a" podNamespace="kube-system" podName="kube-apiserver-localhost" May 8 03:06:47.961280 kubelet[2384]: I0508 03:06:47.961243 2384 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 8 03:06:47.963198 kubelet[2384]: I0508 03:06:47.962260 2384 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 8 03:06:47.973045 systemd[1]: Created slice kubepods-burstable-podf11c1a596cc34caf0a7a3c821540240a.slice - libcontainer container kubepods-burstable-podf11c1a596cc34caf0a7a3c821540240a.slice. May 8 03:06:47.986186 systemd[1]: Created slice kubepods-burstable-podb20b39a8540dba87b5883a6f0f602dba.slice - libcontainer container kubepods-burstable-podb20b39a8540dba87b5883a6f0f602dba.slice. May 8 03:06:47.998493 systemd[1]: Created slice kubepods-burstable-pod6ece95f10dbffa04b25ec3439a115512.slice - libcontainer container kubepods-burstable-pod6ece95f10dbffa04b25ec3439a115512.slice. May 8 03:06:48.046129 kubelet[2384]: I0508 03:06:48.046080 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:48.046129 kubelet[2384]: I0508 03:06:48.046120 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:48.046288 kubelet[2384]: I0508 03:06:48.046151 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:48.046288 kubelet[2384]: I0508 03:06:48.046167 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:48.046288 kubelet[2384]: I0508 03:06:48.046187 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:48.046288 kubelet[2384]: I0508 03:06:48.046207 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:48.046288 kubelet[2384]: I0508 03:06:48.046222 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:48.046382 kubelet[2384]: I0508 03:06:48.046239 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:48.046382 kubelet[2384]: I0508 03:06:48.046254 2384 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 8 03:06:48.245497 kubelet[2384]: E0508 03:06:48.245384 2384 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.7:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.7:6443: connect: connection refused" interval="800ms" May 8 03:06:48.284875 kubelet[2384]: E0508 03:06:48.284789 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.285536 containerd[1496]: time="2025-05-08T03:06:48.285408979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:f11c1a596cc34caf0a7a3c821540240a,Namespace:kube-system,Attempt:0,}" May 8 03:06:48.297036 kubelet[2384]: E0508 03:06:48.296984 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.297433 containerd[1496]: time="2025-05-08T03:06:48.297401099Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,}" May 8 03:06:48.300766 kubelet[2384]: E0508 03:06:48.300701 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.301286 containerd[1496]: time="2025-05-08T03:06:48.301237299Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,}" May 8 03:06:48.348892 kubelet[2384]: I0508 03:06:48.348815 2384 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:48.349167 kubelet[2384]: E0508 03:06:48.349142 2384 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.7:6443/api/v1/nodes\": dial tcp 10.0.0.7:6443: connect: connection refused" node="localhost" May 8 03:06:48.459155 kubelet[2384]: W0508 03:06:48.459092 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.7:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.459155 kubelet[2384]: E0508 03:06:48.459161 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.7:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.562189 kubelet[2384]: W0508 03:06:48.562038 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.7:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.562189 kubelet[2384]: E0508 03:06:48.562098 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.7:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.619676 kubelet[2384]: W0508 03:06:48.619571 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.7:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.619676 kubelet[2384]: E0508 03:06:48.619652 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.7:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.680885 kubelet[2384]: W0508 03:06:48.680823 2384 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.7:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.680885 kubelet[2384]: E0508 03:06:48.680887 2384 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.7:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.7:6443: connect: connection refused May 8 03:06:48.851077 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3933046646.mount: Deactivated successfully. May 8 03:06:48.855714 containerd[1496]: time="2025-05-08T03:06:48.855645499Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 03:06:48.858054 containerd[1496]: time="2025-05-08T03:06:48.858007939Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" May 8 03:06:48.858670 containerd[1496]: time="2025-05-08T03:06:48.858639539Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 03:06:48.859945 containerd[1496]: time="2025-05-08T03:06:48.859921139Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 8 03:06:48.861030 containerd[1496]: time="2025-05-08T03:06:48.860999419Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 03:06:48.862086 containerd[1496]: time="2025-05-08T03:06:48.862052619Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 03:06:48.863262 containerd[1496]: time="2025-05-08T03:06:48.863212779Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 8 03:06:48.864304 containerd[1496]: time="2025-05-08T03:06:48.864238019Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 03:06:48.866157 containerd[1496]: time="2025-05-08T03:06:48.866107339Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 578.55344ms" May 8 03:06:48.869268 containerd[1496]: time="2025-05-08T03:06:48.867901979Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 569.00228ms" May 8 03:06:48.869803 containerd[1496]: time="2025-05-08T03:06:48.869717179Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 566.87924ms" May 8 03:06:48.897616 containerd[1496]: time="2025-05-08T03:06:48.897564219Z" level=info msg="connecting to shim 37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a" address="unix:///run/containerd/s/11c50ad6eb7ef658a2d6bad781cc05f62b345a665297e7bf4ed61302efaa65a5" namespace=k8s.io protocol=ttrpc version=3 May 8 03:06:48.899823 containerd[1496]: time="2025-05-08T03:06:48.899759539Z" level=info msg="connecting to shim de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b" address="unix:///run/containerd/s/556605f4670f775b450127f1a896ad17dee81cde4f7ec7f32b0b649a9cf00200" namespace=k8s.io protocol=ttrpc version=3 May 8 03:06:48.902529 containerd[1496]: time="2025-05-08T03:06:48.902483659Z" level=info msg="connecting to shim ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b" address="unix:///run/containerd/s/3136fd55803a1e60c5f39e223f64e8188d65eac54287181cc32fa157075f8dae" namespace=k8s.io protocol=ttrpc version=3 May 8 03:06:48.925917 systemd[1]: Started cri-containerd-37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a.scope - libcontainer container 37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a. May 8 03:06:48.929964 systemd[1]: Started cri-containerd-ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b.scope - libcontainer container ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b. May 8 03:06:48.931259 systemd[1]: Started cri-containerd-de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b.scope - libcontainer container de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b. May 8 03:06:48.970972 containerd[1496]: time="2025-05-08T03:06:48.970930979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,} returns sandbox id \"ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b\"" May 8 03:06:48.972103 kubelet[2384]: E0508 03:06:48.972009 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.972499 containerd[1496]: time="2025-05-08T03:06:48.972434619Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:f11c1a596cc34caf0a7a3c821540240a,Namespace:kube-system,Attempt:0,} returns sandbox id \"37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a\"" May 8 03:06:48.973592 containerd[1496]: time="2025-05-08T03:06:48.973545459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,} returns sandbox id \"de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b\"" May 8 03:06:48.973865 kubelet[2384]: E0508 03:06:48.973565 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.974204 kubelet[2384]: E0508 03:06:48.974138 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:48.975997 containerd[1496]: time="2025-05-08T03:06:48.975741299Z" level=info msg="CreateContainer within sandbox \"ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 8 03:06:48.977281 containerd[1496]: time="2025-05-08T03:06:48.976827179Z" level=info msg="CreateContainer within sandbox \"de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 8 03:06:48.977281 containerd[1496]: time="2025-05-08T03:06:48.976980979Z" level=info msg="CreateContainer within sandbox \"37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 8 03:06:48.985832 containerd[1496]: time="2025-05-08T03:06:48.985801339Z" level=info msg="Container 66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8: CDI devices from CRI Config.CDIDevices: []" May 8 03:06:48.987148 containerd[1496]: time="2025-05-08T03:06:48.987121659Z" level=info msg="Container c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749: CDI devices from CRI Config.CDIDevices: []" May 8 03:06:48.990023 containerd[1496]: time="2025-05-08T03:06:48.989984419Z" level=info msg="Container c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4: CDI devices from CRI Config.CDIDevices: []" May 8 03:06:48.994349 containerd[1496]: time="2025-05-08T03:06:48.994308139Z" level=info msg="CreateContainer within sandbox \"ceadebbbd47105ce7cecc0960f8c0254a673b7e49b03258e76462b4a8e9bcd4b\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8\"" May 8 03:06:48.995239 containerd[1496]: time="2025-05-08T03:06:48.995188939Z" level=info msg="StartContainer for \"66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8\"" May 8 03:06:48.996335 containerd[1496]: time="2025-05-08T03:06:48.996261979Z" level=info msg="connecting to shim 66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8" address="unix:///run/containerd/s/3136fd55803a1e60c5f39e223f64e8188d65eac54287181cc32fa157075f8dae" protocol=ttrpc version=3 May 8 03:06:48.997331 containerd[1496]: time="2025-05-08T03:06:48.997295379Z" level=info msg="CreateContainer within sandbox \"de9a53daa5fbea294c27baf8d10aa66d84548ce59d87389a6056c6475b2ac31b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749\"" May 8 03:06:48.997762 containerd[1496]: time="2025-05-08T03:06:48.997710459Z" level=info msg="StartContainer for \"c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749\"" May 8 03:06:48.998905 containerd[1496]: time="2025-05-08T03:06:48.998813659Z" level=info msg="connecting to shim c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749" address="unix:///run/containerd/s/556605f4670f775b450127f1a896ad17dee81cde4f7ec7f32b0b649a9cf00200" protocol=ttrpc version=3 May 8 03:06:49.000553 containerd[1496]: time="2025-05-08T03:06:49.000513059Z" level=info msg="CreateContainer within sandbox \"37dd4e75a3dac0a1fbb18738f6d2810f59df0342ff0c704361f4c6f88722505a\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4\"" May 8 03:06:49.001104 containerd[1496]: time="2025-05-08T03:06:49.001077179Z" level=info msg="StartContainer for \"c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4\"" May 8 03:06:49.002081 containerd[1496]: time="2025-05-08T03:06:49.002045899Z" level=info msg="connecting to shim c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4" address="unix:///run/containerd/s/11c50ad6eb7ef658a2d6bad781cc05f62b345a665297e7bf4ed61302efaa65a5" protocol=ttrpc version=3 May 8 03:06:49.024909 systemd[1]: Started cri-containerd-66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8.scope - libcontainer container 66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8. May 8 03:06:49.026116 systemd[1]: Started cri-containerd-c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749.scope - libcontainer container c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749. May 8 03:06:49.029626 systemd[1]: Started cri-containerd-c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4.scope - libcontainer container c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4. May 8 03:06:49.046601 kubelet[2384]: E0508 03:06:49.046551 2384 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.7:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.7:6443: connect: connection refused" interval="1.6s" May 8 03:06:49.074598 containerd[1496]: time="2025-05-08T03:06:49.070086179Z" level=info msg="StartContainer for \"66a19b45f47156bfc54d16da7cbd73d1dda7913742ec3ba89e608997871d28c8\" returns successfully" May 8 03:06:49.111478 containerd[1496]: time="2025-05-08T03:06:49.109023139Z" level=info msg="StartContainer for \"c69e2beab601721121c3d37317e6a9a9a549dbb4270d474ca61296dd36446749\" returns successfully" May 8 03:06:49.111478 containerd[1496]: time="2025-05-08T03:06:49.109512779Z" level=info msg="StartContainer for \"c3ba741470cfc257acb13a24d6de7d18c3b57149ceb7e8dfaa7a5f60965742a4\" returns successfully" May 8 03:06:49.152949 kubelet[2384]: I0508 03:06:49.150708 2384 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:49.152949 kubelet[2384]: E0508 03:06:49.151086 2384 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.7:6443/api/v1/nodes\": dial tcp 10.0.0.7:6443: connect: connection refused" node="localhost" May 8 03:06:49.675078 kubelet[2384]: E0508 03:06:49.674717 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:49.677076 kubelet[2384]: E0508 03:06:49.677035 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:49.679312 kubelet[2384]: E0508 03:06:49.679208 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:50.682901 kubelet[2384]: E0508 03:06:50.682778 2384 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:50.753101 kubelet[2384]: I0508 03:06:50.753077 2384 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:51.053692 kubelet[2384]: E0508 03:06:51.053627 2384 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" May 8 03:06:51.136981 kubelet[2384]: I0508 03:06:51.136934 2384 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 8 03:06:51.194961 kubelet[2384]: E0508 03:06:51.194852 2384 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.183d6e6451d53593 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-08 03:06:47.638332819 +0000 UTC m=+0.943114721,LastTimestamp:2025-05-08 03:06:47.638332819 +0000 UTC m=+0.943114721,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 8 03:06:51.639445 kubelet[2384]: I0508 03:06:51.639165 2384 apiserver.go:52] "Watching apiserver" May 8 03:06:51.644407 kubelet[2384]: I0508 03:06:51.644384 2384 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 8 03:06:53.147439 systemd[1]: Reload requested from client PID 2661 ('systemctl') (unit session-7.scope)... May 8 03:06:53.147457 systemd[1]: Reloading... May 8 03:06:53.215820 zram_generator::config[2706]: No configuration found. May 8 03:06:53.282903 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 03:06:53.382102 systemd[1]: Reloading finished in 234 ms. May 8 03:06:53.402306 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:53.419136 systemd[1]: kubelet.service: Deactivated successfully. May 8 03:06:53.419477 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:53.419609 systemd[1]: kubelet.service: Consumed 1.334s CPU time, 114.1M memory peak. May 8 03:06:53.422239 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 03:06:53.542801 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 03:06:53.548533 (kubelet)[2746]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 8 03:06:53.593053 kubelet[2746]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 03:06:53.593053 kubelet[2746]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 03:06:53.593053 kubelet[2746]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 03:06:53.593379 kubelet[2746]: I0508 03:06:53.593088 2746 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 03:06:53.597762 kubelet[2746]: I0508 03:06:53.596919 2746 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 03:06:53.597762 kubelet[2746]: I0508 03:06:53.596942 2746 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 03:06:53.597762 kubelet[2746]: I0508 03:06:53.597092 2746 server.go:927] "Client rotation is on, will bootstrap in background" May 8 03:06:53.598352 kubelet[2746]: I0508 03:06:53.598337 2746 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 8 03:06:53.599507 kubelet[2746]: I0508 03:06:53.599467 2746 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 03:06:53.606921 kubelet[2746]: I0508 03:06:53.606864 2746 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 03:06:53.607107 kubelet[2746]: I0508 03:06:53.607074 2746 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 03:06:53.607248 kubelet[2746]: I0508 03:06:53.607097 2746 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 03:06:53.607318 kubelet[2746]: I0508 03:06:53.607249 2746 topology_manager.go:138] "Creating topology manager with none policy" May 8 03:06:53.607318 kubelet[2746]: I0508 03:06:53.607258 2746 container_manager_linux.go:301] "Creating device plugin manager" May 8 03:06:53.607318 kubelet[2746]: I0508 03:06:53.607287 2746 state_mem.go:36] "Initialized new in-memory state store" May 8 03:06:53.607388 kubelet[2746]: I0508 03:06:53.607376 2746 kubelet.go:400] "Attempting to sync node with API server" May 8 03:06:53.607417 kubelet[2746]: I0508 03:06:53.607388 2746 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 03:06:53.607417 kubelet[2746]: I0508 03:06:53.607408 2746 kubelet.go:312] "Adding apiserver pod source" May 8 03:06:53.607456 kubelet[2746]: I0508 03:06:53.607422 2746 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 03:06:53.608736 kubelet[2746]: I0508 03:06:53.608463 2746 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 8 03:06:53.608885 kubelet[2746]: I0508 03:06:53.608855 2746 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 03:06:53.611216 kubelet[2746]: I0508 03:06:53.609327 2746 server.go:1264] "Started kubelet" May 8 03:06:53.611216 kubelet[2746]: I0508 03:06:53.609620 2746 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 03:06:53.611216 kubelet[2746]: I0508 03:06:53.609656 2746 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 03:06:53.611216 kubelet[2746]: I0508 03:06:53.609952 2746 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 03:06:53.611216 kubelet[2746]: I0508 03:06:53.610518 2746 server.go:455] "Adding debug handlers to kubelet server" May 8 03:06:53.614224 kubelet[2746]: I0508 03:06:53.611919 2746 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 03:06:53.614926 kubelet[2746]: E0508 03:06:53.614904 2746 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 03:06:53.614985 kubelet[2746]: I0508 03:06:53.614947 2746 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 03:06:53.615038 kubelet[2746]: I0508 03:06:53.615020 2746 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 03:06:53.615152 kubelet[2746]: I0508 03:06:53.615141 2746 reconciler.go:26] "Reconciler: start to sync state" May 8 03:06:53.624302 kubelet[2746]: I0508 03:06:53.624197 2746 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 03:06:53.627102 kubelet[2746]: I0508 03:06:53.627026 2746 factory.go:221] Registration of the containerd container factory successfully May 8 03:06:53.627102 kubelet[2746]: I0508 03:06:53.627046 2746 factory.go:221] Registration of the systemd container factory successfully May 8 03:06:53.630522 kubelet[2746]: I0508 03:06:53.630400 2746 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 03:06:53.632872 kubelet[2746]: I0508 03:06:53.632855 2746 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 03:06:53.632981 kubelet[2746]: I0508 03:06:53.632970 2746 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 03:06:53.633037 kubelet[2746]: I0508 03:06:53.633028 2746 kubelet.go:2337] "Starting kubelet main sync loop" May 8 03:06:53.633118 kubelet[2746]: E0508 03:06:53.633103 2746 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665765 2746 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665780 2746 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665799 2746 state_mem.go:36] "Initialized new in-memory state store" May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665933 2746 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665944 2746 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 8 03:06:53.666103 kubelet[2746]: I0508 03:06:53.665961 2746 policy_none.go:49] "None policy: Start" May 8 03:06:53.666980 kubelet[2746]: I0508 03:06:53.666925 2746 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 03:06:53.666980 kubelet[2746]: I0508 03:06:53.666952 2746 state_mem.go:35] "Initializing new in-memory state store" May 8 03:06:53.667083 kubelet[2746]: I0508 03:06:53.667069 2746 state_mem.go:75] "Updated machine memory state" May 8 03:06:53.671340 kubelet[2746]: I0508 03:06:53.671316 2746 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 03:06:53.671742 kubelet[2746]: I0508 03:06:53.671574 2746 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 03:06:53.671742 kubelet[2746]: I0508 03:06:53.671680 2746 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 03:06:53.718225 kubelet[2746]: I0508 03:06:53.718195 2746 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 03:06:53.726717 kubelet[2746]: I0508 03:06:53.726562 2746 kubelet_node_status.go:112] "Node was previously registered" node="localhost" May 8 03:06:53.726717 kubelet[2746]: I0508 03:06:53.726627 2746 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 8 03:06:53.733236 kubelet[2746]: I0508 03:06:53.733198 2746 topology_manager.go:215] "Topology Admit Handler" podUID="f11c1a596cc34caf0a7a3c821540240a" podNamespace="kube-system" podName="kube-apiserver-localhost" May 8 03:06:53.733338 kubelet[2746]: I0508 03:06:53.733309 2746 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 8 03:06:53.734000 kubelet[2746]: I0508 03:06:53.733346 2746 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 8 03:06:53.916965 kubelet[2746]: I0508 03:06:53.916770 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:53.916965 kubelet[2746]: I0508 03:06:53.916812 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:53.916965 kubelet[2746]: I0508 03:06:53.916835 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:53.916965 kubelet[2746]: I0508 03:06:53.916883 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 8 03:06:53.916965 kubelet[2746]: I0508 03:06:53.916937 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:53.917339 kubelet[2746]: I0508 03:06:53.916958 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:53.917339 kubelet[2746]: I0508 03:06:53.916978 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:53.917339 kubelet[2746]: I0508 03:06:53.916995 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f11c1a596cc34caf0a7a3c821540240a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"f11c1a596cc34caf0a7a3c821540240a\") " pod="kube-system/kube-apiserver-localhost" May 8 03:06:53.917339 kubelet[2746]: I0508 03:06:53.917011 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 03:06:54.064330 kubelet[2746]: E0508 03:06:54.064286 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.064330 kubelet[2746]: E0508 03:06:54.064316 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.067045 kubelet[2746]: E0508 03:06:54.067010 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.208982 sudo[2784]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 8 03:06:54.209259 sudo[2784]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 8 03:06:54.608523 kubelet[2746]: I0508 03:06:54.608198 2746 apiserver.go:52] "Watching apiserver" May 8 03:06:54.616073 kubelet[2746]: I0508 03:06:54.616032 2746 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 8 03:06:54.648634 sudo[2784]: pam_unix(sudo:session): session closed for user root May 8 03:06:54.652638 kubelet[2746]: E0508 03:06:54.652592 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.654027 kubelet[2746]: E0508 03:06:54.654008 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.661059 kubelet[2746]: E0508 03:06:54.660475 2746 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 8 03:06:54.661059 kubelet[2746]: E0508 03:06:54.660879 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:54.686411 kubelet[2746]: I0508 03:06:54.686308 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.686246652 podStartE2EDuration="1.686246652s" podCreationTimestamp="2025-05-08 03:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:06:54.686190172 +0000 UTC m=+1.134494355" watchObservedRunningTime="2025-05-08 03:06:54.686246652 +0000 UTC m=+1.134550835" May 8 03:06:54.686571 kubelet[2746]: I0508 03:06:54.686532 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.68652505 podStartE2EDuration="1.68652505s" podCreationTimestamp="2025-05-08 03:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:06:54.673306163 +0000 UTC m=+1.121610346" watchObservedRunningTime="2025-05-08 03:06:54.68652505 +0000 UTC m=+1.134829193" May 8 03:06:54.696194 kubelet[2746]: I0508 03:06:54.696092 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.696080678 podStartE2EDuration="1.696080678s" podCreationTimestamp="2025-05-08 03:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:06:54.696037358 +0000 UTC m=+1.144341541" watchObservedRunningTime="2025-05-08 03:06:54.696080678 +0000 UTC m=+1.144384861" May 8 03:06:55.653253 kubelet[2746]: E0508 03:06:55.653225 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:06:56.282865 sudo[1703]: pam_unix(sudo:session): session closed for user root May 8 03:06:56.284226 sshd[1702]: Connection closed by 10.0.0.1 port 37014 May 8 03:06:56.284657 sshd-session[1699]: pam_unix(sshd:session): session closed for user core May 8 03:06:56.288321 systemd-logind[1481]: Session 7 logged out. Waiting for processes to exit. May 8 03:06:56.288541 systemd[1]: sshd@6-10.0.0.7:22-10.0.0.1:37014.service: Deactivated successfully. May 8 03:06:56.290851 systemd[1]: session-7.scope: Deactivated successfully. May 8 03:06:56.291014 systemd[1]: session-7.scope: Consumed 8.131s CPU time, 285.1M memory peak. May 8 03:06:56.292713 systemd-logind[1481]: Removed session 7. May 8 03:06:58.289215 kubelet[2746]: E0508 03:06:58.289174 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:00.255570 kubelet[2746]: E0508 03:07:00.255536 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:00.662786 kubelet[2746]: E0508 03:07:00.662640 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:01.981330 kubelet[2746]: E0508 03:07:01.981187 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:02.665703 kubelet[2746]: E0508 03:07:02.665637 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:04.092306 update_engine[1483]: I20250508 03:07:04.092229 1483 update_attempter.cc:509] Updating boot flags... May 8 03:07:04.131812 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2834) May 8 03:07:04.178799 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2834) May 8 03:07:04.212879 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2834) May 8 03:07:07.865145 kubelet[2746]: I0508 03:07:07.865106 2746 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 8 03:07:07.880133 containerd[1496]: time="2025-05-08T03:07:07.877968997Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 8 03:07:07.880554 kubelet[2746]: I0508 03:07:07.878331 2746 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 8 03:07:07.880554 kubelet[2746]: I0508 03:07:07.879835 2746 topology_manager.go:215] "Topology Admit Handler" podUID="5a10c27c-9aa4-48fa-a24d-e93bd176b104" podNamespace="kube-system" podName="kube-proxy-v2vld" May 8 03:07:07.880554 kubelet[2746]: I0508 03:07:07.880009 2746 topology_manager.go:215] "Topology Admit Handler" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" podNamespace="kube-system" podName="cilium-tk6vz" May 8 03:07:07.895206 systemd[1]: Created slice kubepods-besteffort-pod5a10c27c_9aa4_48fa_a24d_e93bd176b104.slice - libcontainer container kubepods-besteffort-pod5a10c27c_9aa4_48fa_a24d_e93bd176b104.slice. May 8 03:07:07.908353 kubelet[2746]: I0508 03:07:07.908319 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-lib-modules\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908529 kubelet[2746]: I0508 03:07:07.908515 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-clustermesh-secrets\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908616 kubelet[2746]: I0508 03:07:07.908604 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-cgroup\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908742 kubelet[2746]: I0508 03:07:07.908683 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-xtables-lock\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908742 kubelet[2746]: I0508 03:07:07.908705 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-kernel\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908838 kubelet[2746]: I0508 03:07:07.908827 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/5a10c27c-9aa4-48fa-a24d-e93bd176b104-kube-proxy\") pod \"kube-proxy-v2vld\" (UID: \"5a10c27c-9aa4-48fa-a24d-e93bd176b104\") " pod="kube-system/kube-proxy-v2vld" May 8 03:07:07.908975 kubelet[2746]: I0508 03:07:07.908906 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-run\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908975 kubelet[2746]: I0508 03:07:07.908927 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-net\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908975 kubelet[2746]: I0508 03:07:07.908944 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-bpf-maps\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.908975 kubelet[2746]: I0508 03:07:07.908959 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hostproc\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.909156 kubelet[2746]: I0508 03:07:07.909095 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cni-path\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.909156 kubelet[2746]: I0508 03:07:07.909116 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-config-path\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.909293 kubelet[2746]: I0508 03:07:07.909134 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5a10c27c-9aa4-48fa-a24d-e93bd176b104-xtables-lock\") pod \"kube-proxy-v2vld\" (UID: \"5a10c27c-9aa4-48fa-a24d-e93bd176b104\") " pod="kube-system/kube-proxy-v2vld" May 8 03:07:07.909293 kubelet[2746]: I0508 03:07:07.909237 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a10c27c-9aa4-48fa-a24d-e93bd176b104-lib-modules\") pod \"kube-proxy-v2vld\" (UID: \"5a10c27c-9aa4-48fa-a24d-e93bd176b104\") " pod="kube-system/kube-proxy-v2vld" May 8 03:07:07.909293 kubelet[2746]: I0508 03:07:07.909257 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-etc-cni-netd\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.909293 kubelet[2746]: I0508 03:07:07.909273 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btqdf\" (UniqueName: \"kubernetes.io/projected/5a10c27c-9aa4-48fa-a24d-e93bd176b104-kube-api-access-btqdf\") pod \"kube-proxy-v2vld\" (UID: \"5a10c27c-9aa4-48fa-a24d-e93bd176b104\") " pod="kube-system/kube-proxy-v2vld" May 8 03:07:07.909550 kubelet[2746]: I0508 03:07:07.909409 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hubble-tls\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.909550 kubelet[2746]: I0508 03:07:07.909433 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsr9b\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-kube-api-access-jsr9b\") pod \"cilium-tk6vz\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " pod="kube-system/cilium-tk6vz" May 8 03:07:07.912408 systemd[1]: Created slice kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice - libcontainer container kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice. May 8 03:07:07.938138 kubelet[2746]: I0508 03:07:07.938088 2746 topology_manager.go:215] "Topology Admit Handler" podUID="64bda319-ec49-409f-8ae5-603c3dfaf0b9" podNamespace="kube-system" podName="cilium-operator-599987898-jjdtq" May 8 03:07:07.959500 systemd[1]: Created slice kubepods-besteffort-pod64bda319_ec49_409f_8ae5_603c3dfaf0b9.slice - libcontainer container kubepods-besteffort-pod64bda319_ec49_409f_8ae5_603c3dfaf0b9.slice. May 8 03:07:08.009970 kubelet[2746]: I0508 03:07:08.009906 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbthz\" (UniqueName: \"kubernetes.io/projected/64bda319-ec49-409f-8ae5-603c3dfaf0b9-kube-api-access-zbthz\") pod \"cilium-operator-599987898-jjdtq\" (UID: \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\") " pod="kube-system/cilium-operator-599987898-jjdtq" May 8 03:07:08.010097 kubelet[2746]: I0508 03:07:08.010021 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/64bda319-ec49-409f-8ae5-603c3dfaf0b9-cilium-config-path\") pod \"cilium-operator-599987898-jjdtq\" (UID: \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\") " pod="kube-system/cilium-operator-599987898-jjdtq" May 8 03:07:08.206944 kubelet[2746]: E0508 03:07:08.206890 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.212209 containerd[1496]: time="2025-05-08T03:07:08.212178274Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-v2vld,Uid:5a10c27c-9aa4-48fa-a24d-e93bd176b104,Namespace:kube-system,Attempt:0,}" May 8 03:07:08.219029 kubelet[2746]: E0508 03:07:08.218991 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.219832 containerd[1496]: time="2025-05-08T03:07:08.219642938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tk6vz,Uid:b15fe2b0-8a4d-4808-a529-dda616ef1cb6,Namespace:kube-system,Attempt:0,}" May 8 03:07:08.234024 containerd[1496]: time="2025-05-08T03:07:08.233981866Z" level=info msg="connecting to shim 8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b" address="unix:///run/containerd/s/2139bccb73a5c57214e823d35d35193adc73fa53a1d06e309a315f06741eda81" namespace=k8s.io protocol=ttrpc version=3 May 8 03:07:08.239273 containerd[1496]: time="2025-05-08T03:07:08.239208614Z" level=info msg="connecting to shim 13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" namespace=k8s.io protocol=ttrpc version=3 May 8 03:07:08.257880 systemd[1]: Started cri-containerd-8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b.scope - libcontainer container 8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b. May 8 03:07:08.260576 systemd[1]: Started cri-containerd-13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0.scope - libcontainer container 13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0. May 8 03:07:08.262854 kubelet[2746]: E0508 03:07:08.262827 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.264489 containerd[1496]: time="2025-05-08T03:07:08.264446518Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-jjdtq,Uid:64bda319-ec49-409f-8ae5-603c3dfaf0b9,Namespace:kube-system,Attempt:0,}" May 8 03:07:08.288993 containerd[1496]: time="2025-05-08T03:07:08.288800063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-v2vld,Uid:5a10c27c-9aa4-48fa-a24d-e93bd176b104,Namespace:kube-system,Attempt:0,} returns sandbox id \"8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b\"" May 8 03:07:08.298865 kubelet[2746]: E0508 03:07:08.298841 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.301834 containerd[1496]: time="2025-05-08T03:07:08.301689235Z" level=info msg="connecting to shim eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef" address="unix:///run/containerd/s/e73eb70fa08dd74d1fca39fa4def3d024fcd3a1e405cb0c814f80fd3677ab742" namespace=k8s.io protocol=ttrpc version=3 May 8 03:07:08.302204 containerd[1496]: time="2025-05-08T03:07:08.302035954Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tk6vz,Uid:b15fe2b0-8a4d-4808-a529-dda616ef1cb6,Namespace:kube-system,Attempt:0,} returns sandbox id \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\"" May 8 03:07:08.302794 kubelet[2746]: E0508 03:07:08.302766 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.306402 kubelet[2746]: E0508 03:07:08.305809 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.308342 containerd[1496]: time="2025-05-08T03:07:08.308300980Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 8 03:07:08.310130 containerd[1496]: time="2025-05-08T03:07:08.310085416Z" level=info msg="CreateContainer within sandbox \"8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 8 03:07:08.325574 containerd[1496]: time="2025-05-08T03:07:08.325495502Z" level=info msg="Container be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:08.327018 systemd[1]: Started cri-containerd-eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef.scope - libcontainer container eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef. May 8 03:07:08.335946 containerd[1496]: time="2025-05-08T03:07:08.335908238Z" level=info msg="CreateContainer within sandbox \"8ceb1c506c1a721133fc7581f0116383ec908cd1ea70f889a1ad75d681a47b0b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e\"" May 8 03:07:08.336903 containerd[1496]: time="2025-05-08T03:07:08.336826876Z" level=info msg="StartContainer for \"be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e\"" May 8 03:07:08.339551 containerd[1496]: time="2025-05-08T03:07:08.339501230Z" level=info msg="connecting to shim be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e" address="unix:///run/containerd/s/2139bccb73a5c57214e823d35d35193adc73fa53a1d06e309a315f06741eda81" protocol=ttrpc version=3 May 8 03:07:08.360889 systemd[1]: Started cri-containerd-be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e.scope - libcontainer container be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e. May 8 03:07:08.370254 containerd[1496]: time="2025-05-08T03:07:08.370207882Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-jjdtq,Uid:64bda319-ec49-409f-8ae5-603c3dfaf0b9,Namespace:kube-system,Attempt:0,} returns sandbox id \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\"" May 8 03:07:08.370852 kubelet[2746]: E0508 03:07:08.370828 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.415445 containerd[1496]: time="2025-05-08T03:07:08.415402861Z" level=info msg="StartContainer for \"be03788f2ab1d2fb67037839206632d9a4234c31aa8d79ed65d32c2045b77b6e\" returns successfully" May 8 03:07:08.681193 kubelet[2746]: E0508 03:07:08.680846 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:08.690995 kubelet[2746]: I0508 03:07:08.690925 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-v2vld" podStartSLOduration=1.690906527 podStartE2EDuration="1.690906527s" podCreationTimestamp="2025-05-08 03:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:07:08.690416368 +0000 UTC m=+15.138720551" watchObservedRunningTime="2025-05-08 03:07:08.690906527 +0000 UTC m=+15.139210710" May 8 03:07:13.340882 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount487205886.mount: Deactivated successfully. May 8 03:07:14.575587 containerd[1496]: time="2025-05-08T03:07:14.575535547Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:07:14.576660 containerd[1496]: time="2025-05-08T03:07:14.576442986Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" May 8 03:07:14.577393 containerd[1496]: time="2025-05-08T03:07:14.577339625Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:07:14.578702 containerd[1496]: time="2025-05-08T03:07:14.578647303Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 6.270306763s" May 8 03:07:14.578702 containerd[1496]: time="2025-05-08T03:07:14.578683703Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" May 8 03:07:14.582034 containerd[1496]: time="2025-05-08T03:07:14.581856538Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 8 03:07:14.582946 containerd[1496]: time="2025-05-08T03:07:14.582912816Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 8 03:07:14.589810 containerd[1496]: time="2025-05-08T03:07:14.589182727Z" level=info msg="Container 03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:14.593796 containerd[1496]: time="2025-05-08T03:07:14.593763080Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\"" May 8 03:07:14.594774 containerd[1496]: time="2025-05-08T03:07:14.594301679Z" level=info msg="StartContainer for \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\"" May 8 03:07:14.595115 containerd[1496]: time="2025-05-08T03:07:14.595093398Z" level=info msg="connecting to shim 03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" protocol=ttrpc version=3 May 8 03:07:14.631861 systemd[1]: Started cri-containerd-03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca.scope - libcontainer container 03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca. May 8 03:07:14.668914 containerd[1496]: time="2025-05-08T03:07:14.668870886Z" level=info msg="StartContainer for \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" returns successfully" May 8 03:07:14.692018 kubelet[2746]: E0508 03:07:14.691776 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:14.706422 systemd[1]: cri-containerd-03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca.scope: Deactivated successfully. May 8 03:07:14.733139 containerd[1496]: time="2025-05-08T03:07:14.733032709Z" level=info msg="received exit event container_id:\"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" id:\"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" pid:3180 exited_at:{seconds:1746673634 nanos:724834081}" May 8 03:07:14.740192 containerd[1496]: time="2025-05-08T03:07:14.740151738Z" level=info msg="TaskExit event in podsandbox handler container_id:\"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" id:\"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" pid:3180 exited_at:{seconds:1746673634 nanos:724834081}" May 8 03:07:14.764564 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca-rootfs.mount: Deactivated successfully. May 8 03:07:15.695137 kubelet[2746]: E0508 03:07:15.695104 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:15.699490 containerd[1496]: time="2025-05-08T03:07:15.699443193Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 8 03:07:15.736121 containerd[1496]: time="2025-05-08T03:07:15.735912741Z" level=info msg="Container 8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:15.741108 containerd[1496]: time="2025-05-08T03:07:15.741060614Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\"" May 8 03:07:15.741581 containerd[1496]: time="2025-05-08T03:07:15.741552413Z" level=info msg="StartContainer for \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\"" May 8 03:07:15.742379 containerd[1496]: time="2025-05-08T03:07:15.742334092Z" level=info msg="connecting to shim 8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" protocol=ttrpc version=3 May 8 03:07:15.762916 systemd[1]: Started cri-containerd-8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771.scope - libcontainer container 8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771. May 8 03:07:15.796599 containerd[1496]: time="2025-05-08T03:07:15.796471415Z" level=info msg="StartContainer for \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" returns successfully" May 8 03:07:15.807098 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 8 03:07:15.807483 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 8 03:07:15.807646 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 8 03:07:15.808931 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 03:07:15.811761 containerd[1496]: time="2025-05-08T03:07:15.811008075Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" id:\"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" pid:3224 exited_at:{seconds:1746673635 nanos:810702755}" May 8 03:07:15.811761 containerd[1496]: time="2025-05-08T03:07:15.811089195Z" level=info msg="received exit event container_id:\"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" id:\"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" pid:3224 exited_at:{seconds:1746673635 nanos:810702755}" May 8 03:07:15.811182 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 8 03:07:15.811611 systemd[1]: cri-containerd-8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771.scope: Deactivated successfully. May 8 03:07:15.830423 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771-rootfs.mount: Deactivated successfully. May 8 03:07:15.839496 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 03:07:16.698952 kubelet[2746]: E0508 03:07:16.698911 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:16.702150 containerd[1496]: time="2025-05-08T03:07:16.702105273Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 8 03:07:16.740012 containerd[1496]: time="2025-05-08T03:07:16.739962142Z" level=info msg="Container addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:16.742238 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2317502143.mount: Deactivated successfully. May 8 03:07:16.749925 containerd[1496]: time="2025-05-08T03:07:16.749810369Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\"" May 8 03:07:16.752140 containerd[1496]: time="2025-05-08T03:07:16.751810847Z" level=info msg="StartContainer for \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\"" May 8 03:07:16.753958 containerd[1496]: time="2025-05-08T03:07:16.753932564Z" level=info msg="connecting to shim addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" protocol=ttrpc version=3 May 8 03:07:16.775892 systemd[1]: Started cri-containerd-addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df.scope - libcontainer container addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df. May 8 03:07:16.818694 containerd[1496]: time="2025-05-08T03:07:16.817138480Z" level=info msg="StartContainer for \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" returns successfully" May 8 03:07:16.827300 systemd[1]: cri-containerd-addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df.scope: Deactivated successfully. May 8 03:07:16.839221 containerd[1496]: time="2025-05-08T03:07:16.838991251Z" level=info msg="TaskExit event in podsandbox handler container_id:\"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" id:\"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" pid:3284 exited_at:{seconds:1746673636 nanos:838322772}" May 8 03:07:16.839221 containerd[1496]: time="2025-05-08T03:07:16.839164411Z" level=info msg="received exit event container_id:\"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" id:\"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" pid:3284 exited_at:{seconds:1746673636 nanos:838322772}" May 8 03:07:16.860639 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df-rootfs.mount: Deactivated successfully. May 8 03:07:16.993253 containerd[1496]: time="2025-05-08T03:07:16.993146766Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:07:16.994343 containerd[1496]: time="2025-05-08T03:07:16.994293284Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" May 8 03:07:16.995212 containerd[1496]: time="2025-05-08T03:07:16.995170883Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 03:07:16.996915 containerd[1496]: time="2025-05-08T03:07:16.996876081Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.414985823s" May 8 03:07:16.996995 containerd[1496]: time="2025-05-08T03:07:16.996920161Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" May 8 03:07:16.999675 containerd[1496]: time="2025-05-08T03:07:16.999636677Z" level=info msg="CreateContainer within sandbox \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 8 03:07:17.004970 containerd[1496]: time="2025-05-08T03:07:17.004933670Z" level=info msg="Container 65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:17.010718 containerd[1496]: time="2025-05-08T03:07:17.010668263Z" level=info msg="CreateContainer within sandbox \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\"" May 8 03:07:17.011753 containerd[1496]: time="2025-05-08T03:07:17.011315022Z" level=info msg="StartContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\"" May 8 03:07:17.012120 containerd[1496]: time="2025-05-08T03:07:17.012096821Z" level=info msg="connecting to shim 65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05" address="unix:///run/containerd/s/e73eb70fa08dd74d1fca39fa4def3d024fcd3a1e405cb0c814f80fd3677ab742" protocol=ttrpc version=3 May 8 03:07:17.032904 systemd[1]: Started cri-containerd-65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05.scope - libcontainer container 65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05. May 8 03:07:17.056750 containerd[1496]: time="2025-05-08T03:07:17.056642366Z" level=info msg="StartContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" returns successfully" May 8 03:07:17.705650 kubelet[2746]: E0508 03:07:17.704570 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:17.714932 containerd[1496]: time="2025-05-08T03:07:17.714295226Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 8 03:07:17.720259 kubelet[2746]: E0508 03:07:17.720218 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:17.745507 containerd[1496]: time="2025-05-08T03:07:17.745458907Z" level=info msg="Container 9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:17.749810 kubelet[2746]: I0508 03:07:17.749749 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-jjdtq" podStartSLOduration=2.123891581 podStartE2EDuration="10.749693422s" podCreationTimestamp="2025-05-08 03:07:07 +0000 UTC" firstStartedPulling="2025-05-08 03:07:08.371696879 +0000 UTC m=+14.820001062" lastFinishedPulling="2025-05-08 03:07:16.99749872 +0000 UTC m=+23.445802903" observedRunningTime="2025-05-08 03:07:17.749636222 +0000 UTC m=+24.197940405" watchObservedRunningTime="2025-05-08 03:07:17.749693422 +0000 UTC m=+24.197997565" May 8 03:07:17.752905 containerd[1496]: time="2025-05-08T03:07:17.752862658Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\"" May 8 03:07:17.754399 containerd[1496]: time="2025-05-08T03:07:17.754364336Z" level=info msg="StartContainer for \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\"" May 8 03:07:17.755367 containerd[1496]: time="2025-05-08T03:07:17.755176335Z" level=info msg="connecting to shim 9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" protocol=ttrpc version=3 May 8 03:07:17.777914 systemd[1]: Started cri-containerd-9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273.scope - libcontainer container 9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273. May 8 03:07:17.799301 systemd[1]: cri-containerd-9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273.scope: Deactivated successfully. May 8 03:07:17.800983 containerd[1496]: time="2025-05-08T03:07:17.800939278Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" id:\"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" pid:3362 exited_at:{seconds:1746673637 nanos:799842839}" May 8 03:07:17.844470 containerd[1496]: time="2025-05-08T03:07:17.842525026Z" level=info msg="received exit event container_id:\"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" id:\"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" pid:3362 exited_at:{seconds:1746673637 nanos:799842839}" May 8 03:07:17.844920 containerd[1496]: time="2025-05-08T03:07:17.813789662Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice/cri-containerd-9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273.scope/memory.events\": no such file or directory" May 8 03:07:17.845556 containerd[1496]: time="2025-05-08T03:07:17.845134783Z" level=info msg="StartContainer for \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" returns successfully" May 8 03:07:17.862672 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273-rootfs.mount: Deactivated successfully. May 8 03:07:18.724613 kubelet[2746]: E0508 03:07:18.724562 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:18.724613 kubelet[2746]: E0508 03:07:18.724597 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:18.727303 containerd[1496]: time="2025-05-08T03:07:18.726690740Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 8 03:07:18.745930 containerd[1496]: time="2025-05-08T03:07:18.745876158Z" level=info msg="Container cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:18.752627 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3901474951.mount: Deactivated successfully. May 8 03:07:18.753899 containerd[1496]: time="2025-05-08T03:07:18.753855669Z" level=info msg="CreateContainer within sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\"" May 8 03:07:18.754343 containerd[1496]: time="2025-05-08T03:07:18.754315428Z" level=info msg="StartContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\"" May 8 03:07:18.755402 containerd[1496]: time="2025-05-08T03:07:18.755377027Z" level=info msg="connecting to shim cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6" address="unix:///run/containerd/s/18224d7934bd0215be9a4fd06fdd72ca394893152f37d5e1276e0082209966dd" protocol=ttrpc version=3 May 8 03:07:18.783886 systemd[1]: Started cri-containerd-cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6.scope - libcontainer container cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6. May 8 03:07:18.821145 containerd[1496]: time="2025-05-08T03:07:18.821088230Z" level=info msg="StartContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" returns successfully" May 8 03:07:18.939484 containerd[1496]: time="2025-05-08T03:07:18.938790253Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" id:\"376a8efbdebc406c46cad62b63ca705b51d762bc85cb57f3bcbd02aa26c4a20f\" pid:3431 exited_at:{seconds:1746673638 nanos:938310813}" May 8 03:07:18.971535 kubelet[2746]: I0508 03:07:18.971506 2746 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 8 03:07:19.004880 kubelet[2746]: I0508 03:07:19.003148 2746 topology_manager.go:215] "Topology Admit Handler" podUID="2c6bce78-07b2-45cc-b364-148a48e32a7a" podNamespace="kube-system" podName="coredns-7db6d8ff4d-hv5dm" May 8 03:07:19.004880 kubelet[2746]: I0508 03:07:19.003683 2746 topology_manager.go:215] "Topology Admit Handler" podUID="e6760b87-c22c-43cb-8b86-5bc52385c3c9" podNamespace="kube-system" podName="coredns-7db6d8ff4d-2z7v2" May 8 03:07:19.024521 systemd[1]: Created slice kubepods-burstable-pod2c6bce78_07b2_45cc_b364_148a48e32a7a.slice - libcontainer container kubepods-burstable-pod2c6bce78_07b2_45cc_b364_148a48e32a7a.slice. May 8 03:07:19.029553 systemd[1]: Created slice kubepods-burstable-pode6760b87_c22c_43cb_8b86_5bc52385c3c9.slice - libcontainer container kubepods-burstable-pode6760b87_c22c_43cb_8b86_5bc52385c3c9.slice. May 8 03:07:19.088536 kubelet[2746]: I0508 03:07:19.088499 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwlgp\" (UniqueName: \"kubernetes.io/projected/2c6bce78-07b2-45cc-b364-148a48e32a7a-kube-api-access-kwlgp\") pod \"coredns-7db6d8ff4d-hv5dm\" (UID: \"2c6bce78-07b2-45cc-b364-148a48e32a7a\") " pod="kube-system/coredns-7db6d8ff4d-hv5dm" May 8 03:07:19.088536 kubelet[2746]: I0508 03:07:19.088536 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6760b87-c22c-43cb-8b86-5bc52385c3c9-config-volume\") pod \"coredns-7db6d8ff4d-2z7v2\" (UID: \"e6760b87-c22c-43cb-8b86-5bc52385c3c9\") " pod="kube-system/coredns-7db6d8ff4d-2z7v2" May 8 03:07:19.088691 kubelet[2746]: I0508 03:07:19.088560 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6bce78-07b2-45cc-b364-148a48e32a7a-config-volume\") pod \"coredns-7db6d8ff4d-hv5dm\" (UID: \"2c6bce78-07b2-45cc-b364-148a48e32a7a\") " pod="kube-system/coredns-7db6d8ff4d-hv5dm" May 8 03:07:19.088691 kubelet[2746]: I0508 03:07:19.088578 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m72qg\" (UniqueName: \"kubernetes.io/projected/e6760b87-c22c-43cb-8b86-5bc52385c3c9-kube-api-access-m72qg\") pod \"coredns-7db6d8ff4d-2z7v2\" (UID: \"e6760b87-c22c-43cb-8b86-5bc52385c3c9\") " pod="kube-system/coredns-7db6d8ff4d-2z7v2" May 8 03:07:19.328149 kubelet[2746]: E0508 03:07:19.328045 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:19.328798 containerd[1496]: time="2025-05-08T03:07:19.328758261Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hv5dm,Uid:2c6bce78-07b2-45cc-b364-148a48e32a7a,Namespace:kube-system,Attempt:0,}" May 8 03:07:19.332396 kubelet[2746]: E0508 03:07:19.332355 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:19.332869 containerd[1496]: time="2025-05-08T03:07:19.332811456Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-2z7v2,Uid:e6760b87-c22c-43cb-8b86-5bc52385c3c9,Namespace:kube-system,Attempt:0,}" May 8 03:07:19.732319 kubelet[2746]: E0508 03:07:19.732233 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:19.743181 kubelet[2746]: I0508 03:07:19.743128 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-tk6vz" podStartSLOduration=6.470002049 podStartE2EDuration="12.743112247s" podCreationTimestamp="2025-05-08 03:07:07 +0000 UTC" firstStartedPulling="2025-05-08 03:07:08.307949981 +0000 UTC m=+14.756254164" lastFinishedPulling="2025-05-08 03:07:14.581060179 +0000 UTC m=+21.029364362" observedRunningTime="2025-05-08 03:07:19.743080047 +0000 UTC m=+26.191384230" watchObservedRunningTime="2025-05-08 03:07:19.743112247 +0000 UTC m=+26.191416430" May 8 03:07:20.731821 kubelet[2746]: E0508 03:07:20.731777 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:21.052910 systemd-networkd[1424]: cilium_host: Link UP May 8 03:07:21.053051 systemd-networkd[1424]: cilium_net: Link UP May 8 03:07:21.053179 systemd-networkd[1424]: cilium_net: Gained carrier May 8 03:07:21.053294 systemd-networkd[1424]: cilium_host: Gained carrier May 8 03:07:21.143753 systemd-networkd[1424]: cilium_vxlan: Link UP May 8 03:07:21.143760 systemd-networkd[1424]: cilium_vxlan: Gained carrier May 8 03:07:21.439515 kernel: NET: Registered PF_ALG protocol family May 8 03:07:21.516024 systemd-networkd[1424]: cilium_host: Gained IPv6LL May 8 03:07:21.734946 kubelet[2746]: E0508 03:07:21.734830 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:21.779905 systemd-networkd[1424]: cilium_net: Gained IPv6LL May 8 03:07:22.029829 systemd-networkd[1424]: lxc_health: Link UP May 8 03:07:22.030041 systemd-networkd[1424]: lxc_health: Gained carrier May 8 03:07:22.455759 kernel: eth0: renamed from tmp1037d May 8 03:07:22.468570 systemd-networkd[1424]: lxc46dec57dd470: Link UP May 8 03:07:22.468767 systemd-networkd[1424]: lxcf83c45347e22: Link UP May 8 03:07:22.476741 kernel: eth0: renamed from tmp64739 May 8 03:07:22.490447 systemd-networkd[1424]: lxc46dec57dd470: Gained carrier May 8 03:07:22.490756 systemd-networkd[1424]: lxcf83c45347e22: Gained carrier May 8 03:07:22.736731 kubelet[2746]: E0508 03:07:22.736613 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:22.740841 systemd-networkd[1424]: cilium_vxlan: Gained IPv6LL May 8 03:07:23.506810 systemd[1]: Started sshd@7-10.0.0.7:22-10.0.0.1:38364.service - OpenSSH per-connection server daemon (10.0.0.1:38364). May 8 03:07:23.563925 sshd[3913]: Accepted publickey for core from 10.0.0.1 port 38364 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:23.565184 sshd-session[3913]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:23.570003 systemd-logind[1481]: New session 8 of user core. May 8 03:07:23.582896 systemd[1]: Started session-8.scope - Session 8 of User core. May 8 03:07:23.700893 systemd-networkd[1424]: lxc_health: Gained IPv6LL May 8 03:07:23.712663 sshd[3915]: Connection closed by 10.0.0.1 port 38364 May 8 03:07:23.713184 sshd-session[3913]: pam_unix(sshd:session): session closed for user core May 8 03:07:23.716623 systemd[1]: sshd@7-10.0.0.7:22-10.0.0.1:38364.service: Deactivated successfully. May 8 03:07:23.718366 systemd[1]: session-8.scope: Deactivated successfully. May 8 03:07:23.719092 systemd-logind[1481]: Session 8 logged out. Waiting for processes to exit. May 8 03:07:23.721522 systemd-logind[1481]: Removed session 8. May 8 03:07:24.020886 systemd-networkd[1424]: lxc46dec57dd470: Gained IPv6LL May 8 03:07:24.147881 systemd-networkd[1424]: lxcf83c45347e22: Gained IPv6LL May 8 03:07:25.960285 containerd[1496]: time="2025-05-08T03:07:25.958277919Z" level=info msg="connecting to shim 1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb" address="unix:///run/containerd/s/238e1b1e6cd372e15ad5f6af77c3a294f412f047b016629afd9b2e230ea1ef47" namespace=k8s.io protocol=ttrpc version=3 May 8 03:07:25.969818 containerd[1496]: time="2025-05-08T03:07:25.969329951Z" level=info msg="connecting to shim 6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f" address="unix:///run/containerd/s/828add0fb8237d79d85acd2995d82b02ce603fa040a117ed470dfe7a43872439" namespace=k8s.io protocol=ttrpc version=3 May 8 03:07:25.989896 systemd[1]: Started cri-containerd-1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb.scope - libcontainer container 1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb. May 8 03:07:25.993222 systemd[1]: Started cri-containerd-6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f.scope - libcontainer container 6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f. May 8 03:07:26.003573 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 03:07:26.007622 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 03:07:26.033469 containerd[1496]: time="2025-05-08T03:07:26.033409624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-2z7v2,Uid:e6760b87-c22c-43cb-8b86-5bc52385c3c9,Namespace:kube-system,Attempt:0,} returns sandbox id \"6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f\"" May 8 03:07:26.034977 kubelet[2746]: E0508 03:07:26.034287 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:26.036412 containerd[1496]: time="2025-05-08T03:07:26.035947343Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hv5dm,Uid:2c6bce78-07b2-45cc-b364-148a48e32a7a,Namespace:kube-system,Attempt:0,} returns sandbox id \"1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb\"" May 8 03:07:26.038904 kubelet[2746]: E0508 03:07:26.038867 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:26.040378 containerd[1496]: time="2025-05-08T03:07:26.040347539Z" level=info msg="CreateContainer within sandbox \"6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 03:07:26.046509 containerd[1496]: time="2025-05-08T03:07:26.045623896Z" level=info msg="CreateContainer within sandbox \"1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 03:07:26.055138 containerd[1496]: time="2025-05-08T03:07:26.054428410Z" level=info msg="Container ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:26.057797 containerd[1496]: time="2025-05-08T03:07:26.057744687Z" level=info msg="Container 5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e: CDI devices from CRI Config.CDIDevices: []" May 8 03:07:26.062765 containerd[1496]: time="2025-05-08T03:07:26.062622764Z" level=info msg="CreateContainer within sandbox \"1037df0f7f1d742f731cdfc6a756a6ff24274907f29f1f301306fca07117b3eb\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8\"" May 8 03:07:26.063137 containerd[1496]: time="2025-05-08T03:07:26.063112964Z" level=info msg="StartContainer for \"ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8\"" May 8 03:07:26.065179 containerd[1496]: time="2025-05-08T03:07:26.065147362Z" level=info msg="connecting to shim ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8" address="unix:///run/containerd/s/238e1b1e6cd372e15ad5f6af77c3a294f412f047b016629afd9b2e230ea1ef47" protocol=ttrpc version=3 May 8 03:07:26.066879 containerd[1496]: time="2025-05-08T03:07:26.066852081Z" level=info msg="CreateContainer within sandbox \"6473972f3688c36d109fb7660ae43cfbbb4ad517a3acca530f92c5d796092e7f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e\"" May 8 03:07:26.067309 containerd[1496]: time="2025-05-08T03:07:26.067280041Z" level=info msg="StartContainer for \"5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e\"" May 8 03:07:26.068747 containerd[1496]: time="2025-05-08T03:07:26.068694680Z" level=info msg="connecting to shim 5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e" address="unix:///run/containerd/s/828add0fb8237d79d85acd2995d82b02ce603fa040a117ed470dfe7a43872439" protocol=ttrpc version=3 May 8 03:07:26.087676 systemd[1]: Started cri-containerd-ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8.scope - libcontainer container ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8. May 8 03:07:26.091863 systemd[1]: Started cri-containerd-5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e.scope - libcontainer container 5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e. May 8 03:07:26.123299 containerd[1496]: time="2025-05-08T03:07:26.123268562Z" level=info msg="StartContainer for \"ab01563b638cc35c1b3dbf0c285c16452b2120f9619ae4298adf6ad7e9fb1eb8\" returns successfully" May 8 03:07:26.139069 containerd[1496]: time="2025-05-08T03:07:26.139027831Z" level=info msg="StartContainer for \"5c03f19b4ec8aab0335faa2bb90f7b4e01cc5e6e67f6c9f033c89df54508377e\" returns successfully" May 8 03:07:26.751253 kubelet[2746]: E0508 03:07:26.751185 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:26.751827 kubelet[2746]: E0508 03:07:26.751581 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:26.761391 kubelet[2746]: I0508 03:07:26.760965 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-2z7v2" podStartSLOduration=19.760952157 podStartE2EDuration="19.760952157s" podCreationTimestamp="2025-05-08 03:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:07:26.759445918 +0000 UTC m=+33.207750101" watchObservedRunningTime="2025-05-08 03:07:26.760952157 +0000 UTC m=+33.209256300" May 8 03:07:26.782309 kubelet[2746]: I0508 03:07:26.781663 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-hv5dm" podStartSLOduration=19.781647382 podStartE2EDuration="19.781647382s" podCreationTimestamp="2025-05-08 03:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:07:26.77143119 +0000 UTC m=+33.219735373" watchObservedRunningTime="2025-05-08 03:07:26.781647382 +0000 UTC m=+33.229951565" May 8 03:07:27.753226 kubelet[2746]: E0508 03:07:27.753140 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:27.753226 kubelet[2746]: E0508 03:07:27.753217 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:28.728108 systemd[1]: Started sshd@8-10.0.0.7:22-10.0.0.1:38378.service - OpenSSH per-connection server daemon (10.0.0.1:38378). May 8 03:07:28.754682 kubelet[2746]: E0508 03:07:28.754654 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:28.800961 sshd[4106]: Accepted publickey for core from 10.0.0.1 port 38378 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:28.802362 sshd-session[4106]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:28.806586 systemd-logind[1481]: New session 9 of user core. May 8 03:07:28.823921 systemd[1]: Started session-9.scope - Session 9 of User core. May 8 03:07:28.948462 sshd[4108]: Connection closed by 10.0.0.1 port 38378 May 8 03:07:28.948990 sshd-session[4106]: pam_unix(sshd:session): session closed for user core May 8 03:07:28.952906 systemd[1]: sshd@8-10.0.0.7:22-10.0.0.1:38378.service: Deactivated successfully. May 8 03:07:28.955150 systemd[1]: session-9.scope: Deactivated successfully. May 8 03:07:28.956379 systemd-logind[1481]: Session 9 logged out. Waiting for processes to exit. May 8 03:07:28.957292 systemd-logind[1481]: Removed session 9. May 8 03:07:33.963346 systemd[1]: Started sshd@9-10.0.0.7:22-10.0.0.1:39536.service - OpenSSH per-connection server daemon (10.0.0.1:39536). May 8 03:07:34.025786 sshd[4123]: Accepted publickey for core from 10.0.0.1 port 39536 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:34.026931 sshd-session[4123]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:34.030658 systemd-logind[1481]: New session 10 of user core. May 8 03:07:34.039897 systemd[1]: Started session-10.scope - Session 10 of User core. May 8 03:07:34.161880 sshd[4125]: Connection closed by 10.0.0.1 port 39536 May 8 03:07:34.162647 sshd-session[4123]: pam_unix(sshd:session): session closed for user core May 8 03:07:34.166433 systemd[1]: sshd@9-10.0.0.7:22-10.0.0.1:39536.service: Deactivated successfully. May 8 03:07:34.168407 systemd[1]: session-10.scope: Deactivated successfully. May 8 03:07:34.170347 systemd-logind[1481]: Session 10 logged out. Waiting for processes to exit. May 8 03:07:34.171405 systemd-logind[1481]: Removed session 10. May 8 03:07:36.607746 kubelet[2746]: I0508 03:07:36.606762 2746 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 8 03:07:36.610769 kubelet[2746]: E0508 03:07:36.610734 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:36.772055 kubelet[2746]: E0508 03:07:36.771961 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:07:39.179112 systemd[1]: Started sshd@10-10.0.0.7:22-10.0.0.1:39546.service - OpenSSH per-connection server daemon (10.0.0.1:39546). May 8 03:07:39.237254 sshd[4143]: Accepted publickey for core from 10.0.0.1 port 39546 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:39.237773 sshd-session[4143]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:39.242947 systemd-logind[1481]: New session 11 of user core. May 8 03:07:39.249904 systemd[1]: Started session-11.scope - Session 11 of User core. May 8 03:07:39.378979 sshd[4145]: Connection closed by 10.0.0.1 port 39546 May 8 03:07:39.378099 sshd-session[4143]: pam_unix(sshd:session): session closed for user core May 8 03:07:39.396390 systemd[1]: sshd@10-10.0.0.7:22-10.0.0.1:39546.service: Deactivated successfully. May 8 03:07:39.398308 systemd[1]: session-11.scope: Deactivated successfully. May 8 03:07:39.399419 systemd-logind[1481]: Session 11 logged out. Waiting for processes to exit. May 8 03:07:39.401383 systemd[1]: Started sshd@11-10.0.0.7:22-10.0.0.1:39558.service - OpenSSH per-connection server daemon (10.0.0.1:39558). May 8 03:07:39.402995 systemd-logind[1481]: Removed session 11. May 8 03:07:39.457638 sshd[4161]: Accepted publickey for core from 10.0.0.1 port 39558 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:39.458824 sshd-session[4161]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:39.463202 systemd-logind[1481]: New session 12 of user core. May 8 03:07:39.473892 systemd[1]: Started session-12.scope - Session 12 of User core. May 8 03:07:39.649820 sshd[4164]: Connection closed by 10.0.0.1 port 39558 May 8 03:07:39.650582 sshd-session[4161]: pam_unix(sshd:session): session closed for user core May 8 03:07:39.664031 systemd[1]: sshd@11-10.0.0.7:22-10.0.0.1:39558.service: Deactivated successfully. May 8 03:07:39.668005 systemd[1]: session-12.scope: Deactivated successfully. May 8 03:07:39.672329 systemd-logind[1481]: Session 12 logged out. Waiting for processes to exit. May 8 03:07:39.672539 systemd[1]: Started sshd@12-10.0.0.7:22-10.0.0.1:39574.service - OpenSSH per-connection server daemon (10.0.0.1:39574). May 8 03:07:39.675814 systemd-logind[1481]: Removed session 12. May 8 03:07:39.736565 sshd[4174]: Accepted publickey for core from 10.0.0.1 port 39574 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:39.737923 sshd-session[4174]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:39.741856 systemd-logind[1481]: New session 13 of user core. May 8 03:07:39.748876 systemd[1]: Started session-13.scope - Session 13 of User core. May 8 03:07:39.866104 sshd[4177]: Connection closed by 10.0.0.1 port 39574 May 8 03:07:39.866590 sshd-session[4174]: pam_unix(sshd:session): session closed for user core May 8 03:07:39.870006 systemd[1]: sshd@12-10.0.0.7:22-10.0.0.1:39574.service: Deactivated successfully. May 8 03:07:39.872001 systemd[1]: session-13.scope: Deactivated successfully. May 8 03:07:39.872826 systemd-logind[1481]: Session 13 logged out. Waiting for processes to exit. May 8 03:07:39.873637 systemd-logind[1481]: Removed session 13. May 8 03:07:44.876963 systemd[1]: Started sshd@13-10.0.0.7:22-10.0.0.1:39008.service - OpenSSH per-connection server daemon (10.0.0.1:39008). May 8 03:07:44.924671 sshd[4192]: Accepted publickey for core from 10.0.0.1 port 39008 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:44.925890 sshd-session[4192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:44.929772 systemd-logind[1481]: New session 14 of user core. May 8 03:07:44.937892 systemd[1]: Started session-14.scope - Session 14 of User core. May 8 03:07:45.049775 sshd[4194]: Connection closed by 10.0.0.1 port 39008 May 8 03:07:45.050093 sshd-session[4192]: pam_unix(sshd:session): session closed for user core May 8 03:07:45.053362 systemd-logind[1481]: Session 14 logged out. Waiting for processes to exit. May 8 03:07:45.053613 systemd[1]: sshd@13-10.0.0.7:22-10.0.0.1:39008.service: Deactivated successfully. May 8 03:07:45.055176 systemd[1]: session-14.scope: Deactivated successfully. May 8 03:07:45.056893 systemd-logind[1481]: Removed session 14. May 8 03:07:50.065692 systemd[1]: Started sshd@14-10.0.0.7:22-10.0.0.1:39014.service - OpenSSH per-connection server daemon (10.0.0.1:39014). May 8 03:07:50.112285 sshd[4207]: Accepted publickey for core from 10.0.0.1 port 39014 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:50.113428 sshd-session[4207]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:50.117119 systemd-logind[1481]: New session 15 of user core. May 8 03:07:50.125876 systemd[1]: Started session-15.scope - Session 15 of User core. May 8 03:07:50.257494 sshd[4209]: Connection closed by 10.0.0.1 port 39014 May 8 03:07:50.257855 sshd-session[4207]: pam_unix(sshd:session): session closed for user core May 8 03:07:50.270856 systemd[1]: sshd@14-10.0.0.7:22-10.0.0.1:39014.service: Deactivated successfully. May 8 03:07:50.272269 systemd[1]: session-15.scope: Deactivated successfully. May 8 03:07:50.275702 systemd-logind[1481]: Session 15 logged out. Waiting for processes to exit. May 8 03:07:50.277205 systemd[1]: Started sshd@15-10.0.0.7:22-10.0.0.1:39018.service - OpenSSH per-connection server daemon (10.0.0.1:39018). May 8 03:07:50.280077 systemd-logind[1481]: Removed session 15. May 8 03:07:50.332821 sshd[4221]: Accepted publickey for core from 10.0.0.1 port 39018 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:50.335543 sshd-session[4221]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:50.344066 systemd-logind[1481]: New session 16 of user core. May 8 03:07:50.350887 systemd[1]: Started session-16.scope - Session 16 of User core. May 8 03:07:50.546174 sshd[4224]: Connection closed by 10.0.0.1 port 39018 May 8 03:07:50.546780 sshd-session[4221]: pam_unix(sshd:session): session closed for user core May 8 03:07:50.560505 systemd[1]: sshd@15-10.0.0.7:22-10.0.0.1:39018.service: Deactivated successfully. May 8 03:07:50.562193 systemd[1]: session-16.scope: Deactivated successfully. May 8 03:07:50.563026 systemd-logind[1481]: Session 16 logged out. Waiting for processes to exit. May 8 03:07:50.565579 systemd[1]: Started sshd@16-10.0.0.7:22-10.0.0.1:39022.service - OpenSSH per-connection server daemon (10.0.0.1:39022). May 8 03:07:50.567524 systemd-logind[1481]: Removed session 16. May 8 03:07:50.623631 sshd[4234]: Accepted publickey for core from 10.0.0.1 port 39022 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:50.625077 sshd-session[4234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:50.629435 systemd-logind[1481]: New session 17 of user core. May 8 03:07:50.646904 systemd[1]: Started session-17.scope - Session 17 of User core. May 8 03:07:52.095158 sshd[4237]: Connection closed by 10.0.0.1 port 39022 May 8 03:07:52.096484 sshd-session[4234]: pam_unix(sshd:session): session closed for user core May 8 03:07:52.106712 systemd[1]: sshd@16-10.0.0.7:22-10.0.0.1:39022.service: Deactivated successfully. May 8 03:07:52.110616 systemd[1]: session-17.scope: Deactivated successfully. May 8 03:07:52.113810 systemd-logind[1481]: Session 17 logged out. Waiting for processes to exit. May 8 03:07:52.117279 systemd[1]: Started sshd@17-10.0.0.7:22-10.0.0.1:39032.service - OpenSSH per-connection server daemon (10.0.0.1:39032). May 8 03:07:52.118848 systemd-logind[1481]: Removed session 17. May 8 03:07:52.170148 sshd[4260]: Accepted publickey for core from 10.0.0.1 port 39032 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:52.171660 sshd-session[4260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:52.176633 systemd-logind[1481]: New session 18 of user core. May 8 03:07:52.182915 systemd[1]: Started session-18.scope - Session 18 of User core. May 8 03:07:52.403045 sshd[4263]: Connection closed by 10.0.0.1 port 39032 May 8 03:07:52.403920 sshd-session[4260]: pam_unix(sshd:session): session closed for user core May 8 03:07:52.411226 systemd[1]: sshd@17-10.0.0.7:22-10.0.0.1:39032.service: Deactivated successfully. May 8 03:07:52.413899 systemd[1]: session-18.scope: Deactivated successfully. May 8 03:07:52.417070 systemd-logind[1481]: Session 18 logged out. Waiting for processes to exit. May 8 03:07:52.419101 systemd[1]: Started sshd@18-10.0.0.7:22-10.0.0.1:39034.service - OpenSSH per-connection server daemon (10.0.0.1:39034). May 8 03:07:52.422980 systemd-logind[1481]: Removed session 18. May 8 03:07:52.474194 sshd[4273]: Accepted publickey for core from 10.0.0.1 port 39034 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:52.475991 sshd-session[4273]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:52.481016 systemd-logind[1481]: New session 19 of user core. May 8 03:07:52.490947 systemd[1]: Started session-19.scope - Session 19 of User core. May 8 03:07:52.603623 sshd[4276]: Connection closed by 10.0.0.1 port 39034 May 8 03:07:52.604340 sshd-session[4273]: pam_unix(sshd:session): session closed for user core May 8 03:07:52.608637 systemd[1]: sshd@18-10.0.0.7:22-10.0.0.1:39034.service: Deactivated successfully. May 8 03:07:52.611684 systemd[1]: session-19.scope: Deactivated successfully. May 8 03:07:52.612596 systemd-logind[1481]: Session 19 logged out. Waiting for processes to exit. May 8 03:07:52.613572 systemd-logind[1481]: Removed session 19. May 8 03:07:57.615043 systemd[1]: Started sshd@19-10.0.0.7:22-10.0.0.1:54618.service - OpenSSH per-connection server daemon (10.0.0.1:54618). May 8 03:07:57.669924 sshd[4295]: Accepted publickey for core from 10.0.0.1 port 54618 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:07:57.671276 sshd-session[4295]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:07:57.675263 systemd-logind[1481]: New session 20 of user core. May 8 03:07:57.686877 systemd[1]: Started session-20.scope - Session 20 of User core. May 8 03:07:57.798008 sshd[4297]: Connection closed by 10.0.0.1 port 54618 May 8 03:07:57.798378 sshd-session[4295]: pam_unix(sshd:session): session closed for user core May 8 03:07:57.802079 systemd[1]: sshd@19-10.0.0.7:22-10.0.0.1:54618.service: Deactivated successfully. May 8 03:07:57.804103 systemd[1]: session-20.scope: Deactivated successfully. May 8 03:07:57.805839 systemd-logind[1481]: Session 20 logged out. Waiting for processes to exit. May 8 03:07:57.806667 systemd-logind[1481]: Removed session 20. May 8 03:08:01.634705 kubelet[2746]: E0508 03:08:01.634671 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:02.810090 systemd[1]: Started sshd@20-10.0.0.7:22-10.0.0.1:47114.service - OpenSSH per-connection server daemon (10.0.0.1:47114). May 8 03:08:02.858659 sshd[4310]: Accepted publickey for core from 10.0.0.1 port 47114 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:02.859844 sshd-session[4310]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:02.863588 systemd-logind[1481]: New session 21 of user core. May 8 03:08:02.872902 systemd[1]: Started session-21.scope - Session 21 of User core. May 8 03:08:02.991742 sshd[4312]: Connection closed by 10.0.0.1 port 47114 May 8 03:08:02.992092 sshd-session[4310]: pam_unix(sshd:session): session closed for user core May 8 03:08:02.996557 systemd[1]: sshd@20-10.0.0.7:22-10.0.0.1:47114.service: Deactivated successfully. May 8 03:08:02.998191 systemd[1]: session-21.scope: Deactivated successfully. May 8 03:08:02.999914 systemd-logind[1481]: Session 21 logged out. Waiting for processes to exit. May 8 03:08:03.000932 systemd-logind[1481]: Removed session 21. May 8 03:08:08.006897 systemd[1]: Started sshd@21-10.0.0.7:22-10.0.0.1:47118.service - OpenSSH per-connection server daemon (10.0.0.1:47118). May 8 03:08:08.060952 sshd[4325]: Accepted publickey for core from 10.0.0.1 port 47118 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:08.062495 sshd-session[4325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:08.068964 systemd-logind[1481]: New session 22 of user core. May 8 03:08:08.080927 systemd[1]: Started session-22.scope - Session 22 of User core. May 8 03:08:08.188116 sshd[4327]: Connection closed by 10.0.0.1 port 47118 May 8 03:08:08.188444 sshd-session[4325]: pam_unix(sshd:session): session closed for user core May 8 03:08:08.191933 systemd[1]: sshd@21-10.0.0.7:22-10.0.0.1:47118.service: Deactivated successfully. May 8 03:08:08.194481 systemd[1]: session-22.scope: Deactivated successfully. May 8 03:08:08.195496 systemd-logind[1481]: Session 22 logged out. Waiting for processes to exit. May 8 03:08:08.196476 systemd-logind[1481]: Removed session 22. May 8 03:08:11.634179 kubelet[2746]: E0508 03:08:11.634140 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:13.213167 systemd[1]: Started sshd@22-10.0.0.7:22-10.0.0.1:47534.service - OpenSSH per-connection server daemon (10.0.0.1:47534). May 8 03:08:13.265853 sshd[4342]: Accepted publickey for core from 10.0.0.1 port 47534 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:13.267280 sshd-session[4342]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:13.271785 systemd-logind[1481]: New session 23 of user core. May 8 03:08:13.281894 systemd[1]: Started session-23.scope - Session 23 of User core. May 8 03:08:13.402313 sshd[4344]: Connection closed by 10.0.0.1 port 47534 May 8 03:08:13.401485 sshd-session[4342]: pam_unix(sshd:session): session closed for user core May 8 03:08:13.418983 systemd[1]: sshd@22-10.0.0.7:22-10.0.0.1:47534.service: Deactivated successfully. May 8 03:08:13.420910 systemd[1]: session-23.scope: Deactivated successfully. May 8 03:08:13.424096 systemd-logind[1481]: Session 23 logged out. Waiting for processes to exit. May 8 03:08:13.425497 systemd[1]: Started sshd@23-10.0.0.7:22-10.0.0.1:47536.service - OpenSSH per-connection server daemon (10.0.0.1:47536). May 8 03:08:13.426594 systemd-logind[1481]: Removed session 23. May 8 03:08:13.478955 sshd[4356]: Accepted publickey for core from 10.0.0.1 port 47536 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:13.480119 sshd-session[4356]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:13.485192 systemd-logind[1481]: New session 24 of user core. May 8 03:08:13.496883 systemd[1]: Started session-24.scope - Session 24 of User core. May 8 03:08:14.634570 kubelet[2746]: E0508 03:08:14.634533 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:16.229805 containerd[1496]: time="2025-05-08T03:08:16.229691643Z" level=info msg="StopContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" with timeout 30 (s)" May 8 03:08:16.232922 containerd[1496]: time="2025-05-08T03:08:16.232846253Z" level=info msg="Stop container \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" with signal terminated" May 8 03:08:16.249253 systemd[1]: cri-containerd-65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05.scope: Deactivated successfully. May 8 03:08:16.252846 containerd[1496]: time="2025-05-08T03:08:16.251581432Z" level=info msg="received exit event container_id:\"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" id:\"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" pid:3329 exited_at:{seconds:1746673696 nanos:251274791}" May 8 03:08:16.252846 containerd[1496]: time="2025-05-08T03:08:16.251991193Z" level=info msg="TaskExit event in podsandbox handler container_id:\"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" id:\"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" pid:3329 exited_at:{seconds:1746673696 nanos:251274791}" May 8 03:08:16.269325 containerd[1496]: time="2025-05-08T03:08:16.269242728Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 8 03:08:16.271695 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05-rootfs.mount: Deactivated successfully. May 8 03:08:16.274085 containerd[1496]: time="2025-05-08T03:08:16.273938263Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" id:\"42405ea847310416b424a011dee27bd9efa441751aa829589b3ea56bbc4a69b0\" pid:4389 exited_at:{seconds:1746673696 nanos:273606342}" May 8 03:08:16.276712 containerd[1496]: time="2025-05-08T03:08:16.276677831Z" level=info msg="StopContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" with timeout 2 (s)" May 8 03:08:16.277209 containerd[1496]: time="2025-05-08T03:08:16.277123393Z" level=info msg="Stop container \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" with signal terminated" May 8 03:08:16.283504 systemd-networkd[1424]: lxc_health: Link DOWN May 8 03:08:16.283509 systemd-networkd[1424]: lxc_health: Lost carrier May 8 03:08:16.289997 containerd[1496]: time="2025-05-08T03:08:16.289274151Z" level=info msg="StopContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" returns successfully" May 8 03:08:16.294037 containerd[1496]: time="2025-05-08T03:08:16.293995286Z" level=info msg="StopPodSandbox for \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\"" May 8 03:08:16.294525 containerd[1496]: time="2025-05-08T03:08:16.294503448Z" level=info msg="Container to stop \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.300523 systemd[1]: cri-containerd-cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6.scope: Deactivated successfully. May 8 03:08:16.301509 containerd[1496]: time="2025-05-08T03:08:16.301405470Z" level=info msg="received exit event container_id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" pid:3399 exited_at:{seconds:1746673696 nanos:301200789}" May 8 03:08:16.301509 containerd[1496]: time="2025-05-08T03:08:16.301473270Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" id:\"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" pid:3399 exited_at:{seconds:1746673696 nanos:301200789}" May 8 03:08:16.301472 systemd[1]: cri-containerd-cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6.scope: Consumed 6.406s CPU time, 122.2M memory peak, 152K read from disk, 12.9M written to disk. May 8 03:08:16.314146 systemd[1]: cri-containerd-eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef.scope: Deactivated successfully. May 8 03:08:16.317767 containerd[1496]: time="2025-05-08T03:08:16.317708961Z" level=info msg="TaskExit event in podsandbox handler container_id:\"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" id:\"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" pid:2972 exit_status:137 exited_at:{seconds:1746673696 nanos:316988239}" May 8 03:08:16.325076 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6-rootfs.mount: Deactivated successfully. May 8 03:08:16.342604 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef-rootfs.mount: Deactivated successfully. May 8 03:08:16.345538 containerd[1496]: time="2025-05-08T03:08:16.345496129Z" level=info msg="StopContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" returns successfully" May 8 03:08:16.346061 containerd[1496]: time="2025-05-08T03:08:16.346038411Z" level=info msg="shim disconnected" id=eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef namespace=k8s.io May 8 03:08:16.361230 containerd[1496]: time="2025-05-08T03:08:16.346063971Z" level=warning msg="cleaning up after shim disconnected" id=eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef namespace=k8s.io May 8 03:08:16.361230 containerd[1496]: time="2025-05-08T03:08:16.361228899Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 8 03:08:16.361346 containerd[1496]: time="2025-05-08T03:08:16.346082531Z" level=info msg="StopPodSandbox for \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\"" May 8 03:08:16.361431 containerd[1496]: time="2025-05-08T03:08:16.361404699Z" level=info msg="Container to stop \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.361431 containerd[1496]: time="2025-05-08T03:08:16.361423299Z" level=info msg="Container to stop \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.361483 containerd[1496]: time="2025-05-08T03:08:16.361433859Z" level=info msg="Container to stop \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.361483 containerd[1496]: time="2025-05-08T03:08:16.361443059Z" level=info msg="Container to stop \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.361483 containerd[1496]: time="2025-05-08T03:08:16.361450740Z" level=info msg="Container to stop \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 03:08:16.370247 systemd[1]: cri-containerd-13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0.scope: Deactivated successfully. May 8 03:08:16.382524 containerd[1496]: time="2025-05-08T03:08:16.382446006Z" level=info msg="received exit event sandbox_id:\"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" exit_status:137 exited_at:{seconds:1746673696 nanos:316988239}" May 8 03:08:16.384221 containerd[1496]: time="2025-05-08T03:08:16.383313169Z" level=info msg="TaskExit event in podsandbox handler container_id:\"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" id:\"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" pid:2918 exit_status:137 exited_at:{seconds:1746673696 nanos:374070259}" May 8 03:08:16.384400 containerd[1496]: time="2025-05-08T03:08:16.384339652Z" level=info msg="TearDown network for sandbox \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" successfully" May 8 03:08:16.384400 containerd[1496]: time="2025-05-08T03:08:16.384381452Z" level=info msg="StopPodSandbox for \"eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef\" returns successfully" May 8 03:08:16.386187 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-eb4410e79bc102f982a5c0605b0a81f05288cf01e1a9962d8fda033ede8814ef-shm.mount: Deactivated successfully. May 8 03:08:16.408655 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0-rootfs.mount: Deactivated successfully. May 8 03:08:16.412220 containerd[1496]: time="2025-05-08T03:08:16.412181020Z" level=info msg="shim disconnected" id=13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0 namespace=k8s.io May 8 03:08:16.412325 containerd[1496]: time="2025-05-08T03:08:16.412210420Z" level=warning msg="cleaning up after shim disconnected" id=13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0 namespace=k8s.io May 8 03:08:16.412325 containerd[1496]: time="2025-05-08T03:08:16.412247460Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 8 03:08:16.421449 containerd[1496]: time="2025-05-08T03:08:16.421395089Z" level=info msg="received exit event sandbox_id:\"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" exit_status:137 exited_at:{seconds:1746673696 nanos:374070259}" May 8 03:08:16.421563 containerd[1496]: time="2025-05-08T03:08:16.421527529Z" level=info msg="TearDown network for sandbox \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" successfully" May 8 03:08:16.421563 containerd[1496]: time="2025-05-08T03:08:16.421554010Z" level=info msg="StopPodSandbox for \"13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0\" returns successfully" May 8 03:08:16.446559 kubelet[2746]: I0508 03:08:16.446512 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-etc-cni-netd\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.446559 kubelet[2746]: I0508 03:08:16.446560 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-net\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446607 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-config-path\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446625 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cni-path\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446641 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-bpf-maps\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446657 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-lib-modules\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446679 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-cgroup\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447062 kubelet[2746]: I0508 03:08:16.446697 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/64bda319-ec49-409f-8ae5-603c3dfaf0b9-cilium-config-path\") pod \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\" (UID: \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446714 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-clustermesh-secrets\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446754 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-run\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446771 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hostproc\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446785 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-xtables-lock\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446800 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-kernel\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447253 kubelet[2746]: I0508 03:08:16.446819 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsr9b\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-kube-api-access-jsr9b\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.447400 kubelet[2746]: I0508 03:08:16.446835 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbthz\" (UniqueName: \"kubernetes.io/projected/64bda319-ec49-409f-8ae5-603c3dfaf0b9-kube-api-access-zbthz\") pod \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\" (UID: \"64bda319-ec49-409f-8ae5-603c3dfaf0b9\") " May 8 03:08:16.447400 kubelet[2746]: I0508 03:08:16.446851 2746 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hubble-tls\") pod \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\" (UID: \"b15fe2b0-8a4d-4808-a529-dda616ef1cb6\") " May 8 03:08:16.450771 kubelet[2746]: I0508 03:08:16.450648 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.455428 kubelet[2746]: I0508 03:08:16.455290 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cni-path" (OuterVolumeSpecName: "cni-path") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.455428 kubelet[2746]: I0508 03:08:16.455342 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.455527 kubelet[2746]: I0508 03:08:16.455352 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.455527 kubelet[2746]: I0508 03:08:16.455380 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.455527 kubelet[2746]: I0508 03:08:16.455396 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.456018 kubelet[2746]: I0508 03:08:16.455612 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 8 03:08:16.456018 kubelet[2746]: I0508 03:08:16.455852 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.456018 kubelet[2746]: I0508 03:08:16.455879 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.456018 kubelet[2746]: I0508 03:08:16.455897 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hostproc" (OuterVolumeSpecName: "hostproc") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.457960 kubelet[2746]: I0508 03:08:16.457255 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 03:08:16.457960 kubelet[2746]: I0508 03:08:16.457373 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 03:08:16.458219 kubelet[2746]: I0508 03:08:16.458187 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 8 03:08:16.458294 kubelet[2746]: I0508 03:08:16.458274 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-kube-api-access-jsr9b" (OuterVolumeSpecName: "kube-api-access-jsr9b") pod "b15fe2b0-8a4d-4808-a529-dda616ef1cb6" (UID: "b15fe2b0-8a4d-4808-a529-dda616ef1cb6"). InnerVolumeSpecName "kube-api-access-jsr9b". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 03:08:16.458576 kubelet[2746]: I0508 03:08:16.458551 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64bda319-ec49-409f-8ae5-603c3dfaf0b9-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "64bda319-ec49-409f-8ae5-603c3dfaf0b9" (UID: "64bda319-ec49-409f-8ae5-603c3dfaf0b9"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 8 03:08:16.458636 kubelet[2746]: I0508 03:08:16.458607 2746 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64bda319-ec49-409f-8ae5-603c3dfaf0b9-kube-api-access-zbthz" (OuterVolumeSpecName: "kube-api-access-zbthz") pod "64bda319-ec49-409f-8ae5-603c3dfaf0b9" (UID: "64bda319-ec49-409f-8ae5-603c3dfaf0b9"). InnerVolumeSpecName "kube-api-access-zbthz". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547025 2746 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547069 2746 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/64bda319-ec49-409f-8ae5-603c3dfaf0b9-cilium-config-path\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547088 2746 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-xtables-lock\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547106 2746 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547123 2746 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-run\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547139 2746 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hostproc\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547153 2746 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-zbthz\" (UniqueName: \"kubernetes.io/projected/64bda319-ec49-409f-8ae5-603c3dfaf0b9-kube-api-access-zbthz\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548760 kubelet[2746]: I0508 03:08:16.547170 2746 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-jsr9b\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-kube-api-access-jsr9b\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547184 2746 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-hubble-tls\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547197 2746 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547211 2746 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547224 2746 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cni-path\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547236 2746 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-config-path\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547243 2746 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-lib-modules\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547251 2746 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.548998 kubelet[2746]: I0508 03:08:16.547259 2746 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b15fe2b0-8a4d-4808-a529-dda616ef1cb6-bpf-maps\") on node \"localhost\" DevicePath \"\"" May 8 03:08:16.849199 kubelet[2746]: I0508 03:08:16.849087 2746 scope.go:117] "RemoveContainer" containerID="65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05" May 8 03:08:16.851572 containerd[1496]: time="2025-05-08T03:08:16.851322449Z" level=info msg="RemoveContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\"" May 8 03:08:16.857931 systemd[1]: Removed slice kubepods-besteffort-pod64bda319_ec49_409f_8ae5_603c3dfaf0b9.slice - libcontainer container kubepods-besteffort-pod64bda319_ec49_409f_8ae5_603c3dfaf0b9.slice. May 8 03:08:16.863552 systemd[1]: Removed slice kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice - libcontainer container kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice. May 8 03:08:16.863814 systemd[1]: kubepods-burstable-podb15fe2b0_8a4d_4808_a529_dda616ef1cb6.slice: Consumed 6.554s CPU time, 122.5M memory peak, 164K read from disk, 12.9M written to disk. May 8 03:08:16.874826 containerd[1496]: time="2025-05-08T03:08:16.874486242Z" level=info msg="RemoveContainer for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" returns successfully" May 8 03:08:16.876284 kubelet[2746]: I0508 03:08:16.875883 2746 scope.go:117] "RemoveContainer" containerID="65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05" May 8 03:08:16.876393 containerd[1496]: time="2025-05-08T03:08:16.876155167Z" level=error msg="ContainerStatus for \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\": not found" May 8 03:08:16.876433 kubelet[2746]: E0508 03:08:16.876408 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\": not found" containerID="65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05" May 8 03:08:16.877349 kubelet[2746]: I0508 03:08:16.876439 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05"} err="failed to get container status \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\": rpc error: code = NotFound desc = an error occurred when try to find container \"65cbd935d6f5ea460fe784c6f242c0012afac86033a7c4f5e830b714d4336a05\": not found" May 8 03:08:16.877349 kubelet[2746]: I0508 03:08:16.876704 2746 scope.go:117] "RemoveContainer" containerID="cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6" May 8 03:08:16.878494 containerd[1496]: time="2025-05-08T03:08:16.878463774Z" level=info msg="RemoveContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\"" May 8 03:08:16.889899 containerd[1496]: time="2025-05-08T03:08:16.889852690Z" level=info msg="RemoveContainer for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" returns successfully" May 8 03:08:16.890102 kubelet[2746]: I0508 03:08:16.890075 2746 scope.go:117] "RemoveContainer" containerID="9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273" May 8 03:08:16.891626 containerd[1496]: time="2025-05-08T03:08:16.891585976Z" level=info msg="RemoveContainer for \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\"" May 8 03:08:16.901906 containerd[1496]: time="2025-05-08T03:08:16.901847168Z" level=info msg="RemoveContainer for \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" returns successfully" May 8 03:08:16.902111 kubelet[2746]: I0508 03:08:16.902072 2746 scope.go:117] "RemoveContainer" containerID="addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df" May 8 03:08:16.904333 containerd[1496]: time="2025-05-08T03:08:16.904303536Z" level=info msg="RemoveContainer for \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\"" May 8 03:08:16.910208 containerd[1496]: time="2025-05-08T03:08:16.910172395Z" level=info msg="RemoveContainer for \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" returns successfully" May 8 03:08:16.910368 kubelet[2746]: I0508 03:08:16.910327 2746 scope.go:117] "RemoveContainer" containerID="8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771" May 8 03:08:16.912251 containerd[1496]: time="2025-05-08T03:08:16.911785320Z" level=info msg="RemoveContainer for \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\"" May 8 03:08:16.914376 containerd[1496]: time="2025-05-08T03:08:16.914326048Z" level=info msg="RemoveContainer for \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" returns successfully" May 8 03:08:16.914683 kubelet[2746]: I0508 03:08:16.914657 2746 scope.go:117] "RemoveContainer" containerID="03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca" May 8 03:08:16.916164 containerd[1496]: time="2025-05-08T03:08:16.916136733Z" level=info msg="RemoveContainer for \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\"" May 8 03:08:16.918513 containerd[1496]: time="2025-05-08T03:08:16.918473981Z" level=info msg="RemoveContainer for \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" returns successfully" May 8 03:08:16.918688 kubelet[2746]: I0508 03:08:16.918658 2746 scope.go:117] "RemoveContainer" containerID="cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6" May 8 03:08:16.918984 containerd[1496]: time="2025-05-08T03:08:16.918951822Z" level=error msg="ContainerStatus for \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\": not found" May 8 03:08:16.919116 kubelet[2746]: E0508 03:08:16.919091 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\": not found" containerID="cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6" May 8 03:08:16.919154 kubelet[2746]: I0508 03:08:16.919125 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6"} err="failed to get container status \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\": rpc error: code = NotFound desc = an error occurred when try to find container \"cc2cc6ae1efd25ec342b3b685e7ff5ea5c469cd3f9e4b20b979ce17e978c1ab6\": not found" May 8 03:08:16.919154 kubelet[2746]: I0508 03:08:16.919147 2746 scope.go:117] "RemoveContainer" containerID="9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273" May 8 03:08:16.919469 containerd[1496]: time="2025-05-08T03:08:16.919324304Z" level=error msg="ContainerStatus for \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\": not found" May 8 03:08:16.919531 kubelet[2746]: E0508 03:08:16.919455 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\": not found" containerID="9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273" May 8 03:08:16.919531 kubelet[2746]: I0508 03:08:16.919479 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273"} err="failed to get container status \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\": rpc error: code = NotFound desc = an error occurred when try to find container \"9c86547232a8ddfbba096792dfd90580f5bb214e28adcb529767148dd6e3a273\": not found" May 8 03:08:16.919531 kubelet[2746]: I0508 03:08:16.919496 2746 scope.go:117] "RemoveContainer" containerID="addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df" May 8 03:08:16.919804 containerd[1496]: time="2025-05-08T03:08:16.919717785Z" level=error msg="ContainerStatus for \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\": not found" May 8 03:08:16.919883 kubelet[2746]: E0508 03:08:16.919860 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\": not found" containerID="addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df" May 8 03:08:16.919923 kubelet[2746]: I0508 03:08:16.919886 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df"} err="failed to get container status \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\": rpc error: code = NotFound desc = an error occurred when try to find container \"addd535ca74c9d6535340388a3743a206641963ca5948cd557a224b32ebec1df\": not found" May 8 03:08:16.919923 kubelet[2746]: I0508 03:08:16.919903 2746 scope.go:117] "RemoveContainer" containerID="8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771" May 8 03:08:16.920104 containerd[1496]: time="2025-05-08T03:08:16.920064546Z" level=error msg="ContainerStatus for \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\": not found" May 8 03:08:16.920203 kubelet[2746]: E0508 03:08:16.920183 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\": not found" containerID="8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771" May 8 03:08:16.920232 kubelet[2746]: I0508 03:08:16.920206 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771"} err="failed to get container status \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\": rpc error: code = NotFound desc = an error occurred when try to find container \"8fb49aeb39a68917c5723f91cab28461c46ccf57b9e912feae6640607b9f3771\": not found" May 8 03:08:16.920232 kubelet[2746]: I0508 03:08:16.920221 2746 scope.go:117] "RemoveContainer" containerID="03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca" May 8 03:08:16.920552 containerd[1496]: time="2025-05-08T03:08:16.920524907Z" level=error msg="ContainerStatus for \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\": not found" May 8 03:08:16.920648 kubelet[2746]: E0508 03:08:16.920630 2746 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\": not found" containerID="03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca" May 8 03:08:16.920681 kubelet[2746]: I0508 03:08:16.920651 2746 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca"} err="failed to get container status \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\": rpc error: code = NotFound desc = an error occurred when try to find container \"03a6965b40ba3e5596314fbd7f65f44025d1a2ccd934e13c21a1aa75b26226ca\": not found" May 8 03:08:17.271591 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-13b2620cee53cff658bfc9d2896d5fdf9fefbc359e75c00e54df6685ceac73d0-shm.mount: Deactivated successfully. May 8 03:08:17.271692 systemd[1]: var-lib-kubelet-pods-64bda319\x2dec49\x2d409f\x2d8ae5\x2d603c3dfaf0b9-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dzbthz.mount: Deactivated successfully. May 8 03:08:17.271770 systemd[1]: var-lib-kubelet-pods-b15fe2b0\x2d8a4d\x2d4808\x2da529\x2ddda616ef1cb6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djsr9b.mount: Deactivated successfully. May 8 03:08:17.271827 systemd[1]: var-lib-kubelet-pods-b15fe2b0\x2d8a4d\x2d4808\x2da529\x2ddda616ef1cb6-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 8 03:08:17.271878 systemd[1]: var-lib-kubelet-pods-b15fe2b0\x2d8a4d\x2d4808\x2da529\x2ddda616ef1cb6-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 8 03:08:17.636704 kubelet[2746]: I0508 03:08:17.636602 2746 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64bda319-ec49-409f-8ae5-603c3dfaf0b9" path="/var/lib/kubelet/pods/64bda319-ec49-409f-8ae5-603c3dfaf0b9/volumes" May 8 03:08:17.638700 kubelet[2746]: I0508 03:08:17.638670 2746 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" path="/var/lib/kubelet/pods/b15fe2b0-8a4d-4808-a529-dda616ef1cb6/volumes" May 8 03:08:18.185172 sshd[4359]: Connection closed by 10.0.0.1 port 47536 May 8 03:08:18.185741 sshd-session[4356]: pam_unix(sshd:session): session closed for user core May 8 03:08:18.199123 systemd[1]: sshd@23-10.0.0.7:22-10.0.0.1:47536.service: Deactivated successfully. May 8 03:08:18.200997 systemd[1]: session-24.scope: Deactivated successfully. May 8 03:08:18.201769 systemd[1]: session-24.scope: Consumed 2.056s CPU time, 28.3M memory peak. May 8 03:08:18.202770 systemd-logind[1481]: Session 24 logged out. Waiting for processes to exit. May 8 03:08:18.204932 systemd[1]: Started sshd@24-10.0.0.7:22-10.0.0.1:47540.service - OpenSSH per-connection server daemon (10.0.0.1:47540). May 8 03:08:18.206169 systemd-logind[1481]: Removed session 24. May 8 03:08:18.281163 sshd[4514]: Accepted publickey for core from 10.0.0.1 port 47540 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:18.282435 sshd-session[4514]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:18.286779 systemd-logind[1481]: New session 25 of user core. May 8 03:08:18.292853 systemd[1]: Started session-25.scope - Session 25 of User core. May 8 03:08:18.688540 kubelet[2746]: E0508 03:08:18.688498 2746 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 8 03:08:19.173292 sshd[4517]: Connection closed by 10.0.0.1 port 47540 May 8 03:08:19.174257 sshd-session[4514]: pam_unix(sshd:session): session closed for user core May 8 03:08:19.186816 kubelet[2746]: I0508 03:08:19.186664 2746 topology_manager.go:215] "Topology Admit Handler" podUID="687509c4-a361-479e-a5fe-9c8609ab71e9" podNamespace="kube-system" podName="cilium-j4lld" May 8 03:08:19.187044 kubelet[2746]: E0508 03:08:19.187020 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="clean-cilium-state" May 8 03:08:19.187143 kubelet[2746]: E0508 03:08:19.187130 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="cilium-agent" May 8 03:08:19.187205 kubelet[2746]: E0508 03:08:19.187196 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="mount-cgroup" May 8 03:08:19.187267 kubelet[2746]: E0508 03:08:19.187258 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="apply-sysctl-overwrites" May 8 03:08:19.187319 kubelet[2746]: E0508 03:08:19.187310 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="mount-bpf-fs" May 8 03:08:19.187375 kubelet[2746]: E0508 03:08:19.187357 2746 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="64bda319-ec49-409f-8ae5-603c3dfaf0b9" containerName="cilium-operator" May 8 03:08:19.188148 kubelet[2746]: I0508 03:08:19.187438 2746 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15fe2b0-8a4d-4808-a529-dda616ef1cb6" containerName="cilium-agent" May 8 03:08:19.188148 kubelet[2746]: I0508 03:08:19.187449 2746 memory_manager.go:354] "RemoveStaleState removing state" podUID="64bda319-ec49-409f-8ae5-603c3dfaf0b9" containerName="cilium-operator" May 8 03:08:19.188092 systemd[1]: sshd@24-10.0.0.7:22-10.0.0.1:47540.service: Deactivated successfully. May 8 03:08:19.191297 systemd[1]: session-25.scope: Deactivated successfully. May 8 03:08:19.195162 systemd-logind[1481]: Session 25 logged out. Waiting for processes to exit. May 8 03:08:19.200998 systemd[1]: Started sshd@25-10.0.0.7:22-10.0.0.1:47554.service - OpenSSH per-connection server daemon (10.0.0.1:47554). May 8 03:08:19.206308 systemd-logind[1481]: Removed session 25. May 8 03:08:19.217773 systemd[1]: Created slice kubepods-burstable-pod687509c4_a361_479e_a5fe_9c8609ab71e9.slice - libcontainer container kubepods-burstable-pod687509c4_a361_479e_a5fe_9c8609ab71e9.slice. May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260211 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-host-proc-sys-net\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260252 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drqn4\" (UniqueName: \"kubernetes.io/projected/687509c4-a361-479e-a5fe-9c8609ab71e9-kube-api-access-drqn4\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260274 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-bpf-maps\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260291 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-host-proc-sys-kernel\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260307 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-etc-cni-netd\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262582 kubelet[2746]: I0508 03:08:19.260330 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-cilium-cgroup\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260344 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-cni-path\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260366 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-xtables-lock\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260384 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/687509c4-a361-479e-a5fe-9c8609ab71e9-clustermesh-secrets\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260400 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/687509c4-a361-479e-a5fe-9c8609ab71e9-cilium-ipsec-secrets\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260415 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/687509c4-a361-479e-a5fe-9c8609ab71e9-hubble-tls\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262822 kubelet[2746]: I0508 03:08:19.260431 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-hostproc\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262944 kubelet[2746]: I0508 03:08:19.260446 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-lib-modules\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262944 kubelet[2746]: I0508 03:08:19.260461 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/687509c4-a361-479e-a5fe-9c8609ab71e9-cilium-config-path\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.262944 kubelet[2746]: I0508 03:08:19.260476 2746 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/687509c4-a361-479e-a5fe-9c8609ab71e9-cilium-run\") pod \"cilium-j4lld\" (UID: \"687509c4-a361-479e-a5fe-9c8609ab71e9\") " pod="kube-system/cilium-j4lld" May 8 03:08:19.276533 sshd[4528]: Accepted publickey for core from 10.0.0.1 port 47554 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:19.277882 sshd-session[4528]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:19.281540 systemd-logind[1481]: New session 26 of user core. May 8 03:08:19.288888 systemd[1]: Started session-26.scope - Session 26 of User core. May 8 03:08:19.337537 sshd[4531]: Connection closed by 10.0.0.1 port 47554 May 8 03:08:19.338791 sshd-session[4528]: pam_unix(sshd:session): session closed for user core May 8 03:08:19.347985 systemd[1]: sshd@25-10.0.0.7:22-10.0.0.1:47554.service: Deactivated successfully. May 8 03:08:19.349487 systemd[1]: session-26.scope: Deactivated successfully. May 8 03:08:19.350247 systemd-logind[1481]: Session 26 logged out. Waiting for processes to exit. May 8 03:08:19.352154 systemd[1]: Started sshd@26-10.0.0.7:22-10.0.0.1:47570.service - OpenSSH per-connection server daemon (10.0.0.1:47570). May 8 03:08:19.353046 systemd-logind[1481]: Removed session 26. May 8 03:08:19.398750 sshd[4537]: Accepted publickey for core from 10.0.0.1 port 47570 ssh2: RSA SHA256:cWacyJtaZbzObjv3vWfzlr+7uozH1ItWFzOaJiT28ZY May 8 03:08:19.399937 sshd-session[4537]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 03:08:19.403605 systemd-logind[1481]: New session 27 of user core. May 8 03:08:19.416873 systemd[1]: Started session-27.scope - Session 27 of User core. May 8 03:08:19.523825 kubelet[2746]: E0508 03:08:19.523785 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:19.524605 containerd[1496]: time="2025-05-08T03:08:19.524319902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-j4lld,Uid:687509c4-a361-479e-a5fe-9c8609ab71e9,Namespace:kube-system,Attempt:0,}" May 8 03:08:19.543413 containerd[1496]: time="2025-05-08T03:08:19.543225077Z" level=info msg="connecting to shim f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" namespace=k8s.io protocol=ttrpc version=3 May 8 03:08:19.566881 systemd[1]: Started cri-containerd-f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c.scope - libcontainer container f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c. May 8 03:08:19.589113 containerd[1496]: time="2025-05-08T03:08:19.589071732Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-j4lld,Uid:687509c4-a361-479e-a5fe-9c8609ab71e9,Namespace:kube-system,Attempt:0,} returns sandbox id \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\"" May 8 03:08:19.589997 kubelet[2746]: E0508 03:08:19.589966 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:19.593444 containerd[1496]: time="2025-05-08T03:08:19.593260304Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 8 03:08:19.598182 containerd[1496]: time="2025-05-08T03:08:19.598141878Z" level=info msg="Container c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d: CDI devices from CRI Config.CDIDevices: []" May 8 03:08:19.605206 containerd[1496]: time="2025-05-08T03:08:19.605163579Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\"" May 8 03:08:19.605879 containerd[1496]: time="2025-05-08T03:08:19.605851821Z" level=info msg="StartContainer for \"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\"" May 8 03:08:19.606657 containerd[1496]: time="2025-05-08T03:08:19.606628183Z" level=info msg="connecting to shim c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" protocol=ttrpc version=3 May 8 03:08:19.623959 systemd[1]: Started cri-containerd-c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d.scope - libcontainer container c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d. May 8 03:08:19.651077 containerd[1496]: time="2025-05-08T03:08:19.651039233Z" level=info msg="StartContainer for \"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\" returns successfully" May 8 03:08:19.662242 systemd[1]: cri-containerd-c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d.scope: Deactivated successfully. May 8 03:08:19.663441 containerd[1496]: time="2025-05-08T03:08:19.663374389Z" level=info msg="received exit event container_id:\"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\" id:\"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\" pid:4610 exited_at:{seconds:1746673699 nanos:663092828}" May 8 03:08:19.663562 containerd[1496]: time="2025-05-08T03:08:19.663412349Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\" id:\"c18a7045e6cb398c05a9e0ba6dbab3db13e32899ebb9663af5f62e32008d951d\" pid:4610 exited_at:{seconds:1746673699 nanos:663092828}" May 8 03:08:19.873697 kubelet[2746]: E0508 03:08:19.873595 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:19.877431 containerd[1496]: time="2025-05-08T03:08:19.877394416Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 8 03:08:19.883087 containerd[1496]: time="2025-05-08T03:08:19.883051353Z" level=info msg="Container 227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32: CDI devices from CRI Config.CDIDevices: []" May 8 03:08:19.889878 containerd[1496]: time="2025-05-08T03:08:19.889841253Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\"" May 8 03:08:19.892918 containerd[1496]: time="2025-05-08T03:08:19.892876342Z" level=info msg="StartContainer for \"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\"" May 8 03:08:19.893873 containerd[1496]: time="2025-05-08T03:08:19.893618504Z" level=info msg="connecting to shim 227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" protocol=ttrpc version=3 May 8 03:08:19.919927 systemd[1]: Started cri-containerd-227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32.scope - libcontainer container 227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32. May 8 03:08:19.942593 containerd[1496]: time="2025-05-08T03:08:19.942499967Z" level=info msg="StartContainer for \"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\" returns successfully" May 8 03:08:19.951290 systemd[1]: cri-containerd-227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32.scope: Deactivated successfully. May 8 03:08:19.952572 containerd[1496]: time="2025-05-08T03:08:19.952537436Z" level=info msg="received exit event container_id:\"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\" id:\"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\" pid:4655 exited_at:{seconds:1746673699 nanos:952013875}" May 8 03:08:19.952644 containerd[1496]: time="2025-05-08T03:08:19.952620917Z" level=info msg="TaskExit event in podsandbox handler container_id:\"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\" id:\"227cf9cfe848497bfe49d6fa100eb6267b76efd58b7e0e54e271167e0591ff32\" pid:4655 exited_at:{seconds:1746673699 nanos:952013875}" May 8 03:08:20.878689 kubelet[2746]: E0508 03:08:20.878659 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:20.881380 containerd[1496]: time="2025-05-08T03:08:20.880870332Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 8 03:08:20.918372 containerd[1496]: time="2025-05-08T03:08:20.918313159Z" level=info msg="Container 438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e: CDI devices from CRI Config.CDIDevices: []" May 8 03:08:20.926150 containerd[1496]: time="2025-05-08T03:08:20.926096141Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\"" May 8 03:08:20.926609 containerd[1496]: time="2025-05-08T03:08:20.926580543Z" level=info msg="StartContainer for \"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\"" May 8 03:08:20.927949 containerd[1496]: time="2025-05-08T03:08:20.927922387Z" level=info msg="connecting to shim 438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" protocol=ttrpc version=3 May 8 03:08:20.950904 systemd[1]: Started cri-containerd-438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e.scope - libcontainer container 438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e. May 8 03:08:20.984678 systemd[1]: cri-containerd-438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e.scope: Deactivated successfully. May 8 03:08:20.986432 containerd[1496]: time="2025-05-08T03:08:20.985980592Z" level=info msg="TaskExit event in podsandbox handler container_id:\"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\" id:\"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\" pid:4700 exited_at:{seconds:1746673700 nanos:984508788}" May 8 03:08:20.986432 containerd[1496]: time="2025-05-08T03:08:20.986054353Z" level=info msg="StartContainer for \"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\" returns successfully" May 8 03:08:20.986432 containerd[1496]: time="2025-05-08T03:08:20.986063233Z" level=info msg="received exit event container_id:\"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\" id:\"438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e\" pid:4700 exited_at:{seconds:1746673700 nanos:984508788}" May 8 03:08:21.005599 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-438441c071dbc3a96ee9f7614a7064092ec16616ff8f6badd62523e3623f0b2e-rootfs.mount: Deactivated successfully. May 8 03:08:21.882818 kubelet[2746]: E0508 03:08:21.882790 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:21.886811 containerd[1496]: time="2025-05-08T03:08:21.886354742Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 8 03:08:21.894495 containerd[1496]: time="2025-05-08T03:08:21.894063364Z" level=info msg="Container 8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011: CDI devices from CRI Config.CDIDevices: []" May 8 03:08:21.902420 containerd[1496]: time="2025-05-08T03:08:21.902374067Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\"" May 8 03:08:21.902797 containerd[1496]: time="2025-05-08T03:08:21.902774908Z" level=info msg="StartContainer for \"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\"" May 8 03:08:21.903497 containerd[1496]: time="2025-05-08T03:08:21.903474310Z" level=info msg="connecting to shim 8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" protocol=ttrpc version=3 May 8 03:08:21.923881 systemd[1]: Started cri-containerd-8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011.scope - libcontainer container 8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011. May 8 03:08:21.945759 systemd[1]: cri-containerd-8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011.scope: Deactivated successfully. May 8 03:08:21.947302 containerd[1496]: time="2025-05-08T03:08:21.946922631Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\" id:\"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\" pid:4739 exited_at:{seconds:1746673701 nanos:945976469}" May 8 03:08:21.949151 containerd[1496]: time="2025-05-08T03:08:21.949117717Z" level=info msg="received exit event container_id:\"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\" id:\"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\" pid:4739 exited_at:{seconds:1746673701 nanos:945976469}" May 8 03:08:21.956523 containerd[1496]: time="2025-05-08T03:08:21.956458938Z" level=info msg="StartContainer for \"8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011\" returns successfully" May 8 03:08:21.967525 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8709069321c3b9e80c89d9184ed280eba497fe9641bfc4a5feb821f147744011-rootfs.mount: Deactivated successfully. May 8 03:08:22.634462 kubelet[2746]: E0508 03:08:22.634411 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:22.890113 kubelet[2746]: E0508 03:08:22.890011 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:22.896753 containerd[1496]: time="2025-05-08T03:08:22.894066050Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 8 03:08:23.070796 containerd[1496]: time="2025-05-08T03:08:23.070005443Z" level=info msg="Container 30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf: CDI devices from CRI Config.CDIDevices: []" May 8 03:08:23.089826 containerd[1496]: time="2025-05-08T03:08:23.089787576Z" level=info msg="CreateContainer within sandbox \"f8feb6b6d361de309105a307b6cc156e2c0a8698c409c69093c4af109bea816c\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\"" May 8 03:08:23.091057 containerd[1496]: time="2025-05-08T03:08:23.091026899Z" level=info msg="StartContainer for \"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\"" May 8 03:08:23.091904 containerd[1496]: time="2025-05-08T03:08:23.091855061Z" level=info msg="connecting to shim 30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf" address="unix:///run/containerd/s/5024769ca5b0c65364653a888d32d9606dcc12c86815f5af0d584bca15bab134" protocol=ttrpc version=3 May 8 03:08:23.117901 systemd[1]: Started cri-containerd-30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf.scope - libcontainer container 30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf. May 8 03:08:23.161636 containerd[1496]: time="2025-05-08T03:08:23.161343006Z" level=info msg="StartContainer for \"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" returns successfully" May 8 03:08:23.212037 containerd[1496]: time="2025-05-08T03:08:23.211989140Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" id:\"9a98f3bbd967c033e5697006447826455e6338c5f914fbeb1b34bcbf668a09c7\" pid:4808 exited_at:{seconds:1746673703 nanos:211610179}" May 8 03:08:23.405771 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) May 8 03:08:23.897302 kubelet[2746]: E0508 03:08:23.897271 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:23.922493 kubelet[2746]: I0508 03:08:23.922428 2746 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-j4lld" podStartSLOduration=4.922413944 podStartE2EDuration="4.922413944s" podCreationTimestamp="2025-05-08 03:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 03:08:23.922117783 +0000 UTC m=+90.370421966" watchObservedRunningTime="2025-05-08 03:08:23.922413944 +0000 UTC m=+90.370718127" May 8 03:08:25.526558 kubelet[2746]: E0508 03:08:25.525906 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:25.787903 containerd[1496]: time="2025-05-08T03:08:25.787614968Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" id:\"1a8911053265e5559f2017f85f0883e16ec3e8925b413a1b7cc2a5b99e51ef8a\" pid:5201 exit_status:1 exited_at:{seconds:1746673705 nanos:787229767}" May 8 03:08:26.218827 systemd-networkd[1424]: lxc_health: Link UP May 8 03:08:26.231830 systemd-networkd[1424]: lxc_health: Gained carrier May 8 03:08:27.526782 kubelet[2746]: E0508 03:08:27.526749 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:27.763909 systemd-networkd[1424]: lxc_health: Gained IPv6LL May 8 03:08:27.903534 kubelet[2746]: E0508 03:08:27.903130 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:27.942099 containerd[1496]: time="2025-05-08T03:08:27.941171517Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" id:\"56dec08dd8ae044bf011a16c500f374be8ae930fc13b45ea9b31520c072cdd24\" pid:5348 exited_at:{seconds:1746673707 nanos:940896996}" May 8 03:08:28.904641 kubelet[2746]: E0508 03:08:28.904431 2746 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 03:08:30.038156 containerd[1496]: time="2025-05-08T03:08:30.038103033Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" id:\"607796624f47b4f169070cd21c178b77eca9c439d4de112db5718cea7fd29148\" pid:5380 exited_at:{seconds:1746673710 nanos:37811792}" May 8 03:08:32.187875 containerd[1496]: time="2025-05-08T03:08:32.187832866Z" level=info msg="TaskExit event in podsandbox handler container_id:\"30aa786c9c00bda9f253792d770c35bb8c3ee404b37cba234a0374ce884f45cf\" id:\"819ad7adcd861d31649e2dd9f75e5f4ffa57b7748027a2e6e6778ec814ddacd6\" pid:5405 exited_at:{seconds:1746673712 nanos:187505386}" May 8 03:08:32.197548 sshd[4544]: Connection closed by 10.0.0.1 port 47570 May 8 03:08:32.198056 sshd-session[4537]: pam_unix(sshd:session): session closed for user core May 8 03:08:32.201048 systemd[1]: sshd@26-10.0.0.7:22-10.0.0.1:47570.service: Deactivated successfully. May 8 03:08:32.202937 systemd[1]: session-27.scope: Deactivated successfully. May 8 03:08:32.204462 systemd-logind[1481]: Session 27 logged out. Waiting for processes to exit. May 8 03:08:32.205581 systemd-logind[1481]: Removed session 27.