Sep 8 23:51:03.791209 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 8 23:51:03.791231 kernel: Linux version 6.12.45-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Mon Sep 8 22:10:01 -00 2025 Sep 8 23:51:03.791241 kernel: KASLR enabled Sep 8 23:51:03.791246 kernel: efi: EFI v2.7 by EDK II Sep 8 23:51:03.791252 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb832018 ACPI 2.0=0xdbfd0018 RNG=0xdbfd0a18 MEMRESERVE=0xdb838218 Sep 8 23:51:03.791257 kernel: random: crng init done Sep 8 23:51:03.791264 kernel: secureboot: Secure boot disabled Sep 8 23:51:03.791270 kernel: ACPI: Early table checksum verification disabled Sep 8 23:51:03.791276 kernel: ACPI: RSDP 0x00000000DBFD0018 000024 (v02 BOCHS ) Sep 8 23:51:03.791283 kernel: ACPI: XSDT 0x00000000DBFD0F18 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 8 23:51:03.791288 kernel: ACPI: FACP 0x00000000DBFD0B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791294 kernel: ACPI: DSDT 0x00000000DBF0E018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791300 kernel: ACPI: APIC 0x00000000DBFD0C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791306 kernel: ACPI: PPTT 0x00000000DBFD0098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791313 kernel: ACPI: GTDT 0x00000000DBFD0818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791320 kernel: ACPI: MCFG 0x00000000DBFD0A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791326 kernel: ACPI: SPCR 0x00000000DBFD0918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791332 kernel: ACPI: DBG2 0x00000000DBFD0998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791338 kernel: ACPI: IORT 0x00000000DBFD0198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 8 23:51:03.791344 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 8 23:51:03.791350 kernel: ACPI: Use ACPI SPCR as default console: No Sep 8 23:51:03.791365 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 8 23:51:03.791372 kernel: NODE_DATA(0) allocated [mem 0xdc965a00-0xdc96cfff] Sep 8 23:51:03.791378 kernel: Zone ranges: Sep 8 23:51:03.791384 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 8 23:51:03.791391 kernel: DMA32 empty Sep 8 23:51:03.791397 kernel: Normal empty Sep 8 23:51:03.791403 kernel: Device empty Sep 8 23:51:03.791409 kernel: Movable zone start for each node Sep 8 23:51:03.791415 kernel: Early memory node ranges Sep 8 23:51:03.791421 kernel: node 0: [mem 0x0000000040000000-0x00000000db81ffff] Sep 8 23:51:03.791427 kernel: node 0: [mem 0x00000000db820000-0x00000000db82ffff] Sep 8 23:51:03.791433 kernel: node 0: [mem 0x00000000db830000-0x00000000dc09ffff] Sep 8 23:51:03.791439 kernel: node 0: [mem 0x00000000dc0a0000-0x00000000dc2dffff] Sep 8 23:51:03.791445 kernel: node 0: [mem 0x00000000dc2e0000-0x00000000dc36ffff] Sep 8 23:51:03.791451 kernel: node 0: [mem 0x00000000dc370000-0x00000000dc45ffff] Sep 8 23:51:03.791457 kernel: node 0: [mem 0x00000000dc460000-0x00000000dc52ffff] Sep 8 23:51:03.791465 kernel: node 0: [mem 0x00000000dc530000-0x00000000dc5cffff] Sep 8 23:51:03.791470 kernel: node 0: [mem 0x00000000dc5d0000-0x00000000dce1ffff] Sep 8 23:51:03.791476 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Sep 8 23:51:03.791485 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Sep 8 23:51:03.791492 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Sep 8 23:51:03.791498 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Sep 8 23:51:03.791506 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 8 23:51:03.791513 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 8 23:51:03.791520 kernel: cma: Reserved 16 MiB at 0x00000000d8000000 on node -1 Sep 8 23:51:03.791526 kernel: psci: probing for conduit method from ACPI. Sep 8 23:51:03.791533 kernel: psci: PSCIv1.1 detected in firmware. Sep 8 23:51:03.791539 kernel: psci: Using standard PSCI v0.2 function IDs Sep 8 23:51:03.791545 kernel: psci: Trusted OS migration not required Sep 8 23:51:03.791552 kernel: psci: SMC Calling Convention v1.1 Sep 8 23:51:03.791558 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 8 23:51:03.791565 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Sep 8 23:51:03.791572 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Sep 8 23:51:03.791579 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 8 23:51:03.791585 kernel: Detected PIPT I-cache on CPU0 Sep 8 23:51:03.791592 kernel: CPU features: detected: GIC system register CPU interface Sep 8 23:51:03.791598 kernel: CPU features: detected: Spectre-v4 Sep 8 23:51:03.791605 kernel: CPU features: detected: Spectre-BHB Sep 8 23:51:03.791611 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 8 23:51:03.791617 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 8 23:51:03.791624 kernel: CPU features: detected: ARM erratum 1418040 Sep 8 23:51:03.791630 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 8 23:51:03.791637 kernel: alternatives: applying boot alternatives Sep 8 23:51:03.791644 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=56d35272d6799b20efe64172ddb761aa9d752bf4ee92cd36e6693ce5e7a3b83d Sep 8 23:51:03.791652 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 8 23:51:03.791658 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 8 23:51:03.791665 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 8 23:51:03.791671 kernel: Fallback order for Node 0: 0 Sep 8 23:51:03.791677 kernel: Built 1 zonelists, mobility grouping on. Total pages: 643072 Sep 8 23:51:03.791684 kernel: Policy zone: DMA Sep 8 23:51:03.791690 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 8 23:51:03.791712 kernel: software IO TLB: SWIOTLB bounce buffer size adjusted to 2MB Sep 8 23:51:03.791744 kernel: software IO TLB: area num 4. Sep 8 23:51:03.791759 kernel: software IO TLB: SWIOTLB bounce buffer size roundup to 4MB Sep 8 23:51:03.791765 kernel: software IO TLB: mapped [mem 0x00000000d7c00000-0x00000000d8000000] (4MB) Sep 8 23:51:03.791777 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 8 23:51:03.791787 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 8 23:51:03.791797 kernel: rcu: RCU event tracing is enabled. Sep 8 23:51:03.791810 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 8 23:51:03.791817 kernel: Trampoline variant of Tasks RCU enabled. Sep 8 23:51:03.791824 kernel: Tracing variant of Tasks RCU enabled. Sep 8 23:51:03.791832 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 8 23:51:03.791839 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 8 23:51:03.791846 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 8 23:51:03.791853 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Sep 8 23:51:03.791860 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 8 23:51:03.791868 kernel: GICv3: 256 SPIs implemented Sep 8 23:51:03.791875 kernel: GICv3: 0 Extended SPIs implemented Sep 8 23:51:03.791882 kernel: Root IRQ handler: gic_handle_irq Sep 8 23:51:03.791889 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Sep 8 23:51:03.791903 kernel: GICv3: GICD_CTRL.DS=1, SCR_EL3.FIQ=0 Sep 8 23:51:03.791910 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 8 23:51:03.791916 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 8 23:51:03.791923 kernel: ITS@0x0000000008080000: allocated 8192 Devices @40110000 (indirect, esz 8, psz 64K, shr 1) Sep 8 23:51:03.791930 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @40120000 (flat, esz 8, psz 64K, shr 1) Sep 8 23:51:03.791937 kernel: GICv3: using LPI property table @0x0000000040130000 Sep 8 23:51:03.791944 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040140000 Sep 8 23:51:03.791951 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 8 23:51:03.791959 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 8 23:51:03.791966 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 8 23:51:03.791977 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 8 23:51:03.791985 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 8 23:51:03.791991 kernel: arm-pv: using stolen time PV Sep 8 23:51:03.791998 kernel: Console: colour dummy device 80x25 Sep 8 23:51:03.792005 kernel: ACPI: Core revision 20240827 Sep 8 23:51:03.792012 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 8 23:51:03.792019 kernel: pid_max: default: 32768 minimum: 301 Sep 8 23:51:03.792025 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 8 23:51:03.792033 kernel: landlock: Up and running. Sep 8 23:51:03.792085 kernel: SELinux: Initializing. Sep 8 23:51:03.792092 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 8 23:51:03.792100 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 8 23:51:03.792113 kernel: rcu: Hierarchical SRCU implementation. Sep 8 23:51:03.792120 kernel: rcu: Max phase no-delay instances is 400. Sep 8 23:51:03.792130 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 8 23:51:03.792142 kernel: Remapping and enabling EFI services. Sep 8 23:51:03.792148 kernel: smp: Bringing up secondary CPUs ... Sep 8 23:51:03.792170 kernel: Detected PIPT I-cache on CPU1 Sep 8 23:51:03.792187 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 8 23:51:03.792194 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040150000 Sep 8 23:51:03.792202 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 8 23:51:03.792209 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 8 23:51:03.792216 kernel: Detected PIPT I-cache on CPU2 Sep 8 23:51:03.792223 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 8 23:51:03.792230 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040160000 Sep 8 23:51:03.792238 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 8 23:51:03.792245 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 8 23:51:03.792252 kernel: Detected PIPT I-cache on CPU3 Sep 8 23:51:03.792260 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 8 23:51:03.792268 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040170000 Sep 8 23:51:03.792275 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 8 23:51:03.792282 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 8 23:51:03.792289 kernel: smp: Brought up 1 node, 4 CPUs Sep 8 23:51:03.792296 kernel: SMP: Total of 4 processors activated. Sep 8 23:51:03.792305 kernel: CPU: All CPU(s) started at EL1 Sep 8 23:51:03.792312 kernel: CPU features: detected: 32-bit EL0 Support Sep 8 23:51:03.792319 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 8 23:51:03.792326 kernel: CPU features: detected: Common not Private translations Sep 8 23:51:03.792334 kernel: CPU features: detected: CRC32 instructions Sep 8 23:51:03.792341 kernel: CPU features: detected: Enhanced Virtualization Traps Sep 8 23:51:03.792348 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 8 23:51:03.792355 kernel: CPU features: detected: LSE atomic instructions Sep 8 23:51:03.792370 kernel: CPU features: detected: Privileged Access Never Sep 8 23:51:03.792379 kernel: CPU features: detected: RAS Extension Support Sep 8 23:51:03.792387 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 8 23:51:03.792394 kernel: alternatives: applying system-wide alternatives Sep 8 23:51:03.792401 kernel: CPU features: detected: Hardware dirty bit management on CPU0-3 Sep 8 23:51:03.792408 kernel: Memory: 2424544K/2572288K available (11136K kernel code, 2436K rwdata, 9060K rodata, 38912K init, 1038K bss, 125408K reserved, 16384K cma-reserved) Sep 8 23:51:03.792415 kernel: devtmpfs: initialized Sep 8 23:51:03.792422 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 8 23:51:03.792429 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 8 23:51:03.792436 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Sep 8 23:51:03.792449 kernel: 0 pages in range for non-PLT usage Sep 8 23:51:03.792456 kernel: 508576 pages in range for PLT usage Sep 8 23:51:03.792463 kernel: pinctrl core: initialized pinctrl subsystem Sep 8 23:51:03.792470 kernel: SMBIOS 3.0.0 present. Sep 8 23:51:03.792477 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Sep 8 23:51:03.792484 kernel: DMI: Memory slots populated: 1/1 Sep 8 23:51:03.792491 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 8 23:51:03.792498 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 8 23:51:03.792505 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 8 23:51:03.792526 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 8 23:51:03.792533 kernel: audit: initializing netlink subsys (disabled) Sep 8 23:51:03.792540 kernel: audit: type=2000 audit(0.025:1): state=initialized audit_enabled=0 res=1 Sep 8 23:51:03.792547 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 8 23:51:03.792554 kernel: cpuidle: using governor menu Sep 8 23:51:03.792561 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 8 23:51:03.792568 kernel: ASID allocator initialised with 32768 entries Sep 8 23:51:03.792575 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 8 23:51:03.792582 kernel: Serial: AMBA PL011 UART driver Sep 8 23:51:03.792590 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 8 23:51:03.792597 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Sep 8 23:51:03.792604 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Sep 8 23:51:03.792611 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Sep 8 23:51:03.792618 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 8 23:51:03.792625 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Sep 8 23:51:03.792632 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Sep 8 23:51:03.792639 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Sep 8 23:51:03.792646 kernel: ACPI: Added _OSI(Module Device) Sep 8 23:51:03.792654 kernel: ACPI: Added _OSI(Processor Device) Sep 8 23:51:03.792661 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 8 23:51:03.792668 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 8 23:51:03.792675 kernel: ACPI: Interpreter enabled Sep 8 23:51:03.792682 kernel: ACPI: Using GIC for interrupt routing Sep 8 23:51:03.792689 kernel: ACPI: MCFG table detected, 1 entries Sep 8 23:51:03.792698 kernel: ACPI: CPU0 has been hot-added Sep 8 23:51:03.792706 kernel: ACPI: CPU1 has been hot-added Sep 8 23:51:03.792713 kernel: ACPI: CPU2 has been hot-added Sep 8 23:51:03.792720 kernel: ACPI: CPU3 has been hot-added Sep 8 23:51:03.792730 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 8 23:51:03.792737 kernel: printk: legacy console [ttyAMA0] enabled Sep 8 23:51:03.792745 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 8 23:51:03.792893 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 8 23:51:03.792972 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 8 23:51:03.793033 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 8 23:51:03.793123 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 8 23:51:03.793187 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 8 23:51:03.793196 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 8 23:51:03.793203 kernel: PCI host bridge to bus 0000:00 Sep 8 23:51:03.793276 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 8 23:51:03.793328 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 8 23:51:03.793393 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 8 23:51:03.793448 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 8 23:51:03.793527 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 conventional PCI endpoint Sep 8 23:51:03.793595 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Sep 8 23:51:03.793656 kernel: pci 0000:00:01.0: BAR 0 [io 0x0000-0x001f] Sep 8 23:51:03.793716 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff] Sep 8 23:51:03.793775 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref] Sep 8 23:51:03.793833 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8000000000-0x8000003fff 64bit pref]: assigned Sep 8 23:51:03.793891 kernel: pci 0000:00:01.0: BAR 1 [mem 0x10000000-0x10000fff]: assigned Sep 8 23:51:03.793952 kernel: pci 0000:00:01.0: BAR 0 [io 0x1000-0x101f]: assigned Sep 8 23:51:03.794006 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 8 23:51:03.794075 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 8 23:51:03.794138 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 8 23:51:03.794150 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 8 23:51:03.794158 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 8 23:51:03.794165 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 8 23:51:03.794176 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 8 23:51:03.794183 kernel: iommu: Default domain type: Translated Sep 8 23:51:03.794190 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 8 23:51:03.794197 kernel: efivars: Registered efivars operations Sep 8 23:51:03.794204 kernel: vgaarb: loaded Sep 8 23:51:03.794212 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 8 23:51:03.794219 kernel: VFS: Disk quotas dquot_6.6.0 Sep 8 23:51:03.794227 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 8 23:51:03.794235 kernel: pnp: PnP ACPI init Sep 8 23:51:03.794305 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 8 23:51:03.794316 kernel: pnp: PnP ACPI: found 1 devices Sep 8 23:51:03.794323 kernel: NET: Registered PF_INET protocol family Sep 8 23:51:03.794330 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 8 23:51:03.794337 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 8 23:51:03.794345 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 8 23:51:03.794353 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 8 23:51:03.794370 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 8 23:51:03.794381 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 8 23:51:03.794388 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 8 23:51:03.794395 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 8 23:51:03.794402 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 8 23:51:03.794409 kernel: PCI: CLS 0 bytes, default 64 Sep 8 23:51:03.794416 kernel: kvm [1]: HYP mode not available Sep 8 23:51:03.794423 kernel: Initialise system trusted keyrings Sep 8 23:51:03.794431 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 8 23:51:03.794438 kernel: Key type asymmetric registered Sep 8 23:51:03.794446 kernel: Asymmetric key parser 'x509' registered Sep 8 23:51:03.794453 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 8 23:51:03.794460 kernel: io scheduler mq-deadline registered Sep 8 23:51:03.794467 kernel: io scheduler kyber registered Sep 8 23:51:03.794474 kernel: io scheduler bfq registered Sep 8 23:51:03.794481 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 8 23:51:03.794488 kernel: ACPI: button: Power Button [PWRB] Sep 8 23:51:03.794495 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 8 23:51:03.794572 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 8 23:51:03.794584 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 8 23:51:03.794591 kernel: thunder_xcv, ver 1.0 Sep 8 23:51:03.794598 kernel: thunder_bgx, ver 1.0 Sep 8 23:51:03.794605 kernel: nicpf, ver 1.0 Sep 8 23:51:03.794612 kernel: nicvf, ver 1.0 Sep 8 23:51:03.794685 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 8 23:51:03.794742 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-09-08T23:51:03 UTC (1757375463) Sep 8 23:51:03.794751 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 8 23:51:03.794760 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Sep 8 23:51:03.794767 kernel: watchdog: NMI not fully supported Sep 8 23:51:03.794774 kernel: watchdog: Hard watchdog permanently disabled Sep 8 23:51:03.794781 kernel: NET: Registered PF_INET6 protocol family Sep 8 23:51:03.794788 kernel: Segment Routing with IPv6 Sep 8 23:51:03.794794 kernel: In-situ OAM (IOAM) with IPv6 Sep 8 23:51:03.794801 kernel: NET: Registered PF_PACKET protocol family Sep 8 23:51:03.794808 kernel: Key type dns_resolver registered Sep 8 23:51:03.794815 kernel: registered taskstats version 1 Sep 8 23:51:03.794822 kernel: Loading compiled-in X.509 certificates Sep 8 23:51:03.794831 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.45-flatcar: a394eaa34ffd7f1371a823c439a0662c32ae9397' Sep 8 23:51:03.794838 kernel: Demotion targets for Node 0: null Sep 8 23:51:03.794845 kernel: Key type .fscrypt registered Sep 8 23:51:03.794852 kernel: Key type fscrypt-provisioning registered Sep 8 23:51:03.794859 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 8 23:51:03.794866 kernel: ima: Allocated hash algorithm: sha1 Sep 8 23:51:03.794873 kernel: ima: No architecture policies found Sep 8 23:51:03.794880 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 8 23:51:03.794889 kernel: clk: Disabling unused clocks Sep 8 23:51:03.794896 kernel: PM: genpd: Disabling unused power domains Sep 8 23:51:03.794903 kernel: Warning: unable to open an initial console. Sep 8 23:51:03.794910 kernel: Freeing unused kernel memory: 38912K Sep 8 23:51:03.794917 kernel: Run /init as init process Sep 8 23:51:03.794924 kernel: with arguments: Sep 8 23:51:03.794931 kernel: /init Sep 8 23:51:03.794938 kernel: with environment: Sep 8 23:51:03.794945 kernel: HOME=/ Sep 8 23:51:03.794952 kernel: TERM=linux Sep 8 23:51:03.794961 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 8 23:51:03.794969 systemd[1]: Successfully made /usr/ read-only. Sep 8 23:51:03.794980 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 8 23:51:03.794988 systemd[1]: Detected virtualization kvm. Sep 8 23:51:03.794996 systemd[1]: Detected architecture arm64. Sep 8 23:51:03.795003 systemd[1]: Running in initrd. Sep 8 23:51:03.795011 systemd[1]: No hostname configured, using default hostname. Sep 8 23:51:03.795020 systemd[1]: Hostname set to . Sep 8 23:51:03.795028 systemd[1]: Initializing machine ID from VM UUID. Sep 8 23:51:03.795035 systemd[1]: Queued start job for default target initrd.target. Sep 8 23:51:03.795071 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 8 23:51:03.795080 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 8 23:51:03.795089 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 8 23:51:03.795096 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 8 23:51:03.795104 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 8 23:51:03.795116 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 8 23:51:03.795124 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 8 23:51:03.795132 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 8 23:51:03.795140 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 8 23:51:03.795148 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 8 23:51:03.795156 systemd[1]: Reached target paths.target - Path Units. Sep 8 23:51:03.795163 systemd[1]: Reached target slices.target - Slice Units. Sep 8 23:51:03.795173 systemd[1]: Reached target swap.target - Swaps. Sep 8 23:51:03.795181 systemd[1]: Reached target timers.target - Timer Units. Sep 8 23:51:03.795197 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 8 23:51:03.795205 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 8 23:51:03.795213 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 8 23:51:03.795221 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 8 23:51:03.795229 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 8 23:51:03.795237 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 8 23:51:03.795246 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 8 23:51:03.795254 systemd[1]: Reached target sockets.target - Socket Units. Sep 8 23:51:03.795262 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 8 23:51:03.795270 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 8 23:51:03.795278 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 8 23:51:03.795286 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 8 23:51:03.795294 systemd[1]: Starting systemd-fsck-usr.service... Sep 8 23:51:03.795302 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 8 23:51:03.795309 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 8 23:51:03.795319 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 8 23:51:03.795327 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 8 23:51:03.795335 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 8 23:51:03.795383 systemd-journald[245]: Collecting audit messages is disabled. Sep 8 23:51:03.795419 systemd[1]: Finished systemd-fsck-usr.service. Sep 8 23:51:03.795431 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 8 23:51:03.795439 systemd-journald[245]: Journal started Sep 8 23:51:03.795459 systemd-journald[245]: Runtime Journal (/run/log/journal/a38e81aaad9b4c3eb32f5e5652895cfd) is 6M, max 48.5M, 42.4M free. Sep 8 23:51:03.790157 systemd-modules-load[246]: Inserted module 'overlay' Sep 8 23:51:03.797442 systemd[1]: Started systemd-journald.service - Journal Service. Sep 8 23:51:03.798595 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 8 23:51:03.805272 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 8 23:51:03.805096 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 8 23:51:03.807466 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 8 23:51:03.809630 systemd-modules-load[246]: Inserted module 'br_netfilter' Sep 8 23:51:03.810456 kernel: Bridge firewalling registered Sep 8 23:51:03.814277 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 8 23:51:03.816422 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 8 23:51:03.819213 systemd-tmpfiles[267]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 8 23:51:03.821213 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 8 23:51:03.823246 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 8 23:51:03.824566 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 8 23:51:03.836204 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 8 23:51:03.840279 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 8 23:51:03.842398 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 8 23:51:03.845109 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 8 23:51:03.848807 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 8 23:51:03.857079 dracut-cmdline[287]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=56d35272d6799b20efe64172ddb761aa9d752bf4ee92cd36e6693ce5e7a3b83d Sep 8 23:51:03.890856 systemd-resolved[295]: Positive Trust Anchors: Sep 8 23:51:03.890875 systemd-resolved[295]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 8 23:51:03.890907 systemd-resolved[295]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 8 23:51:03.897453 systemd-resolved[295]: Defaulting to hostname 'linux'. Sep 8 23:51:03.898590 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 8 23:51:03.901007 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 8 23:51:03.949098 kernel: SCSI subsystem initialized Sep 8 23:51:03.953061 kernel: Loading iSCSI transport class v2.0-870. Sep 8 23:51:03.961077 kernel: iscsi: registered transport (tcp) Sep 8 23:51:03.974422 kernel: iscsi: registered transport (qla4xxx) Sep 8 23:51:03.974464 kernel: QLogic iSCSI HBA Driver Sep 8 23:51:03.992650 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 8 23:51:04.008959 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 8 23:51:04.011609 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 8 23:51:04.062290 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 8 23:51:04.064661 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 8 23:51:04.134093 kernel: raid6: neonx8 gen() 15471 MB/s Sep 8 23:51:04.151069 kernel: raid6: neonx4 gen() 15659 MB/s Sep 8 23:51:04.168067 kernel: raid6: neonx2 gen() 13193 MB/s Sep 8 23:51:04.185070 kernel: raid6: neonx1 gen() 10451 MB/s Sep 8 23:51:04.202078 kernel: raid6: int64x8 gen() 6875 MB/s Sep 8 23:51:04.219069 kernel: raid6: int64x4 gen() 7289 MB/s Sep 8 23:51:04.236069 kernel: raid6: int64x2 gen() 6035 MB/s Sep 8 23:51:04.253068 kernel: raid6: int64x1 gen() 4993 MB/s Sep 8 23:51:04.253097 kernel: raid6: using algorithm neonx4 gen() 15659 MB/s Sep 8 23:51:04.270065 kernel: raid6: .... xor() 12322 MB/s, rmw enabled Sep 8 23:51:04.270080 kernel: raid6: using neon recovery algorithm Sep 8 23:51:04.275406 kernel: xor: measuring software checksum speed Sep 8 23:51:04.275431 kernel: 8regs : 21601 MB/sec Sep 8 23:51:04.276057 kernel: 32regs : 21670 MB/sec Sep 8 23:51:04.277073 kernel: arm64_neon : 24683 MB/sec Sep 8 23:51:04.277089 kernel: xor: using function: arm64_neon (24683 MB/sec) Sep 8 23:51:04.330075 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 8 23:51:04.336736 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 8 23:51:04.339728 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 8 23:51:04.377179 systemd-udevd[501]: Using default interface naming scheme 'v255'. Sep 8 23:51:04.381254 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 8 23:51:04.383409 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 8 23:51:04.410965 dracut-pre-trigger[507]: rd.md=0: removing MD RAID activation Sep 8 23:51:04.435560 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 8 23:51:04.438283 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 8 23:51:04.502061 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 8 23:51:04.505804 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 8 23:51:04.549062 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Sep 8 23:51:04.555470 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 8 23:51:04.560527 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 8 23:51:04.560583 kernel: GPT:9289727 != 19775487 Sep 8 23:51:04.560594 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 8 23:51:04.560603 kernel: GPT:9289727 != 19775487 Sep 8 23:51:04.561602 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 8 23:51:04.561642 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 8 23:51:04.563412 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 8 23:51:04.563558 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 8 23:51:04.566527 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 8 23:51:04.568338 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 8 23:51:04.598973 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 8 23:51:04.607616 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 8 23:51:04.614016 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 8 23:51:04.621738 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 8 23:51:04.630446 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 8 23:51:04.636855 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 8 23:51:04.638308 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 8 23:51:04.641188 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 8 23:51:04.643242 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 8 23:51:04.645306 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 8 23:51:04.648097 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 8 23:51:04.649972 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 8 23:51:04.674884 disk-uuid[595]: Primary Header is updated. Sep 8 23:51:04.674884 disk-uuid[595]: Secondary Entries is updated. Sep 8 23:51:04.674884 disk-uuid[595]: Secondary Header is updated. Sep 8 23:51:04.680374 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 8 23:51:04.683611 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 8 23:51:05.688055 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 8 23:51:05.688121 disk-uuid[600]: The operation has completed successfully. Sep 8 23:51:05.716797 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 8 23:51:05.717161 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 8 23:51:05.745835 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 8 23:51:05.776271 sh[614]: Success Sep 8 23:51:05.789513 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 8 23:51:05.789575 kernel: device-mapper: uevent: version 1.0.3 Sep 8 23:51:05.791059 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 8 23:51:05.799067 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Sep 8 23:51:05.827506 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 8 23:51:05.829657 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 8 23:51:05.847960 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 8 23:51:05.855285 kernel: BTRFS: device fsid b6aa4556-53d3-40d0-8c29-11204db15da4 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (626) Sep 8 23:51:05.855320 kernel: BTRFS info (device dm-0): first mount of filesystem b6aa4556-53d3-40d0-8c29-11204db15da4 Sep 8 23:51:05.855331 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Sep 8 23:51:05.860063 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 8 23:51:05.860097 kernel: BTRFS info (device dm-0): enabling free space tree Sep 8 23:51:05.860790 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 8 23:51:05.862379 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 8 23:51:05.863985 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 8 23:51:05.864867 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 8 23:51:05.866783 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 8 23:51:05.895077 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (656) Sep 8 23:51:05.895162 kernel: BTRFS info (device vda6): first mount of filesystem 0ac87192-1b33-43df-818c-9161f04c3e9c Sep 8 23:51:05.896838 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 8 23:51:05.900065 kernel: BTRFS info (device vda6): turning on async discard Sep 8 23:51:05.900104 kernel: BTRFS info (device vda6): enabling free space tree Sep 8 23:51:05.904088 kernel: BTRFS info (device vda6): last unmount of filesystem 0ac87192-1b33-43df-818c-9161f04c3e9c Sep 8 23:51:05.904973 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 8 23:51:05.907128 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 8 23:51:05.979154 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 8 23:51:05.983237 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 8 23:51:06.021219 systemd-networkd[808]: lo: Link UP Sep 8 23:51:06.021232 systemd-networkd[808]: lo: Gained carrier Sep 8 23:51:06.022037 systemd-networkd[808]: Enumeration completed Sep 8 23:51:06.022273 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 8 23:51:06.023185 systemd-networkd[808]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 8 23:51:06.023189 systemd-networkd[808]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 8 23:51:06.024014 systemd[1]: Reached target network.target - Network. Sep 8 23:51:06.024033 systemd-networkd[808]: eth0: Link UP Sep 8 23:51:06.024214 systemd-networkd[808]: eth0: Gained carrier Sep 8 23:51:06.024225 systemd-networkd[808]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 8 23:51:06.034291 ignition[698]: Ignition 2.21.0 Sep 8 23:51:06.034308 ignition[698]: Stage: fetch-offline Sep 8 23:51:06.034349 ignition[698]: no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:06.034358 ignition[698]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:06.034581 ignition[698]: parsed url from cmdline: "" Sep 8 23:51:06.034585 ignition[698]: no config URL provided Sep 8 23:51:06.034590 ignition[698]: reading system config file "/usr/lib/ignition/user.ign" Sep 8 23:51:06.034598 ignition[698]: no config at "/usr/lib/ignition/user.ign" Sep 8 23:51:06.034624 ignition[698]: op(1): [started] loading QEMU firmware config module Sep 8 23:51:06.034630 ignition[698]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 8 23:51:06.048294 ignition[698]: op(1): [finished] loading QEMU firmware config module Sep 8 23:51:06.049113 systemd-networkd[808]: eth0: DHCPv4 address 10.0.0.130/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 8 23:51:06.092752 ignition[698]: parsing config with SHA512: 2c9e2982cbaf88cc7a74d595e450d6ecfd66b6c5be951d05882b198468123ed6d726408df488b1d69bdf920f8627074dd2f0e34082857ac1169b5cef97ac5b56 Sep 8 23:51:06.097549 unknown[698]: fetched base config from "system" Sep 8 23:51:06.097564 unknown[698]: fetched user config from "qemu" Sep 8 23:51:06.097915 ignition[698]: fetch-offline: fetch-offline passed Sep 8 23:51:06.099657 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 8 23:51:06.097969 ignition[698]: Ignition finished successfully Sep 8 23:51:06.103865 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 8 23:51:06.104848 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 8 23:51:06.134186 ignition[815]: Ignition 2.21.0 Sep 8 23:51:06.134202 ignition[815]: Stage: kargs Sep 8 23:51:06.134349 ignition[815]: no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:06.134359 ignition[815]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:06.136181 ignition[815]: kargs: kargs passed Sep 8 23:51:06.136242 ignition[815]: Ignition finished successfully Sep 8 23:51:06.140099 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 8 23:51:06.142933 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 8 23:51:06.168282 ignition[824]: Ignition 2.21.0 Sep 8 23:51:06.168298 ignition[824]: Stage: disks Sep 8 23:51:06.168465 ignition[824]: no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:06.168486 ignition[824]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:06.170287 ignition[824]: disks: disks passed Sep 8 23:51:06.172559 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 8 23:51:06.170361 ignition[824]: Ignition finished successfully Sep 8 23:51:06.173750 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 8 23:51:06.175009 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 8 23:51:06.176862 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 8 23:51:06.178318 systemd[1]: Reached target sysinit.target - System Initialization. Sep 8 23:51:06.180059 systemd[1]: Reached target basic.target - Basic System. Sep 8 23:51:06.183011 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 8 23:51:06.219216 systemd-resolved[295]: Detected conflict on linux IN A 10.0.0.130 Sep 8 23:51:06.219231 systemd-resolved[295]: Hostname conflict, changing published hostname from 'linux' to 'linux3'. Sep 8 23:51:06.222245 systemd-fsck[834]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 8 23:51:06.226540 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 8 23:51:06.228836 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 8 23:51:06.295069 kernel: EXT4-fs (vda9): mounted filesystem 12f0e8f7-98bc-449e-b11f-df07384be687 r/w with ordered data mode. Quota mode: none. Sep 8 23:51:06.295112 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 8 23:51:06.296457 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 8 23:51:06.300640 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 8 23:51:06.303077 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 8 23:51:06.304080 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 8 23:51:06.304122 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 8 23:51:06.304148 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 8 23:51:06.317876 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 8 23:51:06.320257 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 8 23:51:06.325066 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (843) Sep 8 23:51:06.325128 kernel: BTRFS info (device vda6): first mount of filesystem 0ac87192-1b33-43df-818c-9161f04c3e9c Sep 8 23:51:06.326833 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 8 23:51:06.332063 kernel: BTRFS info (device vda6): turning on async discard Sep 8 23:51:06.332097 kernel: BTRFS info (device vda6): enabling free space tree Sep 8 23:51:06.334626 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 8 23:51:06.358835 initrd-setup-root[867]: cut: /sysroot/etc/passwd: No such file or directory Sep 8 23:51:06.362877 initrd-setup-root[874]: cut: /sysroot/etc/group: No such file or directory Sep 8 23:51:06.367374 initrd-setup-root[881]: cut: /sysroot/etc/shadow: No such file or directory Sep 8 23:51:06.370958 initrd-setup-root[888]: cut: /sysroot/etc/gshadow: No such file or directory Sep 8 23:51:06.443988 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 8 23:51:06.446299 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 8 23:51:06.448001 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 8 23:51:06.467072 kernel: BTRFS info (device vda6): last unmount of filesystem 0ac87192-1b33-43df-818c-9161f04c3e9c Sep 8 23:51:06.483214 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 8 23:51:06.495684 ignition[957]: INFO : Ignition 2.21.0 Sep 8 23:51:06.495684 ignition[957]: INFO : Stage: mount Sep 8 23:51:06.497886 ignition[957]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:06.497886 ignition[957]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:06.500646 ignition[957]: INFO : mount: mount passed Sep 8 23:51:06.500646 ignition[957]: INFO : Ignition finished successfully Sep 8 23:51:06.503098 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 8 23:51:06.504850 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 8 23:51:06.854245 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 8 23:51:06.855781 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 8 23:51:06.876063 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (970) Sep 8 23:51:06.878061 kernel: BTRFS info (device vda6): first mount of filesystem 0ac87192-1b33-43df-818c-9161f04c3e9c Sep 8 23:51:06.878088 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 8 23:51:06.880448 kernel: BTRFS info (device vda6): turning on async discard Sep 8 23:51:06.880466 kernel: BTRFS info (device vda6): enabling free space tree Sep 8 23:51:06.881851 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 8 23:51:06.908261 ignition[987]: INFO : Ignition 2.21.0 Sep 8 23:51:06.908261 ignition[987]: INFO : Stage: files Sep 8 23:51:06.910637 ignition[987]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:06.910637 ignition[987]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:06.910637 ignition[987]: DEBUG : files: compiled without relabeling support, skipping Sep 8 23:51:06.913158 ignition[987]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 8 23:51:06.913158 ignition[987]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 8 23:51:06.913158 ignition[987]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 8 23:51:06.913158 ignition[987]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 8 23:51:06.913158 ignition[987]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 8 23:51:06.913100 unknown[987]: wrote ssh authorized keys file for user: core Sep 8 23:51:06.919861 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 8 23:51:06.919861 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Sep 8 23:51:06.964597 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 8 23:51:07.168063 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 8 23:51:07.168063 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 8 23:51:07.170942 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Sep 8 23:51:07.238196 systemd-networkd[808]: eth0: Gained IPv6LL Sep 8 23:51:07.380389 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 8 23:51:07.516977 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 8 23:51:07.516977 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 8 23:51:07.521920 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 8 23:51:07.540677 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 8 23:51:07.540677 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 8 23:51:07.540677 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Sep 8 23:51:07.798822 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 8 23:51:08.526291 ignition[987]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 8 23:51:08.526291 ignition[987]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 8 23:51:08.530285 ignition[987]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Sep 8 23:51:08.545394 ignition[987]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 8 23:51:08.548925 ignition[987]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 8 23:51:08.550477 ignition[987]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Sep 8 23:51:08.550477 ignition[987]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Sep 8 23:51:08.550477 ignition[987]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Sep 8 23:51:08.550477 ignition[987]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 8 23:51:08.550477 ignition[987]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 8 23:51:08.550477 ignition[987]: INFO : files: files passed Sep 8 23:51:08.550477 ignition[987]: INFO : Ignition finished successfully Sep 8 23:51:08.551143 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 8 23:51:08.558199 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 8 23:51:08.563206 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 8 23:51:08.573740 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 8 23:51:08.573842 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 8 23:51:08.577443 initrd-setup-root-after-ignition[1016]: grep: /sysroot/oem/oem-release: No such file or directory Sep 8 23:51:08.580528 initrd-setup-root-after-ignition[1022]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 8 23:51:08.581912 initrd-setup-root-after-ignition[1018]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 8 23:51:08.581912 initrd-setup-root-after-ignition[1018]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 8 23:51:08.581795 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 8 23:51:08.583272 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 8 23:51:08.586085 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 8 23:51:08.648094 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 8 23:51:08.648205 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 8 23:51:08.650395 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 8 23:51:08.652105 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 8 23:51:08.653823 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 8 23:51:08.654633 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 8 23:51:08.697421 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 8 23:51:08.700071 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 8 23:51:08.723371 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 8 23:51:08.724736 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 8 23:51:08.726681 systemd[1]: Stopped target timers.target - Timer Units. Sep 8 23:51:08.728334 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 8 23:51:08.728463 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 8 23:51:08.730853 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 8 23:51:08.732714 systemd[1]: Stopped target basic.target - Basic System. Sep 8 23:51:08.734237 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 8 23:51:08.735920 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 8 23:51:08.737853 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 8 23:51:08.739749 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 8 23:51:08.741646 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 8 23:51:08.743319 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 8 23:51:08.745219 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 8 23:51:08.747027 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 8 23:51:08.748707 systemd[1]: Stopped target swap.target - Swaps. Sep 8 23:51:08.750168 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 8 23:51:08.750301 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 8 23:51:08.752450 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 8 23:51:08.754214 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 8 23:51:08.755980 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 8 23:51:08.757579 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 8 23:51:08.758902 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 8 23:51:08.759021 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 8 23:51:08.761589 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 8 23:51:08.761704 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 8 23:51:08.763585 systemd[1]: Stopped target paths.target - Path Units. Sep 8 23:51:08.764970 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 8 23:51:08.768150 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 8 23:51:08.769487 systemd[1]: Stopped target slices.target - Slice Units. Sep 8 23:51:08.771332 systemd[1]: Stopped target sockets.target - Socket Units. Sep 8 23:51:08.772770 systemd[1]: iscsid.socket: Deactivated successfully. Sep 8 23:51:08.772869 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 8 23:51:08.774396 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 8 23:51:08.774474 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 8 23:51:08.775918 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 8 23:51:08.776033 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 8 23:51:08.777808 systemd[1]: ignition-files.service: Deactivated successfully. Sep 8 23:51:08.777916 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 8 23:51:08.780206 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 8 23:51:08.785253 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 8 23:51:08.785390 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 8 23:51:08.791260 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 8 23:51:08.792642 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 8 23:51:08.792757 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 8 23:51:08.795323 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 8 23:51:08.795423 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 8 23:51:08.802852 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 8 23:51:08.810429 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 8 23:51:08.820239 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 8 23:51:08.821447 ignition[1042]: INFO : Ignition 2.21.0 Sep 8 23:51:08.821447 ignition[1042]: INFO : Stage: umount Sep 8 23:51:08.822724 ignition[1042]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 8 23:51:08.822724 ignition[1042]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 8 23:51:08.824659 ignition[1042]: INFO : umount: umount passed Sep 8 23:51:08.824659 ignition[1042]: INFO : Ignition finished successfully Sep 8 23:51:08.825579 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 8 23:51:08.825670 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 8 23:51:08.826826 systemd[1]: Stopped target network.target - Network. Sep 8 23:51:08.827913 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 8 23:51:08.827968 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 8 23:51:08.829922 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 8 23:51:08.829968 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 8 23:51:08.832231 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 8 23:51:08.832305 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 8 23:51:08.833437 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 8 23:51:08.833479 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 8 23:51:08.835770 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 8 23:51:08.839249 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 8 23:51:08.850825 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 8 23:51:08.850934 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 8 23:51:08.853726 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 8 23:51:08.853985 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 8 23:51:08.854022 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 8 23:51:08.860859 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 8 23:51:08.861354 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 8 23:51:08.861453 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 8 23:51:08.864847 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 8 23:51:08.865280 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 8 23:51:08.867805 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 8 23:51:08.867843 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 8 23:51:08.870849 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 8 23:51:08.872474 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 8 23:51:08.872548 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 8 23:51:08.875076 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 8 23:51:08.875128 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 8 23:51:08.877336 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 8 23:51:08.877382 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 8 23:51:08.878933 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 8 23:51:08.881883 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 8 23:51:08.892498 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 8 23:51:08.892634 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 8 23:51:08.894699 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 8 23:51:08.894755 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 8 23:51:08.896000 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 8 23:51:08.896032 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 8 23:51:08.897914 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 8 23:51:08.897963 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 8 23:51:08.900581 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 8 23:51:08.900629 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 8 23:51:08.902956 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 8 23:51:08.903087 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 8 23:51:08.906393 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 8 23:51:08.908141 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 8 23:51:08.908202 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 8 23:51:08.910968 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 8 23:51:08.911016 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 8 23:51:08.913867 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 8 23:51:08.913916 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 8 23:51:08.918283 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 8 23:51:08.919166 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 8 23:51:08.920652 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 8 23:51:08.920729 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 8 23:51:08.922835 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 8 23:51:08.922915 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 8 23:51:08.924959 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 8 23:51:08.925071 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 8 23:51:08.926883 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 8 23:51:08.928807 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 8 23:51:08.947531 systemd[1]: Switching root. Sep 8 23:51:08.985100 systemd-journald[245]: Journal stopped Sep 8 23:51:09.786418 systemd-journald[245]: Received SIGTERM from PID 1 (systemd). Sep 8 23:51:09.786588 kernel: SELinux: policy capability network_peer_controls=1 Sep 8 23:51:09.786609 kernel: SELinux: policy capability open_perms=1 Sep 8 23:51:09.786618 kernel: SELinux: policy capability extended_socket_class=1 Sep 8 23:51:09.786627 kernel: SELinux: policy capability always_check_network=0 Sep 8 23:51:09.786638 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 8 23:51:09.786647 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 8 23:51:09.786656 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 8 23:51:09.786727 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 8 23:51:09.786747 kernel: SELinux: policy capability userspace_initial_context=0 Sep 8 23:51:09.786757 kernel: audit: type=1403 audit(1757375469.160:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 8 23:51:09.786772 systemd[1]: Successfully loaded SELinux policy in 45.892ms. Sep 8 23:51:09.786788 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.608ms. Sep 8 23:51:09.786802 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 8 23:51:09.786816 systemd[1]: Detected virtualization kvm. Sep 8 23:51:09.786826 systemd[1]: Detected architecture arm64. Sep 8 23:51:09.786836 systemd[1]: Detected first boot. Sep 8 23:51:09.786846 systemd[1]: Initializing machine ID from VM UUID. Sep 8 23:51:09.786855 zram_generator::config[1088]: No configuration found. Sep 8 23:51:09.786866 kernel: NET: Registered PF_VSOCK protocol family Sep 8 23:51:09.786875 systemd[1]: Populated /etc with preset unit settings. Sep 8 23:51:09.786885 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 8 23:51:09.786896 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 8 23:51:09.786906 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 8 23:51:09.786916 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 8 23:51:09.786926 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 8 23:51:09.786936 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 8 23:51:09.786946 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 8 23:51:09.786956 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 8 23:51:09.786966 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 8 23:51:09.786980 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 8 23:51:09.786989 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 8 23:51:09.787003 systemd[1]: Created slice user.slice - User and Session Slice. Sep 8 23:51:09.787013 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 8 23:51:09.787031 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 8 23:51:09.787113 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 8 23:51:09.787126 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 8 23:51:09.787141 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 8 23:51:09.787152 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 8 23:51:09.787164 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Sep 8 23:51:09.787174 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 8 23:51:09.787184 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 8 23:51:09.787194 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 8 23:51:09.787204 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 8 23:51:09.787220 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 8 23:51:09.787231 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 8 23:51:09.787240 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 8 23:51:09.787252 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 8 23:51:09.787261 systemd[1]: Reached target slices.target - Slice Units. Sep 8 23:51:09.787271 systemd[1]: Reached target swap.target - Swaps. Sep 8 23:51:09.787281 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 8 23:51:09.787290 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 8 23:51:09.787300 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 8 23:51:09.787324 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 8 23:51:09.787334 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 8 23:51:09.787344 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 8 23:51:09.787357 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 8 23:51:09.787366 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 8 23:51:09.787376 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 8 23:51:09.787390 systemd[1]: Mounting media.mount - External Media Directory... Sep 8 23:51:09.787400 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 8 23:51:09.787410 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 8 23:51:09.787420 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 8 23:51:09.787430 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 8 23:51:09.787440 systemd[1]: Reached target machines.target - Containers. Sep 8 23:51:09.787451 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 8 23:51:09.787461 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 8 23:51:09.787470 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 8 23:51:09.787480 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 8 23:51:09.787490 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 8 23:51:09.787500 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 8 23:51:09.787509 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 8 23:51:09.787519 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 8 23:51:09.787531 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 8 23:51:09.787541 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 8 23:51:09.787550 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 8 23:51:09.787560 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 8 23:51:09.787571 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 8 23:51:09.787581 systemd[1]: Stopped systemd-fsck-usr.service. Sep 8 23:51:09.787592 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 8 23:51:09.787601 kernel: fuse: init (API version 7.41) Sep 8 23:51:09.787612 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 8 23:51:09.787622 kernel: loop: module loaded Sep 8 23:51:09.787631 kernel: ACPI: bus type drm_connector registered Sep 8 23:51:09.787641 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 8 23:51:09.787651 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 8 23:51:09.787661 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 8 23:51:09.787672 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 8 23:51:09.787683 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 8 23:51:09.787693 systemd[1]: verity-setup.service: Deactivated successfully. Sep 8 23:51:09.787703 systemd[1]: Stopped verity-setup.service. Sep 8 23:51:09.787745 systemd-journald[1163]: Collecting audit messages is disabled. Sep 8 23:51:09.787768 systemd-journald[1163]: Journal started Sep 8 23:51:09.787790 systemd-journald[1163]: Runtime Journal (/run/log/journal/a38e81aaad9b4c3eb32f5e5652895cfd) is 6M, max 48.5M, 42.4M free. Sep 8 23:51:09.571115 systemd[1]: Queued start job for default target multi-user.target. Sep 8 23:51:09.596125 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 8 23:51:09.596530 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 8 23:51:09.790349 systemd[1]: Started systemd-journald.service - Journal Service. Sep 8 23:51:09.791037 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 8 23:51:09.792395 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 8 23:51:09.794534 systemd[1]: Mounted media.mount - External Media Directory. Sep 8 23:51:09.795764 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 8 23:51:09.797109 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 8 23:51:09.798373 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 8 23:51:09.799665 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 8 23:51:09.801279 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 8 23:51:09.802825 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 8 23:51:09.802986 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 8 23:51:09.804495 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 8 23:51:09.804672 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 8 23:51:09.806398 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 8 23:51:09.806680 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 8 23:51:09.808150 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 8 23:51:09.808318 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 8 23:51:09.809886 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 8 23:51:09.810069 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 8 23:51:09.811378 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 8 23:51:09.811529 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 8 23:51:09.812940 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 8 23:51:09.814629 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 8 23:51:09.816331 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 8 23:51:09.817987 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 8 23:51:09.830037 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 8 23:51:09.832400 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 8 23:51:09.834260 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 8 23:51:09.835174 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 8 23:51:09.835208 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 8 23:51:09.836928 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 8 23:51:09.845935 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 8 23:51:09.847071 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 8 23:51:09.848514 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 8 23:51:09.850224 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 8 23:51:09.851277 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 8 23:51:09.854196 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 8 23:51:09.855270 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 8 23:51:09.856515 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 8 23:51:09.858762 systemd-journald[1163]: Time spent on flushing to /var/log/journal/a38e81aaad9b4c3eb32f5e5652895cfd is 15.055ms for 889 entries. Sep 8 23:51:09.858762 systemd-journald[1163]: System Journal (/var/log/journal/a38e81aaad9b4c3eb32f5e5652895cfd) is 8M, max 195.6M, 187.6M free. Sep 8 23:51:09.888418 systemd-journald[1163]: Received client request to flush runtime journal. Sep 8 23:51:09.888469 kernel: loop0: detected capacity change from 0 to 119320 Sep 8 23:51:09.858627 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 8 23:51:09.864181 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 8 23:51:09.866794 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 8 23:51:09.868171 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 8 23:51:09.869382 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 8 23:51:09.870641 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 8 23:51:09.874573 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 8 23:51:09.879291 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 8 23:51:09.891376 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 8 23:51:09.896241 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 8 23:51:09.897063 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 8 23:51:09.910448 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 8 23:51:09.917146 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 8 23:51:09.919908 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 8 23:51:09.922107 kernel: loop1: detected capacity change from 0 to 203944 Sep 8 23:51:09.941822 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Sep 8 23:51:09.941841 systemd-tmpfiles[1222]: ACLs are not supported, ignoring. Sep 8 23:51:09.947098 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 8 23:51:09.954061 kernel: loop2: detected capacity change from 0 to 100608 Sep 8 23:51:09.982066 kernel: loop3: detected capacity change from 0 to 119320 Sep 8 23:51:09.988066 kernel: loop4: detected capacity change from 0 to 203944 Sep 8 23:51:09.995066 kernel: loop5: detected capacity change from 0 to 100608 Sep 8 23:51:09.999679 (sd-merge)[1228]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 8 23:51:10.000113 (sd-merge)[1228]: Merged extensions into '/usr'. Sep 8 23:51:10.003591 systemd[1]: Reload requested from client PID 1204 ('systemd-sysext') (unit systemd-sysext.service)... Sep 8 23:51:10.003612 systemd[1]: Reloading... Sep 8 23:51:10.056066 zram_generator::config[1253]: No configuration found. Sep 8 23:51:10.116270 ldconfig[1199]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 8 23:51:10.220696 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 8 23:51:10.221146 systemd[1]: Reloading finished in 217 ms. Sep 8 23:51:10.237353 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 8 23:51:10.238554 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 8 23:51:10.251192 systemd[1]: Starting ensure-sysext.service... Sep 8 23:51:10.252838 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 8 23:51:10.261554 systemd[1]: Reload requested from client PID 1288 ('systemctl') (unit ensure-sysext.service)... Sep 8 23:51:10.261569 systemd[1]: Reloading... Sep 8 23:51:10.268308 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 8 23:51:10.268604 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 8 23:51:10.268906 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 8 23:51:10.269213 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 8 23:51:10.269910 systemd-tmpfiles[1289]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 8 23:51:10.270297 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Sep 8 23:51:10.270426 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Sep 8 23:51:10.272702 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Sep 8 23:51:10.272804 systemd-tmpfiles[1289]: Skipping /boot Sep 8 23:51:10.278577 systemd-tmpfiles[1289]: Detected autofs mount point /boot during canonicalization of boot. Sep 8 23:51:10.278676 systemd-tmpfiles[1289]: Skipping /boot Sep 8 23:51:10.296063 zram_generator::config[1316]: No configuration found. Sep 8 23:51:10.440434 systemd[1]: Reloading finished in 178 ms. Sep 8 23:51:10.463683 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 8 23:51:10.489419 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 8 23:51:10.496819 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 8 23:51:10.499126 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 8 23:51:10.501247 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 8 23:51:10.510285 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 8 23:51:10.512566 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 8 23:51:10.517357 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 8 23:51:10.520820 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 8 23:51:10.522555 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 8 23:51:10.526458 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 8 23:51:10.530097 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 8 23:51:10.531186 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 8 23:51:10.531292 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 8 23:51:10.532246 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 8 23:51:10.536570 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 8 23:51:10.536894 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 8 23:51:10.539633 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 8 23:51:10.539821 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 8 23:51:10.542269 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 8 23:51:10.542436 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 8 23:51:10.553751 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 8 23:51:10.557541 systemd-udevd[1357]: Using default interface naming scheme 'v255'. Sep 8 23:51:10.558497 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 8 23:51:10.559713 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 8 23:51:10.563322 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 8 23:51:10.566325 augenrules[1387]: No rules Sep 8 23:51:10.569356 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 8 23:51:10.571248 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 8 23:51:10.571381 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 8 23:51:10.573362 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 8 23:51:10.578308 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 8 23:51:10.579135 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 8 23:51:10.580425 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 8 23:51:10.581920 systemd[1]: audit-rules.service: Deactivated successfully. Sep 8 23:51:10.582111 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 8 23:51:10.584573 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 8 23:51:10.587368 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 8 23:51:10.587545 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 8 23:51:10.589213 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 8 23:51:10.590106 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 8 23:51:10.592481 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 8 23:51:10.592621 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 8 23:51:10.595108 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 8 23:51:10.626388 systemd[1]: Finished ensure-sysext.service. Sep 8 23:51:10.637572 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 8 23:51:10.643483 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 8 23:51:10.645692 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 8 23:51:10.648893 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 8 23:51:10.652270 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 8 23:51:10.659391 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 8 23:51:10.661254 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 8 23:51:10.661309 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 8 23:51:10.664273 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 8 23:51:10.668251 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 8 23:51:10.669113 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 8 23:51:10.669562 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 8 23:51:10.669754 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 8 23:51:10.681171 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 8 23:51:10.681378 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 8 23:51:10.683243 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 8 23:51:10.683420 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 8 23:51:10.689259 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 8 23:51:10.691612 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 8 23:51:10.698141 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Sep 8 23:51:10.699280 augenrules[1434]: /sbin/augenrules: No change Sep 8 23:51:10.702815 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 8 23:51:10.712089 augenrules[1463]: No rules Sep 8 23:51:10.714401 systemd[1]: audit-rules.service: Deactivated successfully. Sep 8 23:51:10.714662 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 8 23:51:10.719580 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 8 23:51:10.722200 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 8 23:51:10.722265 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 8 23:51:10.738507 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 8 23:51:10.740511 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 8 23:51:10.789985 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 8 23:51:10.853583 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 8 23:51:10.871651 systemd-networkd[1441]: lo: Link UP Sep 8 23:51:10.871659 systemd-networkd[1441]: lo: Gained carrier Sep 8 23:51:10.872401 systemd-networkd[1441]: Enumeration completed Sep 8 23:51:10.872538 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 8 23:51:10.872805 systemd-networkd[1441]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 8 23:51:10.872808 systemd-networkd[1441]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 8 23:51:10.873321 systemd-networkd[1441]: eth0: Link UP Sep 8 23:51:10.873435 systemd-networkd[1441]: eth0: Gained carrier Sep 8 23:51:10.873449 systemd-networkd[1441]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 8 23:51:10.875159 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 8 23:51:10.877439 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 8 23:51:10.880315 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 8 23:51:10.881622 systemd[1]: Reached target time-set.target - System Time Set. Sep 8 23:51:10.889101 systemd-resolved[1355]: Positive Trust Anchors: Sep 8 23:51:10.889103 systemd-networkd[1441]: eth0: DHCPv4 address 10.0.0.130/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 8 23:51:10.889119 systemd-resolved[1355]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 8 23:51:10.889150 systemd-resolved[1355]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 8 23:51:10.890079 systemd-timesyncd[1442]: Network configuration changed, trying to establish connection. Sep 8 23:51:10.891060 systemd-timesyncd[1442]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 8 23:51:10.891108 systemd-timesyncd[1442]: Initial clock synchronization to Mon 2025-09-08 23:51:11.177066 UTC. Sep 8 23:51:10.895120 systemd-resolved[1355]: Defaulting to hostname 'linux'. Sep 8 23:51:10.896473 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 8 23:51:10.897432 systemd[1]: Reached target network.target - Network. Sep 8 23:51:10.898126 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 8 23:51:10.899221 systemd[1]: Reached target sysinit.target - System Initialization. Sep 8 23:51:10.900073 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 8 23:51:10.900962 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 8 23:51:10.902175 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 8 23:51:10.903064 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 8 23:51:10.903965 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 8 23:51:10.904972 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 8 23:51:10.905004 systemd[1]: Reached target paths.target - Path Units. Sep 8 23:51:10.905757 systemd[1]: Reached target timers.target - Timer Units. Sep 8 23:51:10.907082 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 8 23:51:10.909007 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 8 23:51:10.911523 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 8 23:51:10.912680 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 8 23:51:10.913686 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 8 23:51:10.917822 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 8 23:51:10.918935 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 8 23:51:10.920592 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 8 23:51:10.921761 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 8 23:51:10.923761 systemd[1]: Reached target sockets.target - Socket Units. Sep 8 23:51:10.924590 systemd[1]: Reached target basic.target - Basic System. Sep 8 23:51:10.925356 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 8 23:51:10.925386 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 8 23:51:10.926306 systemd[1]: Starting containerd.service - containerd container runtime... Sep 8 23:51:10.927964 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 8 23:51:10.929685 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 8 23:51:10.931451 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 8 23:51:10.933101 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 8 23:51:10.933960 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 8 23:51:10.934829 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 8 23:51:10.938522 jq[1509]: false Sep 8 23:51:10.939108 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 8 23:51:10.940930 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 8 23:51:10.943064 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 8 23:51:10.945119 extend-filesystems[1510]: Found /dev/vda6 Sep 8 23:51:10.948146 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 8 23:51:10.949756 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 8 23:51:10.950173 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 8 23:51:10.950894 systemd[1]: Starting update-engine.service - Update Engine... Sep 8 23:51:10.951787 extend-filesystems[1510]: Found /dev/vda9 Sep 8 23:51:10.952789 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 8 23:51:10.955585 extend-filesystems[1510]: Checking size of /dev/vda9 Sep 8 23:51:10.962702 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 8 23:51:10.966606 jq[1529]: true Sep 8 23:51:10.966527 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 8 23:51:10.966692 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 8 23:51:10.966971 systemd[1]: motdgen.service: Deactivated successfully. Sep 8 23:51:10.967190 extend-filesystems[1510]: Resized partition /dev/vda9 Sep 8 23:51:10.967145 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 8 23:51:10.970290 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 8 23:51:10.971472 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 8 23:51:10.974814 extend-filesystems[1536]: resize2fs 1.47.2 (1-Jan-2025) Sep 8 23:51:10.986329 update_engine[1527]: I20250908 23:51:10.986129 1527 main.cc:92] Flatcar Update Engine starting Sep 8 23:51:10.994375 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 8 23:51:10.998859 tar[1537]: linux-arm64/helm Sep 8 23:51:11.007310 jq[1538]: true Sep 8 23:51:11.013729 (ntainerd)[1549]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 8 23:51:11.023515 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 8 23:51:11.024902 dbus-daemon[1507]: [system] SELinux support is enabled Sep 8 23:51:11.025155 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 8 23:51:11.046766 update_engine[1527]: I20250908 23:51:11.032248 1527 update_check_scheduler.cc:74] Next update check in 11m26s Sep 8 23:51:11.030328 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 8 23:51:11.030362 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 8 23:51:11.033261 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 8 23:51:11.033298 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 8 23:51:11.037623 systemd[1]: Started update-engine.service - Update Engine. Sep 8 23:51:11.041198 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 8 23:51:11.046837 systemd-logind[1524]: Watching system buttons on /dev/input/event0 (Power Button) Sep 8 23:51:11.047336 systemd-logind[1524]: New seat seat0. Sep 8 23:51:11.047435 extend-filesystems[1536]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 8 23:51:11.047435 extend-filesystems[1536]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 8 23:51:11.047435 extend-filesystems[1536]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 8 23:51:11.055986 extend-filesystems[1510]: Resized filesystem in /dev/vda9 Sep 8 23:51:11.048375 systemd[1]: Started systemd-logind.service - User Login Management. Sep 8 23:51:11.050714 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 8 23:51:11.052117 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 8 23:51:11.080883 bash[1570]: Updated "/home/core/.ssh/authorized_keys" Sep 8 23:51:11.083129 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 8 23:51:11.086727 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 8 23:51:11.103337 locksmithd[1554]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 8 23:51:11.182920 containerd[1549]: time="2025-09-08T23:51:11Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 8 23:51:11.183840 containerd[1549]: time="2025-09-08T23:51:11.183794714Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 8 23:51:11.193325 containerd[1549]: time="2025-09-08T23:51:11.193288785Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.41µs" Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193393894Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193417467Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193553981Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193571962Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193593464Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193637753Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193649105Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193836536Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193849752Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193865620Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193874238Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194104 containerd[1549]: time="2025-09-08T23:51:11.193939035Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194465 containerd[1549]: time="2025-09-08T23:51:11.194415319Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194548 containerd[1549]: time="2025-09-08T23:51:11.194530371Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 8 23:51:11.194601 containerd[1549]: time="2025-09-08T23:51:11.194585930Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 8 23:51:11.194682 containerd[1549]: time="2025-09-08T23:51:11.194666139Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 8 23:51:11.195531 containerd[1549]: time="2025-09-08T23:51:11.195490314Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 8 23:51:11.195839 containerd[1549]: time="2025-09-08T23:51:11.195712050Z" level=info msg="metadata content store policy set" policy=shared Sep 8 23:51:11.199817 containerd[1549]: time="2025-09-08T23:51:11.199778901Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 8 23:51:11.199898 containerd[1549]: time="2025-09-08T23:51:11.199878707Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 8 23:51:11.199924 containerd[1549]: time="2025-09-08T23:51:11.199903980Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 8 23:51:11.199924 containerd[1549]: time="2025-09-08T23:51:11.199918439Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199930909Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199941433Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199953116Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199964924Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199975613Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 8 23:51:11.199984 containerd[1549]: time="2025-09-08T23:51:11.199985018Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 8 23:51:11.200152 containerd[1549]: time="2025-09-08T23:51:11.199994381Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 8 23:51:11.200152 containerd[1549]: time="2025-09-08T23:51:11.200048448Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 8 23:51:11.200299 containerd[1549]: time="2025-09-08T23:51:11.200262104Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 8 23:51:11.200299 containerd[1549]: time="2025-09-08T23:51:11.200296326Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 8 23:51:11.200343 containerd[1549]: time="2025-09-08T23:51:11.200313064Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 8 23:51:11.200343 containerd[1549]: time="2025-09-08T23:51:11.200323546Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 8 23:51:11.200395 containerd[1549]: time="2025-09-08T23:51:11.200377447Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 8 23:51:11.200416 containerd[1549]: time="2025-09-08T23:51:11.200401145Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 8 23:51:11.200435 containerd[1549]: time="2025-09-08T23:51:11.200416847Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 8 23:51:11.200457 containerd[1549]: time="2025-09-08T23:51:11.200433005Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 8 23:51:11.200457 containerd[1549]: time="2025-09-08T23:51:11.200446842Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 8 23:51:11.200493 containerd[1549]: time="2025-09-08T23:51:11.200457822Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 8 23:51:11.200493 containerd[1549]: time="2025-09-08T23:51:11.200468635Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 8 23:51:11.200766 containerd[1549]: time="2025-09-08T23:51:11.200736690Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 8 23:51:11.200766 containerd[1549]: time="2025-09-08T23:51:11.200763992Z" level=info msg="Start snapshots syncer" Sep 8 23:51:11.200848 containerd[1549]: time="2025-09-08T23:51:11.200831400Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 8 23:51:11.201428 containerd[1549]: time="2025-09-08T23:51:11.201374842Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 8 23:51:11.201526 containerd[1549]: time="2025-09-08T23:51:11.201442540Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 8 23:51:11.201654 containerd[1549]: time="2025-09-08T23:51:11.201587836Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 8 23:51:11.201868 containerd[1549]: time="2025-09-08T23:51:11.201825771Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 8 23:51:11.201960 containerd[1549]: time="2025-09-08T23:51:11.201941155Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 8 23:51:11.201983 containerd[1549]: time="2025-09-08T23:51:11.201964024Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 8 23:51:11.201983 containerd[1549]: time="2025-09-08T23:51:11.201977448Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 8 23:51:11.202049 containerd[1549]: time="2025-09-08T23:51:11.202031970Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 8 23:51:11.202073 containerd[1549]: time="2025-09-08T23:51:11.202064079Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 8 23:51:11.202091 containerd[1549]: time="2025-09-08T23:51:11.202078290Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 8 23:51:11.202109 containerd[1549]: time="2025-09-08T23:51:11.202102485Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 8 23:51:11.202175 containerd[1549]: time="2025-09-08T23:51:11.202156593Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 8 23:51:11.202200 containerd[1549]: time="2025-09-08T23:51:11.202180788Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 8 23:51:11.202236 containerd[1549]: time="2025-09-08T23:51:11.202222136Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 8 23:51:11.202260 containerd[1549]: time="2025-09-08T23:51:11.202239040Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 8 23:51:11.202260 containerd[1549]: time="2025-09-08T23:51:11.202249397Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 8 23:51:11.202304 containerd[1549]: time="2025-09-08T23:51:11.202258636Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 8 23:51:11.202383 containerd[1549]: time="2025-09-08T23:51:11.202266757Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 8 23:51:11.202383 containerd[1549]: time="2025-09-08T23:51:11.202341663Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 8 23:51:11.202383 containerd[1549]: time="2025-09-08T23:51:11.202354962Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 8 23:51:11.202487 containerd[1549]: time="2025-09-08T23:51:11.202432520Z" level=info msg="runtime interface created" Sep 8 23:51:11.202487 containerd[1549]: time="2025-09-08T23:51:11.202483272Z" level=info msg="created NRI interface" Sep 8 23:51:11.202531 containerd[1549]: time="2025-09-08T23:51:11.202496985Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 8 23:51:11.202531 containerd[1549]: time="2025-09-08T23:51:11.202510202Z" level=info msg="Connect containerd service" Sep 8 23:51:11.202565 containerd[1549]: time="2025-09-08T23:51:11.202546785Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 8 23:51:11.203668 containerd[1549]: time="2025-09-08T23:51:11.203586688Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 8 23:51:11.238518 sshd_keygen[1530]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 8 23:51:11.261122 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 8 23:51:11.264408 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 8 23:51:11.276964 containerd[1549]: time="2025-09-08T23:51:11.276791133Z" level=info msg="Start subscribing containerd event" Sep 8 23:51:11.276964 containerd[1549]: time="2025-09-08T23:51:11.276871218Z" level=info msg="Start recovering state" Sep 8 23:51:11.276964 containerd[1549]: time="2025-09-08T23:51:11.276817234Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 8 23:51:11.277346 containerd[1549]: time="2025-09-08T23:51:11.277325171Z" level=info msg="Start event monitor" Sep 8 23:51:11.277433 containerd[1549]: time="2025-09-08T23:51:11.277414660Z" level=info msg="Start cni network conf syncer for default" Sep 8 23:51:11.277570 containerd[1549]: time="2025-09-08T23:51:11.277494953Z" level=info msg="Start streaming server" Sep 8 23:51:11.277570 containerd[1549]: time="2025-09-08T23:51:11.277520764Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 8 23:51:11.277570 containerd[1549]: time="2025-09-08T23:51:11.277528635Z" level=info msg="runtime interface starting up..." Sep 8 23:51:11.277570 containerd[1549]: time="2025-09-08T23:51:11.277533980Z" level=info msg="starting plugins..." Sep 8 23:51:11.277570 containerd[1549]: time="2025-09-08T23:51:11.277552002Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 8 23:51:11.277676 containerd[1549]: time="2025-09-08T23:51:11.277341619Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 8 23:51:11.278017 containerd[1549]: time="2025-09-08T23:51:11.277995763Z" level=info msg="containerd successfully booted in 0.095422s" Sep 8 23:51:11.278113 systemd[1]: Started containerd.service - containerd container runtime. Sep 8 23:51:11.279892 systemd[1]: issuegen.service: Deactivated successfully. Sep 8 23:51:11.282164 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 8 23:51:11.286632 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 8 23:51:11.292265 tar[1537]: linux-arm64/LICENSE Sep 8 23:51:11.292336 tar[1537]: linux-arm64/README.md Sep 8 23:51:11.318523 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 8 23:51:11.322660 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 8 23:51:11.326540 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 8 23:51:11.329314 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Sep 8 23:51:11.330842 systemd[1]: Reached target getty.target - Login Prompts. Sep 8 23:51:12.550564 systemd-networkd[1441]: eth0: Gained IPv6LL Sep 8 23:51:12.552848 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 8 23:51:12.556712 systemd[1]: Reached target network-online.target - Network is Online. Sep 8 23:51:12.558871 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 8 23:51:12.561165 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:12.563030 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 8 23:51:12.579603 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 8 23:51:12.579786 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 8 23:51:12.582380 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 8 23:51:12.584664 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 8 23:51:13.114032 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:13.115630 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 8 23:51:13.117617 (kubelet)[1644]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 8 23:51:13.119139 systemd[1]: Startup finished in 2.041s (kernel) + 5.560s (initrd) + 4.005s (userspace) = 11.608s. Sep 8 23:51:13.500082 kubelet[1644]: E0908 23:51:13.499966 1644 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 8 23:51:13.502516 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 8 23:51:13.502670 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 8 23:51:13.503011 systemd[1]: kubelet.service: Consumed 764ms CPU time, 255.1M memory peak. Sep 8 23:51:17.174851 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 8 23:51:17.177179 systemd[1]: Started sshd@0-10.0.0.130:22-10.0.0.1:42788.service - OpenSSH per-connection server daemon (10.0.0.1:42788). Sep 8 23:51:17.265753 sshd[1657]: Accepted publickey for core from 10.0.0.1 port 42788 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:17.267043 sshd-session[1657]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:17.274688 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 8 23:51:17.277620 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 8 23:51:17.289731 systemd-logind[1524]: New session 1 of user core. Sep 8 23:51:17.311183 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 8 23:51:17.313927 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 8 23:51:17.336382 (systemd)[1662]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 8 23:51:17.340426 systemd-logind[1524]: New session c1 of user core. Sep 8 23:51:17.463906 systemd[1662]: Queued start job for default target default.target. Sep 8 23:51:17.479177 systemd[1662]: Created slice app.slice - User Application Slice. Sep 8 23:51:17.479209 systemd[1662]: Reached target paths.target - Paths. Sep 8 23:51:17.479248 systemd[1662]: Reached target timers.target - Timers. Sep 8 23:51:17.480556 systemd[1662]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 8 23:51:17.491098 systemd[1662]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 8 23:51:17.491267 systemd[1662]: Reached target sockets.target - Sockets. Sep 8 23:51:17.491316 systemd[1662]: Reached target basic.target - Basic System. Sep 8 23:51:17.491347 systemd[1662]: Reached target default.target - Main User Target. Sep 8 23:51:17.491375 systemd[1662]: Startup finished in 144ms. Sep 8 23:51:17.491525 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 8 23:51:17.492925 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 8 23:51:17.557434 systemd[1]: Started sshd@1-10.0.0.130:22-10.0.0.1:42802.service - OpenSSH per-connection server daemon (10.0.0.1:42802). Sep 8 23:51:17.605962 sshd[1673]: Accepted publickey for core from 10.0.0.1 port 42802 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:17.607162 sshd-session[1673]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:17.612956 systemd-logind[1524]: New session 2 of user core. Sep 8 23:51:17.623210 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 8 23:51:17.676022 sshd[1676]: Connection closed by 10.0.0.1 port 42802 Sep 8 23:51:17.676460 sshd-session[1673]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:17.696077 systemd[1]: sshd@1-10.0.0.130:22-10.0.0.1:42802.service: Deactivated successfully. Sep 8 23:51:17.697524 systemd[1]: session-2.scope: Deactivated successfully. Sep 8 23:51:17.699256 systemd-logind[1524]: Session 2 logged out. Waiting for processes to exit. Sep 8 23:51:17.701709 systemd[1]: Started sshd@2-10.0.0.130:22-10.0.0.1:42818.service - OpenSSH per-connection server daemon (10.0.0.1:42818). Sep 8 23:51:17.702194 systemd-logind[1524]: Removed session 2. Sep 8 23:51:17.757305 sshd[1682]: Accepted publickey for core from 10.0.0.1 port 42818 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:17.758556 sshd-session[1682]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:17.763152 systemd-logind[1524]: New session 3 of user core. Sep 8 23:51:17.781245 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 8 23:51:17.828853 sshd[1685]: Connection closed by 10.0.0.1 port 42818 Sep 8 23:51:17.828720 sshd-session[1682]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:17.849707 systemd[1]: sshd@2-10.0.0.130:22-10.0.0.1:42818.service: Deactivated successfully. Sep 8 23:51:17.852991 systemd[1]: session-3.scope: Deactivated successfully. Sep 8 23:51:17.853813 systemd-logind[1524]: Session 3 logged out. Waiting for processes to exit. Sep 8 23:51:17.856229 systemd[1]: Started sshd@3-10.0.0.130:22-10.0.0.1:42822.service - OpenSSH per-connection server daemon (10.0.0.1:42822). Sep 8 23:51:17.856669 systemd-logind[1524]: Removed session 3. Sep 8 23:51:17.915627 sshd[1691]: Accepted publickey for core from 10.0.0.1 port 42822 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:17.916965 sshd-session[1691]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:17.920791 systemd-logind[1524]: New session 4 of user core. Sep 8 23:51:17.929231 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 8 23:51:17.981550 sshd[1694]: Connection closed by 10.0.0.1 port 42822 Sep 8 23:51:17.981941 sshd-session[1691]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:17.994287 systemd[1]: sshd@3-10.0.0.130:22-10.0.0.1:42822.service: Deactivated successfully. Sep 8 23:51:17.996527 systemd[1]: session-4.scope: Deactivated successfully. Sep 8 23:51:17.997705 systemd-logind[1524]: Session 4 logged out. Waiting for processes to exit. Sep 8 23:51:17.999425 systemd[1]: Started sshd@4-10.0.0.130:22-10.0.0.1:42830.service - OpenSSH per-connection server daemon (10.0.0.1:42830). Sep 8 23:51:18.000296 systemd-logind[1524]: Removed session 4. Sep 8 23:51:18.055416 sshd[1700]: Accepted publickey for core from 10.0.0.1 port 42830 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:18.056753 sshd-session[1700]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:18.061914 systemd-logind[1524]: New session 5 of user core. Sep 8 23:51:18.073216 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 8 23:51:18.132023 sudo[1704]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 8 23:51:18.132317 sudo[1704]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 8 23:51:18.148994 sudo[1704]: pam_unix(sudo:session): session closed for user root Sep 8 23:51:18.150431 sshd[1703]: Connection closed by 10.0.0.1 port 42830 Sep 8 23:51:18.150934 sshd-session[1700]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:18.160206 systemd[1]: sshd@4-10.0.0.130:22-10.0.0.1:42830.service: Deactivated successfully. Sep 8 23:51:18.161638 systemd[1]: session-5.scope: Deactivated successfully. Sep 8 23:51:18.163658 systemd-logind[1524]: Session 5 logged out. Waiting for processes to exit. Sep 8 23:51:18.165955 systemd[1]: Started sshd@5-10.0.0.130:22-10.0.0.1:42844.service - OpenSSH per-connection server daemon (10.0.0.1:42844). Sep 8 23:51:18.166585 systemd-logind[1524]: Removed session 5. Sep 8 23:51:18.220754 sshd[1710]: Accepted publickey for core from 10.0.0.1 port 42844 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:18.222248 sshd-session[1710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:18.225980 systemd-logind[1524]: New session 6 of user core. Sep 8 23:51:18.237216 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 8 23:51:18.288215 sudo[1715]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 8 23:51:18.288778 sudo[1715]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 8 23:51:18.402916 sudo[1715]: pam_unix(sudo:session): session closed for user root Sep 8 23:51:18.408188 sudo[1714]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 8 23:51:18.408459 sudo[1714]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 8 23:51:18.416708 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 8 23:51:18.455450 augenrules[1737]: No rules Sep 8 23:51:18.456586 systemd[1]: audit-rules.service: Deactivated successfully. Sep 8 23:51:18.456833 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 8 23:51:18.457678 sudo[1714]: pam_unix(sudo:session): session closed for user root Sep 8 23:51:18.458873 sshd[1713]: Connection closed by 10.0.0.1 port 42844 Sep 8 23:51:18.459275 sshd-session[1710]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:18.470991 systemd[1]: sshd@5-10.0.0.130:22-10.0.0.1:42844.service: Deactivated successfully. Sep 8 23:51:18.472812 systemd[1]: session-6.scope: Deactivated successfully. Sep 8 23:51:18.473747 systemd-logind[1524]: Session 6 logged out. Waiting for processes to exit. Sep 8 23:51:18.476784 systemd[1]: Started sshd@6-10.0.0.130:22-10.0.0.1:42854.service - OpenSSH per-connection server daemon (10.0.0.1:42854). Sep 8 23:51:18.477434 systemd-logind[1524]: Removed session 6. Sep 8 23:51:18.539816 sshd[1746]: Accepted publickey for core from 10.0.0.1 port 42854 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:51:18.541125 sshd-session[1746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:51:18.546000 systemd-logind[1524]: New session 7 of user core. Sep 8 23:51:18.555953 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 8 23:51:18.607670 sudo[1750]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 8 23:51:18.608309 sudo[1750]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 8 23:51:18.891917 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 8 23:51:18.913433 (dockerd)[1771]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 8 23:51:19.121994 dockerd[1771]: time="2025-09-08T23:51:19.121921053Z" level=info msg="Starting up" Sep 8 23:51:19.123293 dockerd[1771]: time="2025-09-08T23:51:19.123258973Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 8 23:51:19.137310 dockerd[1771]: time="2025-09-08T23:51:19.137255226Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 8 23:51:19.155243 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2712996655-merged.mount: Deactivated successfully. Sep 8 23:51:19.180514 dockerd[1771]: time="2025-09-08T23:51:19.180446178Z" level=info msg="Loading containers: start." Sep 8 23:51:19.189091 kernel: Initializing XFRM netlink socket Sep 8 23:51:19.384731 systemd-networkd[1441]: docker0: Link UP Sep 8 23:51:19.389312 dockerd[1771]: time="2025-09-08T23:51:19.389259938Z" level=info msg="Loading containers: done." Sep 8 23:51:19.402106 dockerd[1771]: time="2025-09-08T23:51:19.402064688Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 8 23:51:19.402253 dockerd[1771]: time="2025-09-08T23:51:19.402151623Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 8 23:51:19.402253 dockerd[1771]: time="2025-09-08T23:51:19.402234793Z" level=info msg="Initializing buildkit" Sep 8 23:51:19.424246 dockerd[1771]: time="2025-09-08T23:51:19.424153816Z" level=info msg="Completed buildkit initialization" Sep 8 23:51:19.430991 dockerd[1771]: time="2025-09-08T23:51:19.430952219Z" level=info msg="Daemon has completed initialization" Sep 8 23:51:19.431219 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 8 23:51:19.431641 dockerd[1771]: time="2025-09-08T23:51:19.431516711Z" level=info msg="API listen on /run/docker.sock" Sep 8 23:51:19.945864 containerd[1549]: time="2025-09-08T23:51:19.945702824Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.12\"" Sep 8 23:51:20.542039 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1647049738.mount: Deactivated successfully. Sep 8 23:51:21.630016 containerd[1549]: time="2025-09-08T23:51:21.629951988Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:21.630656 containerd[1549]: time="2025-09-08T23:51:21.630604348Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.12: active requests=0, bytes read=25652443" Sep 8 23:51:21.631457 containerd[1549]: time="2025-09-08T23:51:21.631418859Z" level=info msg="ImageCreate event name:\"sha256:25d00c9505e8a4a7a6c827030f878b50e58bbf63322e01a7d92807bcb4db6b3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:21.635094 containerd[1549]: time="2025-09-08T23:51:21.635012916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:e9011c3bee8c06ecabd7816e119dca4e448c92f7a78acd891de3d2db1dc6c234\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:21.636181 containerd[1549]: time="2025-09-08T23:51:21.636155969Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.12\" with image id \"sha256:25d00c9505e8a4a7a6c827030f878b50e58bbf63322e01a7d92807bcb4db6b3d\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.12\", repo digest \"registry.k8s.io/kube-apiserver@sha256:e9011c3bee8c06ecabd7816e119dca4e448c92f7a78acd891de3d2db1dc6c234\", size \"25649241\" in 1.690402841s" Sep 8 23:51:21.636241 containerd[1549]: time="2025-09-08T23:51:21.636189764Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.12\" returns image reference \"sha256:25d00c9505e8a4a7a6c827030f878b50e58bbf63322e01a7d92807bcb4db6b3d\"" Sep 8 23:51:21.637478 containerd[1549]: time="2025-09-08T23:51:21.637400407Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.12\"" Sep 8 23:51:22.980519 containerd[1549]: time="2025-09-08T23:51:22.979516004Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:22.980519 containerd[1549]: time="2025-09-08T23:51:22.979996447Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.12: active requests=0, bytes read=22460311" Sep 8 23:51:22.981113 containerd[1549]: time="2025-09-08T23:51:22.981087676Z" level=info msg="ImageCreate event name:\"sha256:04df324666956d4cb57096c0edff6bfe1d75e71fb8f508dec8818f2842f821e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:22.986756 containerd[1549]: time="2025-09-08T23:51:22.986308905Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:d2862f94d87320267fddbd55db26556a267aa802e51d6b60f25786b4c428afc8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:22.987356 containerd[1549]: time="2025-09-08T23:51:22.987320525Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.12\" with image id \"sha256:04df324666956d4cb57096c0edff6bfe1d75e71fb8f508dec8818f2842f821e1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.12\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:d2862f94d87320267fddbd55db26556a267aa802e51d6b60f25786b4c428afc8\", size \"23997423\" in 1.349885243s" Sep 8 23:51:22.987406 containerd[1549]: time="2025-09-08T23:51:22.987355772Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.12\" returns image reference \"sha256:04df324666956d4cb57096c0edff6bfe1d75e71fb8f508dec8818f2842f821e1\"" Sep 8 23:51:22.990431 containerd[1549]: time="2025-09-08T23:51:22.990404948Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.12\"" Sep 8 23:51:23.753013 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 8 23:51:23.754417 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:23.876634 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:23.880295 (kubelet)[2052]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 8 23:51:23.918637 kubelet[2052]: E0908 23:51:23.918590 2052 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 8 23:51:23.923025 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 8 23:51:23.923272 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 8 23:51:23.923830 systemd[1]: kubelet.service: Consumed 142ms CPU time, 108.3M memory peak. Sep 8 23:51:24.442374 containerd[1549]: time="2025-09-08T23:51:24.442324482Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:24.443279 containerd[1549]: time="2025-09-08T23:51:24.443222466Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.12: active requests=0, bytes read=17125905" Sep 8 23:51:24.443995 containerd[1549]: time="2025-09-08T23:51:24.443950505Z" level=info msg="ImageCreate event name:\"sha256:00b0619122c2d4fd3b5e102e9850d8c732e08a386b9c172c409b3a5cd552e07d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:24.447144 containerd[1549]: time="2025-09-08T23:51:24.447097897Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:152943b7e30244f4415fd0a5860a2dccd91660fe983d30a28a10edb0cc8f6756\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:24.448763 containerd[1549]: time="2025-09-08T23:51:24.448646355Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.12\" with image id \"sha256:00b0619122c2d4fd3b5e102e9850d8c732e08a386b9c172c409b3a5cd552e07d\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.12\", repo digest \"registry.k8s.io/kube-scheduler@sha256:152943b7e30244f4415fd0a5860a2dccd91660fe983d30a28a10edb0cc8f6756\", size \"18663035\" in 1.458210571s" Sep 8 23:51:24.448763 containerd[1549]: time="2025-09-08T23:51:24.448681576Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.12\" returns image reference \"sha256:00b0619122c2d4fd3b5e102e9850d8c732e08a386b9c172c409b3a5cd552e07d\"" Sep 8 23:51:24.449185 containerd[1549]: time="2025-09-08T23:51:24.449147574Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.12\"" Sep 8 23:51:25.427801 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount394893007.mount: Deactivated successfully. Sep 8 23:51:25.638251 containerd[1549]: time="2025-09-08T23:51:25.638192589Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:25.638742 containerd[1549]: time="2025-09-08T23:51:25.638698524Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.12: active requests=0, bytes read=26916097" Sep 8 23:51:25.639459 containerd[1549]: time="2025-09-08T23:51:25.639434842Z" level=info msg="ImageCreate event name:\"sha256:25c7652bd0d893b147dce9135dc6a68c37da76f9a20dceec1d520782031b2f36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:25.641171 containerd[1549]: time="2025-09-08T23:51:25.641138023Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:90aa6b5f4065937521ff8438bc705317485d0be3f8b00a07145e697d92cc2cc6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:25.642024 containerd[1549]: time="2025-09-08T23:51:25.641769124Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.12\" with image id \"sha256:25c7652bd0d893b147dce9135dc6a68c37da76f9a20dceec1d520782031b2f36\", repo tag \"registry.k8s.io/kube-proxy:v1.31.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:90aa6b5f4065937521ff8438bc705317485d0be3f8b00a07145e697d92cc2cc6\", size \"26915114\" in 1.192588307s" Sep 8 23:51:25.642024 containerd[1549]: time="2025-09-08T23:51:25.641803955Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.12\" returns image reference \"sha256:25c7652bd0d893b147dce9135dc6a68c37da76f9a20dceec1d520782031b2f36\"" Sep 8 23:51:25.642393 containerd[1549]: time="2025-09-08T23:51:25.642370583Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 8 23:51:26.322664 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount282531855.mount: Deactivated successfully. Sep 8 23:51:27.045619 containerd[1549]: time="2025-09-08T23:51:27.045538532Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:27.046402 containerd[1549]: time="2025-09-08T23:51:27.046362075Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951624" Sep 8 23:51:27.047374 containerd[1549]: time="2025-09-08T23:51:27.047333359Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:27.050159 containerd[1549]: time="2025-09-08T23:51:27.050115456Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:27.051735 containerd[1549]: time="2025-09-08T23:51:27.051692044Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.409290381s" Sep 8 23:51:27.051735 containerd[1549]: time="2025-09-08T23:51:27.051725224Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Sep 8 23:51:27.052202 containerd[1549]: time="2025-09-08T23:51:27.052155392Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 8 23:51:27.463848 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2116956315.mount: Deactivated successfully. Sep 8 23:51:27.469055 containerd[1549]: time="2025-09-08T23:51:27.468985547Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 8 23:51:27.469952 containerd[1549]: time="2025-09-08T23:51:27.469919554Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Sep 8 23:51:27.470780 containerd[1549]: time="2025-09-08T23:51:27.470752657Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 8 23:51:27.473479 containerd[1549]: time="2025-09-08T23:51:27.473424450Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 8 23:51:27.474155 containerd[1549]: time="2025-09-08T23:51:27.474126281Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 421.936262ms" Sep 8 23:51:27.474212 containerd[1549]: time="2025-09-08T23:51:27.474161870Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Sep 8 23:51:27.474579 containerd[1549]: time="2025-09-08T23:51:27.474555003Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 8 23:51:28.071531 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1210376727.mount: Deactivated successfully. Sep 8 23:51:30.183908 containerd[1549]: time="2025-09-08T23:51:30.183859035Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:30.184953 containerd[1549]: time="2025-09-08T23:51:30.184339671Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66537163" Sep 8 23:51:30.185354 containerd[1549]: time="2025-09-08T23:51:30.185310449Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:30.188087 containerd[1549]: time="2025-09-08T23:51:30.188014235Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:51:30.189189 containerd[1549]: time="2025-09-08T23:51:30.189156056Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 2.714571821s" Sep 8 23:51:30.189249 containerd[1549]: time="2025-09-08T23:51:30.189188507Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Sep 8 23:51:34.173578 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 8 23:51:34.175338 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:34.360371 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:34.364316 (kubelet)[2214]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 8 23:51:34.398573 kubelet[2214]: E0908 23:51:34.398503 2214 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 8 23:51:34.401180 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 8 23:51:34.401412 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 8 23:51:34.403124 systemd[1]: kubelet.service: Consumed 135ms CPU time, 106.8M memory peak. Sep 8 23:51:35.287404 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:35.287543 systemd[1]: kubelet.service: Consumed 135ms CPU time, 106.8M memory peak. Sep 8 23:51:35.289701 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:35.312278 systemd[1]: Reload requested from client PID 2229 ('systemctl') (unit session-7.scope)... Sep 8 23:51:35.312293 systemd[1]: Reloading... Sep 8 23:51:35.380082 zram_generator::config[2272]: No configuration found. Sep 8 23:51:35.551346 systemd[1]: Reloading finished in 238 ms. Sep 8 23:51:35.597655 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:35.600454 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:35.601335 systemd[1]: kubelet.service: Deactivated successfully. Sep 8 23:51:35.601653 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:35.601792 systemd[1]: kubelet.service: Consumed 92ms CPU time, 95.1M memory peak. Sep 8 23:51:35.603219 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:35.729056 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:35.732465 (kubelet)[2319]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 8 23:51:35.763957 kubelet[2319]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 8 23:51:35.763957 kubelet[2319]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 8 23:51:35.763957 kubelet[2319]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 8 23:51:35.764297 kubelet[2319]: I0908 23:51:35.763998 2319 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 8 23:51:36.396099 kubelet[2319]: I0908 23:51:36.394143 2319 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 8 23:51:36.396099 kubelet[2319]: I0908 23:51:36.394179 2319 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 8 23:51:36.396099 kubelet[2319]: I0908 23:51:36.394529 2319 server.go:934] "Client rotation is on, will bootstrap in background" Sep 8 23:51:36.419327 kubelet[2319]: E0908 23:51:36.419294 2319 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.130:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.130:6443: connect: connection refused" logger="UnhandledError" Sep 8 23:51:36.420249 kubelet[2319]: I0908 23:51:36.420227 2319 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 8 23:51:36.427503 kubelet[2319]: I0908 23:51:36.427474 2319 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 8 23:51:36.431228 kubelet[2319]: I0908 23:51:36.431193 2319 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 8 23:51:36.431596 kubelet[2319]: I0908 23:51:36.431572 2319 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 8 23:51:36.431715 kubelet[2319]: I0908 23:51:36.431683 2319 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 8 23:51:36.431950 kubelet[2319]: I0908 23:51:36.431712 2319 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 8 23:51:36.432065 kubelet[2319]: I0908 23:51:36.432009 2319 topology_manager.go:138] "Creating topology manager with none policy" Sep 8 23:51:36.432065 kubelet[2319]: I0908 23:51:36.432018 2319 container_manager_linux.go:300] "Creating device plugin manager" Sep 8 23:51:36.432222 kubelet[2319]: I0908 23:51:36.432204 2319 state_mem.go:36] "Initialized new in-memory state store" Sep 8 23:51:36.435051 kubelet[2319]: I0908 23:51:36.435019 2319 kubelet.go:408] "Attempting to sync node with API server" Sep 8 23:51:36.435152 kubelet[2319]: I0908 23:51:36.435141 2319 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 8 23:51:36.435243 kubelet[2319]: I0908 23:51:36.435232 2319 kubelet.go:314] "Adding apiserver pod source" Sep 8 23:51:36.435305 kubelet[2319]: I0908 23:51:36.435296 2319 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 8 23:51:36.439060 kubelet[2319]: W0908 23:51:36.438858 2319 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.130:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.130:6443: connect: connection refused Sep 8 23:51:36.439060 kubelet[2319]: E0908 23:51:36.438925 2319 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.130:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.130:6443: connect: connection refused" logger="UnhandledError" Sep 8 23:51:36.439158 kubelet[2319]: W0908 23:51:36.439091 2319 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.130:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.130:6443: connect: connection refused Sep 8 23:51:36.439158 kubelet[2319]: E0908 23:51:36.439122 2319 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.130:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.130:6443: connect: connection refused" logger="UnhandledError" Sep 8 23:51:36.440202 kubelet[2319]: I0908 23:51:36.440172 2319 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 8 23:51:36.440999 kubelet[2319]: I0908 23:51:36.440974 2319 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 8 23:51:36.441196 kubelet[2319]: W0908 23:51:36.441177 2319 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 8 23:51:36.442313 kubelet[2319]: I0908 23:51:36.442279 2319 server.go:1274] "Started kubelet" Sep 8 23:51:36.444202 kubelet[2319]: I0908 23:51:36.444159 2319 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 8 23:51:36.444462 kubelet[2319]: I0908 23:51:36.444445 2319 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 8 23:51:36.444572 kubelet[2319]: I0908 23:51:36.444526 2319 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 8 23:51:36.444654 kubelet[2319]: I0908 23:51:36.444634 2319 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 8 23:51:36.445355 kubelet[2319]: I0908 23:51:36.445324 2319 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 8 23:51:36.445952 kubelet[2319]: I0908 23:51:36.445932 2319 server.go:449] "Adding debug handlers to kubelet server" Sep 8 23:51:36.446889 kubelet[2319]: I0908 23:51:36.446855 2319 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 8 23:51:36.446979 kubelet[2319]: I0908 23:51:36.446962 2319 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 8 23:51:36.447034 kubelet[2319]: I0908 23:51:36.447020 2319 reconciler.go:26] "Reconciler: start to sync state" Sep 8 23:51:36.447517 kubelet[2319]: W0908 23:51:36.447418 2319 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.130:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.130:6443: connect: connection refused Sep 8 23:51:36.447517 kubelet[2319]: E0908 23:51:36.447466 2319 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.130:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.130:6443: connect: connection refused" logger="UnhandledError" Sep 8 23:51:36.447602 kubelet[2319]: E0908 23:51:36.447552 2319 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 8 23:51:36.447708 kubelet[2319]: I0908 23:51:36.447685 2319 factory.go:221] Registration of the systemd container factory successfully Sep 8 23:51:36.447784 kubelet[2319]: I0908 23:51:36.447764 2319 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 8 23:51:36.447994 kubelet[2319]: E0908 23:51:36.447975 2319 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 8 23:51:36.448152 kubelet[2319]: E0908 23:51:36.448129 2319 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.130:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.130:6443: connect: connection refused" interval="200ms" Sep 8 23:51:36.448774 kubelet[2319]: I0908 23:51:36.448739 2319 factory.go:221] Registration of the containerd container factory successfully Sep 8 23:51:36.449887 kubelet[2319]: E0908 23:51:36.445731 2319 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.130:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.130:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.186373b3fa6a6f7b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-08 23:51:36.442244987 +0000 UTC m=+0.706779573,LastTimestamp:2025-09-08 23:51:36.442244987 +0000 UTC m=+0.706779573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 8 23:51:36.459659 kubelet[2319]: I0908 23:51:36.459619 2319 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 8 23:51:36.459659 kubelet[2319]: I0908 23:51:36.459654 2319 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 8 23:51:36.459745 kubelet[2319]: I0908 23:51:36.459670 2319 state_mem.go:36] "Initialized new in-memory state store" Sep 8 23:51:36.549150 kubelet[2319]: E0908 23:51:36.549102 2319 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 8 23:51:36.572114 kubelet[2319]: I0908 23:51:36.572084 2319 policy_none.go:49] "None policy: Start" Sep 8 23:51:36.572828 kubelet[2319]: I0908 23:51:36.572805 2319 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 8 23:51:36.572890 kubelet[2319]: I0908 23:51:36.572836 2319 state_mem.go:35] "Initializing new in-memory state store" Sep 8 23:51:36.575551 kubelet[2319]: I0908 23:51:36.575507 2319 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 8 23:51:36.577714 kubelet[2319]: I0908 23:51:36.577686 2319 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 8 23:51:36.577714 kubelet[2319]: I0908 23:51:36.577710 2319 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 8 23:51:36.577784 kubelet[2319]: I0908 23:51:36.577729 2319 kubelet.go:2321] "Starting kubelet main sync loop" Sep 8 23:51:36.577784 kubelet[2319]: E0908 23:51:36.577774 2319 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 8 23:51:36.578538 kubelet[2319]: W0908 23:51:36.578505 2319 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.130:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.130:6443: connect: connection refused Sep 8 23:51:36.578598 kubelet[2319]: E0908 23:51:36.578548 2319 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.130:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.130:6443: connect: connection refused" logger="UnhandledError" Sep 8 23:51:36.580935 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 8 23:51:36.594892 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 8 23:51:36.597933 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 8 23:51:36.612954 kubelet[2319]: I0908 23:51:36.612924 2319 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 8 23:51:36.613211 kubelet[2319]: I0908 23:51:36.613179 2319 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 8 23:51:36.613211 kubelet[2319]: I0908 23:51:36.613192 2319 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 8 23:51:36.613422 kubelet[2319]: I0908 23:51:36.613402 2319 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 8 23:51:36.614662 kubelet[2319]: E0908 23:51:36.614638 2319 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 8 23:51:36.648892 kubelet[2319]: E0908 23:51:36.648627 2319 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.130:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.130:6443: connect: connection refused" interval="400ms" Sep 8 23:51:36.686400 systemd[1]: Created slice kubepods-burstable-podfec3f691a145cb26ff55e4af388500b7.slice - libcontainer container kubepods-burstable-podfec3f691a145cb26ff55e4af388500b7.slice. Sep 8 23:51:36.709061 systemd[1]: Created slice kubepods-burstable-pod5dc878868de11c6196259ae42039f4ff.slice - libcontainer container kubepods-burstable-pod5dc878868de11c6196259ae42039f4ff.slice. Sep 8 23:51:36.715061 kubelet[2319]: I0908 23:51:36.715021 2319 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 8 23:51:36.715584 kubelet[2319]: E0908 23:51:36.715541 2319 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.130:6443/api/v1/nodes\": dial tcp 10.0.0.130:6443: connect: connection refused" node="localhost" Sep 8 23:51:36.728590 systemd[1]: Created slice kubepods-burstable-pod84af9435cce1efe8d44c40f715df1ed7.slice - libcontainer container kubepods-burstable-pod84af9435cce1efe8d44c40f715df1ed7.slice. Sep 8 23:51:36.748952 kubelet[2319]: I0908 23:51:36.748800 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:36.748952 kubelet[2319]: I0908 23:51:36.748836 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:36.748952 kubelet[2319]: I0908 23:51:36.748856 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:36.748952 kubelet[2319]: I0908 23:51:36.748875 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:36.748952 kubelet[2319]: I0908 23:51:36.748904 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:36.749149 kubelet[2319]: I0908 23:51:36.748923 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:36.749149 kubelet[2319]: I0908 23:51:36.748963 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5dc878868de11c6196259ae42039f4ff-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"5dc878868de11c6196259ae42039f4ff\") " pod="kube-system/kube-scheduler-localhost" Sep 8 23:51:36.749149 kubelet[2319]: I0908 23:51:36.748994 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:36.749149 kubelet[2319]: I0908 23:51:36.749013 2319 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:36.917237 kubelet[2319]: I0908 23:51:36.917137 2319 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 8 23:51:36.917525 kubelet[2319]: E0908 23:51:36.917434 2319 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.130:6443/api/v1/nodes\": dial tcp 10.0.0.130:6443: connect: connection refused" node="localhost" Sep 8 23:51:37.007470 containerd[1549]: time="2025-09-08T23:51:37.007428360Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fec3f691a145cb26ff55e4af388500b7,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:37.027502 containerd[1549]: time="2025-09-08T23:51:37.027459945Z" level=info msg="connecting to shim 79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705" address="unix:///run/containerd/s/98d7d7bb61735a64a986c490ca631524b6b296bdba18ba1ad88116ea7f8bb54f" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:37.027502 containerd[1549]: time="2025-09-08T23:51:37.027498988Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:5dc878868de11c6196259ae42039f4ff,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:37.031484 containerd[1549]: time="2025-09-08T23:51:37.031452535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:84af9435cce1efe8d44c40f715df1ed7,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:37.049080 kubelet[2319]: E0908 23:51:37.048976 2319 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.130:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.130:6443: connect: connection refused" interval="800ms" Sep 8 23:51:37.054285 containerd[1549]: time="2025-09-08T23:51:37.054196129Z" level=info msg="connecting to shim ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0" address="unix:///run/containerd/s/e65863c43c205e8580031ff7559ea86127716772f38723e1bcac1e911922ba58" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:37.057238 systemd[1]: Started cri-containerd-79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705.scope - libcontainer container 79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705. Sep 8 23:51:37.063429 containerd[1549]: time="2025-09-08T23:51:37.063379478Z" level=info msg="connecting to shim 54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95" address="unix:///run/containerd/s/2b28a94583196ee482271540bd2328aab5340c018c6de80275da3ccac951b8ba" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:37.080204 systemd[1]: Started cri-containerd-ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0.scope - libcontainer container ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0. Sep 8 23:51:37.085920 systemd[1]: Started cri-containerd-54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95.scope - libcontainer container 54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95. Sep 8 23:51:37.101866 containerd[1549]: time="2025-09-08T23:51:37.101824693Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fec3f691a145cb26ff55e4af388500b7,Namespace:kube-system,Attempt:0,} returns sandbox id \"79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705\"" Sep 8 23:51:37.104530 containerd[1549]: time="2025-09-08T23:51:37.104498820Z" level=info msg="CreateContainer within sandbox \"79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 8 23:51:37.114907 containerd[1549]: time="2025-09-08T23:51:37.114765291Z" level=info msg="Container 0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:51:37.122795 containerd[1549]: time="2025-09-08T23:51:37.122742261Z" level=info msg="CreateContainer within sandbox \"79c4cf649f11363c895ad6aa214a41dfe617c7a70fe1e589609746df61407705\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48\"" Sep 8 23:51:37.122972 containerd[1549]: time="2025-09-08T23:51:37.122946328Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:5dc878868de11c6196259ae42039f4ff,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0\"" Sep 8 23:51:37.123710 containerd[1549]: time="2025-09-08T23:51:37.123672453Z" level=info msg="StartContainer for \"0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48\"" Sep 8 23:51:37.126185 containerd[1549]: time="2025-09-08T23:51:37.126152485Z" level=info msg="connecting to shim 0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48" address="unix:///run/containerd/s/98d7d7bb61735a64a986c490ca631524b6b296bdba18ba1ad88116ea7f8bb54f" protocol=ttrpc version=3 Sep 8 23:51:37.127847 containerd[1549]: time="2025-09-08T23:51:37.127429182Z" level=info msg="CreateContainer within sandbox \"ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 8 23:51:37.129778 containerd[1549]: time="2025-09-08T23:51:37.129554580Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:84af9435cce1efe8d44c40f715df1ed7,Namespace:kube-system,Attempt:0,} returns sandbox id \"54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95\"" Sep 8 23:51:37.132129 containerd[1549]: time="2025-09-08T23:51:37.132092315Z" level=info msg="CreateContainer within sandbox \"54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 8 23:51:37.136866 containerd[1549]: time="2025-09-08T23:51:37.136824446Z" level=info msg="Container e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:51:37.143256 containerd[1549]: time="2025-09-08T23:51:37.143211692Z" level=info msg="Container 3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:51:37.150215 systemd[1]: Started cri-containerd-0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48.scope - libcontainer container 0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48. Sep 8 23:51:37.151383 containerd[1549]: time="2025-09-08T23:51:37.151246847Z" level=info msg="CreateContainer within sandbox \"ba34215ad49bdcf89d264bf72b039937a38b58652899caba63aa5bfa8ed06ab0\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d\"" Sep 8 23:51:37.151838 containerd[1549]: time="2025-09-08T23:51:37.151812755Z" level=info msg="StartContainer for \"e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d\"" Sep 8 23:51:37.151888 containerd[1549]: time="2025-09-08T23:51:37.151854642Z" level=info msg="CreateContainer within sandbox \"54f06b52ee8cf9a41a18fd38874e7c6db21a71d8bfc4d34b44e8cfe2641afe95\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2\"" Sep 8 23:51:37.152708 containerd[1549]: time="2025-09-08T23:51:37.152256888Z" level=info msg="StartContainer for \"3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2\"" Sep 8 23:51:37.153219 containerd[1549]: time="2025-09-08T23:51:37.153165776Z" level=info msg="connecting to shim e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d" address="unix:///run/containerd/s/e65863c43c205e8580031ff7559ea86127716772f38723e1bcac1e911922ba58" protocol=ttrpc version=3 Sep 8 23:51:37.154164 containerd[1549]: time="2025-09-08T23:51:37.154135532Z" level=info msg="connecting to shim 3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2" address="unix:///run/containerd/s/2b28a94583196ee482271540bd2328aab5340c018c6de80275da3ccac951b8ba" protocol=ttrpc version=3 Sep 8 23:51:37.179207 systemd[1]: Started cri-containerd-3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2.scope - libcontainer container 3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2. Sep 8 23:51:37.183132 systemd[1]: Started cri-containerd-e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d.scope - libcontainer container e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d. Sep 8 23:51:37.195913 containerd[1549]: time="2025-09-08T23:51:37.195877005Z" level=info msg="StartContainer for \"0a82fa9c77763b52420a5dd7a949d0668aec60808666b8dffa275146c0901f48\" returns successfully" Sep 8 23:51:37.226245 containerd[1549]: time="2025-09-08T23:51:37.226207457Z" level=info msg="StartContainer for \"3207a695316ba3ed328861f8130bf4c52a8e782489e4b7a366be9fa9b8a372d2\" returns successfully" Sep 8 23:51:37.239111 containerd[1549]: time="2025-09-08T23:51:37.239070529Z" level=info msg="StartContainer for \"e08283c0e56e4c851d5d704eace3aedd03d57f1d2753757d71859d67ad8c366d\" returns successfully" Sep 8 23:51:37.320472 kubelet[2319]: I0908 23:51:37.320191 2319 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 8 23:51:38.787324 kubelet[2319]: E0908 23:51:38.787283 2319 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 8 23:51:38.858824 kubelet[2319]: I0908 23:51:38.858591 2319 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 8 23:51:38.858824 kubelet[2319]: E0908 23:51:38.858627 2319 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 8 23:51:39.436712 kubelet[2319]: I0908 23:51:39.436645 2319 apiserver.go:52] "Watching apiserver" Sep 8 23:51:39.447614 kubelet[2319]: I0908 23:51:39.447563 2319 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 8 23:51:41.170466 systemd[1]: Reload requested from client PID 2594 ('systemctl') (unit session-7.scope)... Sep 8 23:51:41.170481 systemd[1]: Reloading... Sep 8 23:51:41.255097 zram_generator::config[2643]: No configuration found. Sep 8 23:51:41.520773 systemd[1]: Reloading finished in 349 ms. Sep 8 23:51:41.548479 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:41.565104 systemd[1]: kubelet.service: Deactivated successfully. Sep 8 23:51:41.566113 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:41.566171 systemd[1]: kubelet.service: Consumed 1.083s CPU time, 129.9M memory peak. Sep 8 23:51:41.567987 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 8 23:51:41.711215 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 8 23:51:41.730825 (kubelet)[2679]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 8 23:51:41.777247 kubelet[2679]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 8 23:51:41.777247 kubelet[2679]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 8 23:51:41.777247 kubelet[2679]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 8 23:51:41.777963 kubelet[2679]: I0908 23:51:41.777906 2679 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 8 23:51:41.783557 kubelet[2679]: I0908 23:51:41.783512 2679 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 8 23:51:41.783557 kubelet[2679]: I0908 23:51:41.783543 2679 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 8 23:51:41.783792 kubelet[2679]: I0908 23:51:41.783758 2679 server.go:934] "Client rotation is on, will bootstrap in background" Sep 8 23:51:41.785139 kubelet[2679]: I0908 23:51:41.785110 2679 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 8 23:51:41.788113 kubelet[2679]: I0908 23:51:41.788074 2679 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 8 23:51:41.792713 kubelet[2679]: I0908 23:51:41.792286 2679 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 8 23:51:41.796048 kubelet[2679]: I0908 23:51:41.796008 2679 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 8 23:51:41.796146 kubelet[2679]: I0908 23:51:41.796126 2679 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 8 23:51:41.796240 kubelet[2679]: I0908 23:51:41.796215 2679 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 8 23:51:41.796419 kubelet[2679]: I0908 23:51:41.796239 2679 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 8 23:51:41.796419 kubelet[2679]: I0908 23:51:41.796417 2679 topology_manager.go:138] "Creating topology manager with none policy" Sep 8 23:51:41.796511 kubelet[2679]: I0908 23:51:41.796428 2679 container_manager_linux.go:300] "Creating device plugin manager" Sep 8 23:51:41.796511 kubelet[2679]: I0908 23:51:41.796460 2679 state_mem.go:36] "Initialized new in-memory state store" Sep 8 23:51:41.796570 kubelet[2679]: I0908 23:51:41.796554 2679 kubelet.go:408] "Attempting to sync node with API server" Sep 8 23:51:41.796570 kubelet[2679]: I0908 23:51:41.796565 2679 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 8 23:51:41.798086 kubelet[2679]: I0908 23:51:41.796581 2679 kubelet.go:314] "Adding apiserver pod source" Sep 8 23:51:41.798086 kubelet[2679]: I0908 23:51:41.796590 2679 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 8 23:51:41.798086 kubelet[2679]: I0908 23:51:41.797085 2679 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 8 23:51:41.800319 kubelet[2679]: I0908 23:51:41.800290 2679 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 8 23:51:41.800746 kubelet[2679]: I0908 23:51:41.800719 2679 server.go:1274] "Started kubelet" Sep 8 23:51:41.800889 kubelet[2679]: I0908 23:51:41.800855 2679 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 8 23:51:41.801016 kubelet[2679]: I0908 23:51:41.800968 2679 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 8 23:51:41.801213 kubelet[2679]: I0908 23:51:41.801185 2679 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 8 23:51:41.802226 kubelet[2679]: I0908 23:51:41.802198 2679 server.go:449] "Adding debug handlers to kubelet server" Sep 8 23:51:41.802391 kubelet[2679]: I0908 23:51:41.802371 2679 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 8 23:51:41.802560 kubelet[2679]: I0908 23:51:41.802532 2679 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 8 23:51:41.803311 kubelet[2679]: E0908 23:51:41.803261 2679 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 8 23:51:41.803383 kubelet[2679]: I0908 23:51:41.803317 2679 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 8 23:51:41.804219 kubelet[2679]: I0908 23:51:41.804189 2679 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 8 23:51:41.804387 kubelet[2679]: I0908 23:51:41.804369 2679 reconciler.go:26] "Reconciler: start to sync state" Sep 8 23:51:41.823613 kubelet[2679]: I0908 23:51:41.823570 2679 factory.go:221] Registration of the systemd container factory successfully Sep 8 23:51:41.823726 kubelet[2679]: I0908 23:51:41.823698 2679 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 8 23:51:41.826475 kubelet[2679]: I0908 23:51:41.826442 2679 factory.go:221] Registration of the containerd container factory successfully Sep 8 23:51:41.831870 kubelet[2679]: I0908 23:51:41.831839 2679 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 8 23:51:41.833464 kubelet[2679]: I0908 23:51:41.833436 2679 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 8 23:51:41.833464 kubelet[2679]: I0908 23:51:41.833464 2679 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 8 23:51:41.833558 kubelet[2679]: I0908 23:51:41.833482 2679 kubelet.go:2321] "Starting kubelet main sync loop" Sep 8 23:51:41.833558 kubelet[2679]: E0908 23:51:41.833522 2679 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 8 23:51:41.855446 kubelet[2679]: I0908 23:51:41.855416 2679 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 8 23:51:41.855446 kubelet[2679]: I0908 23:51:41.855437 2679 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 8 23:51:41.855446 kubelet[2679]: I0908 23:51:41.855460 2679 state_mem.go:36] "Initialized new in-memory state store" Sep 8 23:51:41.855629 kubelet[2679]: I0908 23:51:41.855610 2679 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 8 23:51:41.855655 kubelet[2679]: I0908 23:51:41.855627 2679 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 8 23:51:41.855655 kubelet[2679]: I0908 23:51:41.855648 2679 policy_none.go:49] "None policy: Start" Sep 8 23:51:41.856321 kubelet[2679]: I0908 23:51:41.856306 2679 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 8 23:51:41.856354 kubelet[2679]: I0908 23:51:41.856328 2679 state_mem.go:35] "Initializing new in-memory state store" Sep 8 23:51:41.856481 kubelet[2679]: I0908 23:51:41.856466 2679 state_mem.go:75] "Updated machine memory state" Sep 8 23:51:41.860518 kubelet[2679]: I0908 23:51:41.860489 2679 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 8 23:51:41.860908 kubelet[2679]: I0908 23:51:41.860673 2679 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 8 23:51:41.860908 kubelet[2679]: I0908 23:51:41.860692 2679 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 8 23:51:41.860908 kubelet[2679]: I0908 23:51:41.860849 2679 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 8 23:51:41.940635 kubelet[2679]: E0908 23:51:41.940594 2679 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:41.962999 kubelet[2679]: I0908 23:51:41.962965 2679 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 8 23:51:41.968821 kubelet[2679]: I0908 23:51:41.968792 2679 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Sep 8 23:51:41.968896 kubelet[2679]: I0908 23:51:41.968870 2679 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 8 23:51:42.005692 kubelet[2679]: I0908 23:51:42.005641 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5dc878868de11c6196259ae42039f4ff-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"5dc878868de11c6196259ae42039f4ff\") " pod="kube-system/kube-scheduler-localhost" Sep 8 23:51:42.005692 kubelet[2679]: I0908 23:51:42.005683 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:42.005853 kubelet[2679]: I0908 23:51:42.005703 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:42.005853 kubelet[2679]: I0908 23:51:42.005745 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.005853 kubelet[2679]: I0908 23:51:42.005763 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.005853 kubelet[2679]: I0908 23:51:42.005780 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/84af9435cce1efe8d44c40f715df1ed7-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"84af9435cce1efe8d44c40f715df1ed7\") " pod="kube-system/kube-apiserver-localhost" Sep 8 23:51:42.005853 kubelet[2679]: I0908 23:51:42.005794 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.005954 kubelet[2679]: I0908 23:51:42.005812 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.005954 kubelet[2679]: I0908 23:51:42.005828 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fec3f691a145cb26ff55e4af388500b7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fec3f691a145cb26ff55e4af388500b7\") " pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.167904 sudo[2713]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 8 23:51:42.169193 sudo[2713]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 8 23:51:42.483675 sudo[2713]: pam_unix(sudo:session): session closed for user root Sep 8 23:51:42.797123 kubelet[2679]: I0908 23:51:42.797086 2679 apiserver.go:52] "Watching apiserver" Sep 8 23:51:42.805462 kubelet[2679]: I0908 23:51:42.805421 2679 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 8 23:51:42.852624 kubelet[2679]: E0908 23:51:42.852556 2679 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Sep 8 23:51:42.856582 kubelet[2679]: E0908 23:51:42.856135 2679 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Sep 8 23:51:42.866794 kubelet[2679]: I0908 23:51:42.866737 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.866710904 podStartE2EDuration="1.866710904s" podCreationTimestamp="2025-09-08 23:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:51:42.866532282 +0000 UTC m=+1.132431599" watchObservedRunningTime="2025-09-08 23:51:42.866710904 +0000 UTC m=+1.132610220" Sep 8 23:51:42.884910 kubelet[2679]: I0908 23:51:42.884844 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.88482818 podStartE2EDuration="1.88482818s" podCreationTimestamp="2025-09-08 23:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:51:42.875663402 +0000 UTC m=+1.141562718" watchObservedRunningTime="2025-09-08 23:51:42.88482818 +0000 UTC m=+1.150727496" Sep 8 23:51:42.911073 kubelet[2679]: I0908 23:51:42.909373 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.909355105 podStartE2EDuration="1.909355105s" podCreationTimestamp="2025-09-08 23:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:51:42.885401226 +0000 UTC m=+1.151300542" watchObservedRunningTime="2025-09-08 23:51:42.909355105 +0000 UTC m=+1.175254421" Sep 8 23:51:43.930286 sudo[1750]: pam_unix(sudo:session): session closed for user root Sep 8 23:51:43.934598 sshd[1749]: Connection closed by 10.0.0.1 port 42854 Sep 8 23:51:43.932020 sshd-session[1746]: pam_unix(sshd:session): session closed for user core Sep 8 23:51:43.939097 systemd-logind[1524]: Session 7 logged out. Waiting for processes to exit. Sep 8 23:51:43.939326 systemd[1]: sshd@6-10.0.0.130:22-10.0.0.1:42854.service: Deactivated successfully. Sep 8 23:51:43.941120 systemd[1]: session-7.scope: Deactivated successfully. Sep 8 23:51:43.941314 systemd[1]: session-7.scope: Consumed 6.843s CPU time, 258.2M memory peak. Sep 8 23:51:43.943489 systemd-logind[1524]: Removed session 7. Sep 8 23:51:45.683992 kubelet[2679]: I0908 23:51:45.683851 2679 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 8 23:51:45.684389 containerd[1549]: time="2025-09-08T23:51:45.684320182Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 8 23:51:45.684853 kubelet[2679]: I0908 23:51:45.684834 2679 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 8 23:51:46.506738 systemd[1]: Created slice kubepods-besteffort-pod972d9b2c_25c8_4a56_8413_d8147da39701.slice - libcontainer container kubepods-besteffort-pod972d9b2c_25c8_4a56_8413_d8147da39701.slice. Sep 8 23:51:46.522458 systemd[1]: Created slice kubepods-burstable-podd4d502e4_c0c6_4e11_98be_8a23553b9240.slice - libcontainer container kubepods-burstable-podd4d502e4_c0c6_4e11_98be_8a23553b9240.slice. Sep 8 23:51:46.535420 kubelet[2679]: I0908 23:51:46.535304 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-hostproc\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.535698 kubelet[2679]: I0908 23:51:46.535612 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-hubble-tls\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.535698 kubelet[2679]: I0908 23:51:46.535638 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/972d9b2c-25c8-4a56-8413-d8147da39701-kube-proxy\") pod \"kube-proxy-qrrpc\" (UID: \"972d9b2c-25c8-4a56-8413-d8147da39701\") " pod="kube-system/kube-proxy-qrrpc" Sep 8 23:51:46.535865 kubelet[2679]: I0908 23:51:46.535788 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/972d9b2c-25c8-4a56-8413-d8147da39701-xtables-lock\") pod \"kube-proxy-qrrpc\" (UID: \"972d9b2c-25c8-4a56-8413-d8147da39701\") " pod="kube-system/kube-proxy-qrrpc" Sep 8 23:51:46.535865 kubelet[2679]: I0908 23:51:46.535816 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-lib-modules\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536047 kubelet[2679]: I0908 23:51:46.535956 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/972d9b2c-25c8-4a56-8413-d8147da39701-lib-modules\") pod \"kube-proxy-qrrpc\" (UID: \"972d9b2c-25c8-4a56-8413-d8147da39701\") " pod="kube-system/kube-proxy-qrrpc" Sep 8 23:51:46.536047 kubelet[2679]: I0908 23:51:46.535985 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-bpf-maps\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536183 kubelet[2679]: I0908 23:51:46.536137 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-config-path\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536313 kubelet[2679]: I0908 23:51:46.536256 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-kernel\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536313 kubelet[2679]: I0908 23:51:46.536284 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d4d502e4-c0c6-4e11-98be-8a23553b9240-clustermesh-secrets\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536466 kubelet[2679]: I0908 23:51:46.536408 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-run\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536466 kubelet[2679]: I0908 23:51:46.536432 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-cgroup\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536466 kubelet[2679]: I0908 23:51:46.536449 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-xtables-lock\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536652 kubelet[2679]: I0908 23:51:46.536628 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v996j\" (UniqueName: \"kubernetes.io/projected/972d9b2c-25c8-4a56-8413-d8147da39701-kube-api-access-v996j\") pod \"kube-proxy-qrrpc\" (UID: \"972d9b2c-25c8-4a56-8413-d8147da39701\") " pod="kube-system/kube-proxy-qrrpc" Sep 8 23:51:46.536762 kubelet[2679]: I0908 23:51:46.536750 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-net\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536871 kubelet[2679]: I0908 23:51:46.536859 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn9dh\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-kube-api-access-rn9dh\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.536983 kubelet[2679]: I0908 23:51:46.536971 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cni-path\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.537150 kubelet[2679]: I0908 23:51:46.537095 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-etc-cni-netd\") pod \"cilium-zb7sk\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " pod="kube-system/cilium-zb7sk" Sep 8 23:51:46.748978 systemd[1]: Created slice kubepods-besteffort-podb4e7c630_7500_4e07_bd47_d442efe93d8e.slice - libcontainer container kubepods-besteffort-podb4e7c630_7500_4e07_bd47_d442efe93d8e.slice. Sep 8 23:51:46.820650 containerd[1549]: time="2025-09-08T23:51:46.820550841Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qrrpc,Uid:972d9b2c-25c8-4a56-8413-d8147da39701,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:46.826185 containerd[1549]: time="2025-09-08T23:51:46.826153110Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zb7sk,Uid:d4d502e4-c0c6-4e11-98be-8a23553b9240,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:46.843069 kubelet[2679]: I0908 23:51:46.843022 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b4e7c630-7500-4e07-bd47-d442efe93d8e-cilium-config-path\") pod \"cilium-operator-5d85765b45-2lblt\" (UID: \"b4e7c630-7500-4e07-bd47-d442efe93d8e\") " pod="kube-system/cilium-operator-5d85765b45-2lblt" Sep 8 23:51:46.843507 kubelet[2679]: I0908 23:51:46.843450 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqwf4\" (UniqueName: \"kubernetes.io/projected/b4e7c630-7500-4e07-bd47-d442efe93d8e-kube-api-access-cqwf4\") pod \"cilium-operator-5d85765b45-2lblt\" (UID: \"b4e7c630-7500-4e07-bd47-d442efe93d8e\") " pod="kube-system/cilium-operator-5d85765b45-2lblt" Sep 8 23:51:46.882980 containerd[1549]: time="2025-09-08T23:51:46.882739563Z" level=info msg="connecting to shim ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea" address="unix:///run/containerd/s/9fba54bb811fa9eb157c9bf8938f7800578f4179cb95e9f5852344caed580c19" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:46.885883 containerd[1549]: time="2025-09-08T23:51:46.885841532Z" level=info msg="connecting to shim a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:46.908207 systemd[1]: Started cri-containerd-a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206.scope - libcontainer container a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206. Sep 8 23:51:46.911339 systemd[1]: Started cri-containerd-ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea.scope - libcontainer container ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea. Sep 8 23:51:46.936857 containerd[1549]: time="2025-09-08T23:51:46.936819953Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zb7sk,Uid:d4d502e4-c0c6-4e11-98be-8a23553b9240,Namespace:kube-system,Attempt:0,} returns sandbox id \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\"" Sep 8 23:51:46.939030 containerd[1549]: time="2025-09-08T23:51:46.938934467Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 8 23:51:46.951681 containerd[1549]: time="2025-09-08T23:51:46.951646643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qrrpc,Uid:972d9b2c-25c8-4a56-8413-d8147da39701,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea\"" Sep 8 23:51:46.954880 containerd[1549]: time="2025-09-08T23:51:46.954846230Z" level=info msg="CreateContainer within sandbox \"ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 8 23:51:46.964078 containerd[1549]: time="2025-09-08T23:51:46.963931330Z" level=info msg="Container 06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:51:46.970153 containerd[1549]: time="2025-09-08T23:51:46.970119380Z" level=info msg="CreateContainer within sandbox \"ba5372d4a03a957288d8eb2a22680e5b51e38e56c4f39b9a553742011ef69bea\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7\"" Sep 8 23:51:46.970898 containerd[1549]: time="2025-09-08T23:51:46.970839280Z" level=info msg="StartContainer for \"06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7\"" Sep 8 23:51:46.972378 containerd[1549]: time="2025-09-08T23:51:46.972351883Z" level=info msg="connecting to shim 06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7" address="unix:///run/containerd/s/9fba54bb811fa9eb157c9bf8938f7800578f4179cb95e9f5852344caed580c19" protocol=ttrpc version=3 Sep 8 23:51:46.993195 systemd[1]: Started cri-containerd-06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7.scope - libcontainer container 06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7. Sep 8 23:51:47.030087 containerd[1549]: time="2025-09-08T23:51:47.029948923Z" level=info msg="StartContainer for \"06be1745ed1007a5dcfc4ea9fa6ce1222975509960f6cb4242658b8bc1ede7a7\" returns successfully" Sep 8 23:51:47.055021 containerd[1549]: time="2025-09-08T23:51:47.054978807Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-2lblt,Uid:b4e7c630-7500-4e07-bd47-d442efe93d8e,Namespace:kube-system,Attempt:0,}" Sep 8 23:51:47.071257 containerd[1549]: time="2025-09-08T23:51:47.071124591Z" level=info msg="connecting to shim 4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb" address="unix:///run/containerd/s/70df2fef14da57b9fe4be1dbec34ca5da02dc898bc069801baa4fa072b61b4fb" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:51:47.099289 systemd[1]: Started cri-containerd-4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb.scope - libcontainer container 4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb. Sep 8 23:51:47.135283 containerd[1549]: time="2025-09-08T23:51:47.135247435Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-2lblt,Uid:b4e7c630-7500-4e07-bd47-d442efe93d8e,Namespace:kube-system,Attempt:0,} returns sandbox id \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\"" Sep 8 23:51:50.484131 kubelet[2679]: I0908 23:51:50.484031 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-qrrpc" podStartSLOduration=4.484014314 podStartE2EDuration="4.484014314s" podCreationTimestamp="2025-09-08 23:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:51:47.863388484 +0000 UTC m=+6.129287800" watchObservedRunningTime="2025-09-08 23:51:50.484014314 +0000 UTC m=+8.749913630" Sep 8 23:51:56.512490 update_engine[1527]: I20250908 23:51:56.511781 1527 update_attempter.cc:509] Updating boot flags... Sep 8 23:52:05.441066 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1279934870.mount: Deactivated successfully. Sep 8 23:52:06.613951 containerd[1549]: time="2025-09-08T23:52:06.613892898Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Sep 8 23:52:06.616849 containerd[1549]: time="2025-09-08T23:52:06.616801758Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 19.677820026s" Sep 8 23:52:06.616849 containerd[1549]: time="2025-09-08T23:52:06.616844407Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Sep 8 23:52:06.619571 containerd[1549]: time="2025-09-08T23:52:06.619528059Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 8 23:52:06.624452 containerd[1549]: time="2025-09-08T23:52:06.624421303Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 8 23:52:06.627125 containerd[1549]: time="2025-09-08T23:52:06.627019617Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:52:06.629252 containerd[1549]: time="2025-09-08T23:52:06.629208043Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:52:06.636671 containerd[1549]: time="2025-09-08T23:52:06.635946680Z" level=info msg="Container 1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:06.640767 containerd[1549]: time="2025-09-08T23:52:06.640705495Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\"" Sep 8 23:52:06.643029 containerd[1549]: time="2025-09-08T23:52:06.642666473Z" level=info msg="StartContainer for \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\"" Sep 8 23:52:06.657884 containerd[1549]: time="2025-09-08T23:52:06.657829105Z" level=info msg="connecting to shim 1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" protocol=ttrpc version=3 Sep 8 23:52:06.708291 systemd[1]: Started cri-containerd-1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997.scope - libcontainer container 1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997. Sep 8 23:52:06.739845 containerd[1549]: time="2025-09-08T23:52:06.739799222Z" level=info msg="StartContainer for \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" returns successfully" Sep 8 23:52:06.750927 systemd[1]: cri-containerd-1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997.scope: Deactivated successfully. Sep 8 23:52:06.783164 containerd[1549]: time="2025-09-08T23:52:06.783032120Z" level=info msg="received exit event container_id:\"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" id:\"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" pid:3113 exited_at:{seconds:1757375526 nanos:779699850}" Sep 8 23:52:06.783303 containerd[1549]: time="2025-09-08T23:52:06.783104095Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" id:\"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" pid:3113 exited_at:{seconds:1757375526 nanos:779699850}" Sep 8 23:52:06.822709 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997-rootfs.mount: Deactivated successfully. Sep 8 23:52:07.684669 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2960587318.mount: Deactivated successfully. Sep 8 23:52:07.957623 containerd[1549]: time="2025-09-08T23:52:07.957031993Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 8 23:52:07.993276 containerd[1549]: time="2025-09-08T23:52:07.993221391Z" level=info msg="Container ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:07.999002 containerd[1549]: time="2025-09-08T23:52:07.998946761Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\"" Sep 8 23:52:07.999845 containerd[1549]: time="2025-09-08T23:52:07.999825141Z" level=info msg="StartContainer for \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\"" Sep 8 23:52:08.000806 containerd[1549]: time="2025-09-08T23:52:08.000765893Z" level=info msg="connecting to shim ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" protocol=ttrpc version=3 Sep 8 23:52:08.026231 systemd[1]: Started cri-containerd-ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a.scope - libcontainer container ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a. Sep 8 23:52:08.057267 containerd[1549]: time="2025-09-08T23:52:08.057229293Z" level=info msg="StartContainer for \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" returns successfully" Sep 8 23:52:08.068230 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 8 23:52:08.068665 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 8 23:52:08.068916 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 8 23:52:08.071321 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 8 23:52:08.073022 containerd[1549]: time="2025-09-08T23:52:08.072080447Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" id:\"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" pid:3170 exited_at:{seconds:1757375528 nanos:71436561}" Sep 8 23:52:08.072567 systemd[1]: cri-containerd-ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a.scope: Deactivated successfully. Sep 8 23:52:08.076266 containerd[1549]: time="2025-09-08T23:52:08.076217179Z" level=info msg="received exit event container_id:\"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" id:\"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" pid:3170 exited_at:{seconds:1757375528 nanos:71436561}" Sep 8 23:52:08.113165 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 8 23:52:08.681947 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a-rootfs.mount: Deactivated successfully. Sep 8 23:52:08.960694 containerd[1549]: time="2025-09-08T23:52:08.959685591Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 8 23:52:08.974261 containerd[1549]: time="2025-09-08T23:52:08.974205999Z" level=info msg="Container bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:08.978441 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount580834987.mount: Deactivated successfully. Sep 8 23:52:08.997460 containerd[1549]: time="2025-09-08T23:52:08.997404111Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\"" Sep 8 23:52:08.997919 containerd[1549]: time="2025-09-08T23:52:08.997895327Z" level=info msg="StartContainer for \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\"" Sep 8 23:52:08.999381 containerd[1549]: time="2025-09-08T23:52:08.999354573Z" level=info msg="connecting to shim bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" protocol=ttrpc version=3 Sep 8 23:52:09.025343 systemd[1]: Started cri-containerd-bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33.scope - libcontainer container bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33. Sep 8 23:52:09.059799 containerd[1549]: time="2025-09-08T23:52:09.059754170Z" level=info msg="StartContainer for \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" returns successfully" Sep 8 23:52:09.061707 systemd[1]: cri-containerd-bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33.scope: Deactivated successfully. Sep 8 23:52:09.063522 containerd[1549]: time="2025-09-08T23:52:09.063472591Z" level=info msg="received exit event container_id:\"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" id:\"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" pid:3218 exited_at:{seconds:1757375529 nanos:63255470}" Sep 8 23:52:09.063943 containerd[1549]: time="2025-09-08T23:52:09.063766647Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" id:\"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" pid:3218 exited_at:{seconds:1757375529 nanos:63255470}" Sep 8 23:52:09.085419 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33-rootfs.mount: Deactivated successfully. Sep 8 23:52:09.972425 containerd[1549]: time="2025-09-08T23:52:09.972388460Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 8 23:52:09.981062 containerd[1549]: time="2025-09-08T23:52:09.980944833Z" level=info msg="Container 5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:09.983704 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount659647673.mount: Deactivated successfully. Sep 8 23:52:10.000760 containerd[1549]: time="2025-09-08T23:52:10.000646546Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\"" Sep 8 23:52:10.001578 containerd[1549]: time="2025-09-08T23:52:10.001554513Z" level=info msg="StartContainer for \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\"" Sep 8 23:52:10.002457 containerd[1549]: time="2025-09-08T23:52:10.002426632Z" level=info msg="connecting to shim 5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" protocol=ttrpc version=3 Sep 8 23:52:10.020574 systemd[1]: Started cri-containerd-5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b.scope - libcontainer container 5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b. Sep 8 23:52:10.047069 systemd[1]: cri-containerd-5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b.scope: Deactivated successfully. Sep 8 23:52:10.049566 containerd[1549]: time="2025-09-08T23:52:10.049490721Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" id:\"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" pid:3257 exited_at:{seconds:1757375530 nanos:49267401}" Sep 8 23:52:10.049738 containerd[1549]: time="2025-09-08T23:52:10.049545251Z" level=info msg="received exit event container_id:\"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" id:\"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" pid:3257 exited_at:{seconds:1757375530 nanos:49267401}" Sep 8 23:52:10.051121 containerd[1549]: time="2025-09-08T23:52:10.051031481Z" level=info msg="StartContainer for \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" returns successfully" Sep 8 23:52:10.069440 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b-rootfs.mount: Deactivated successfully. Sep 8 23:52:10.381183 containerd[1549]: time="2025-09-08T23:52:10.381114305Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:52:10.382830 containerd[1549]: time="2025-09-08T23:52:10.382779246Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Sep 8 23:52:10.384124 containerd[1549]: time="2025-09-08T23:52:10.384087444Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 8 23:52:10.387266 containerd[1549]: time="2025-09-08T23:52:10.387218131Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 3.767644942s" Sep 8 23:52:10.387266 containerd[1549]: time="2025-09-08T23:52:10.387261299Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Sep 8 23:52:10.392472 containerd[1549]: time="2025-09-08T23:52:10.392432036Z" level=info msg="CreateContainer within sandbox \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 8 23:52:10.405521 containerd[1549]: time="2025-09-08T23:52:10.403678954Z" level=info msg="Container b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:10.411864 containerd[1549]: time="2025-09-08T23:52:10.411816429Z" level=info msg="CreateContainer within sandbox \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\"" Sep 8 23:52:10.412833 containerd[1549]: time="2025-09-08T23:52:10.412797407Z" level=info msg="StartContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\"" Sep 8 23:52:10.415211 containerd[1549]: time="2025-09-08T23:52:10.415162996Z" level=info msg="connecting to shim b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3" address="unix:///run/containerd/s/70df2fef14da57b9fe4be1dbec34ca5da02dc898bc069801baa4fa072b61b4fb" protocol=ttrpc version=3 Sep 8 23:52:10.442306 systemd[1]: Started cri-containerd-b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3.scope - libcontainer container b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3. Sep 8 23:52:10.467970 containerd[1549]: time="2025-09-08T23:52:10.467929919Z" level=info msg="StartContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" returns successfully" Sep 8 23:52:10.792823 systemd[1]: Started sshd@7-10.0.0.130:22-10.0.0.1:57804.service - OpenSSH per-connection server daemon (10.0.0.1:57804). Sep 8 23:52:10.863673 sshd[3328]: Accepted publickey for core from 10.0.0.1 port 57804 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:10.865052 sshd-session[3328]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:10.874324 systemd-logind[1524]: New session 8 of user core. Sep 8 23:52:10.881254 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 8 23:52:10.990381 containerd[1549]: time="2025-09-08T23:52:10.990306394Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 8 23:52:11.017215 containerd[1549]: time="2025-09-08T23:52:11.017162232Z" level=info msg="Container c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:11.020116 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2221971025.mount: Deactivated successfully. Sep 8 23:52:11.028523 containerd[1549]: time="2025-09-08T23:52:11.028462284Z" level=info msg="CreateContainer within sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\"" Sep 8 23:52:11.029511 containerd[1549]: time="2025-09-08T23:52:11.029483582Z" level=info msg="StartContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\"" Sep 8 23:52:11.031308 containerd[1549]: time="2025-09-08T23:52:11.031257891Z" level=info msg="connecting to shim c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d" address="unix:///run/containerd/s/dec743971f901719fc9ec16e8c0f33c08274fad45ee578a37c295710c57f48fd" protocol=ttrpc version=3 Sep 8 23:52:11.063076 kubelet[2679]: I0908 23:52:11.062642 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-2lblt" podStartSLOduration=1.808742452 podStartE2EDuration="25.062622123s" podCreationTimestamp="2025-09-08 23:51:46 +0000 UTC" firstStartedPulling="2025-09-08 23:51:47.136464066 +0000 UTC m=+5.402363382" lastFinishedPulling="2025-09-08 23:52:10.390343737 +0000 UTC m=+28.656243053" observedRunningTime="2025-09-08 23:52:11.061561298 +0000 UTC m=+29.327460614" watchObservedRunningTime="2025-09-08 23:52:11.062622123 +0000 UTC m=+29.328521439" Sep 8 23:52:11.076239 systemd[1]: Started cri-containerd-c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d.scope - libcontainer container c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d. Sep 8 23:52:11.121158 sshd[3331]: Connection closed by 10.0.0.1 port 57804 Sep 8 23:52:11.120995 sshd-session[3328]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:11.127262 systemd[1]: sshd@7-10.0.0.130:22-10.0.0.1:57804.service: Deactivated successfully. Sep 8 23:52:11.130106 systemd[1]: session-8.scope: Deactivated successfully. Sep 8 23:52:11.131430 systemd-logind[1524]: Session 8 logged out. Waiting for processes to exit. Sep 8 23:52:11.134424 systemd-logind[1524]: Removed session 8. Sep 8 23:52:11.151246 containerd[1549]: time="2025-09-08T23:52:11.151156848Z" level=info msg="StartContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" returns successfully" Sep 8 23:52:11.250147 containerd[1549]: time="2025-09-08T23:52:11.250056622Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" id:\"4680e94529ede24371960442a4c2e57661182abc5541902b52f9f01271b4af89\" pid:3397 exited_at:{seconds:1757375531 nanos:249726684}" Sep 8 23:52:11.324389 kubelet[2679]: I0908 23:52:11.324282 2679 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 8 23:52:11.355844 systemd[1]: Created slice kubepods-burstable-pod606f504c_0d21_41fc_9747_3e3a7f084fb0.slice - libcontainer container kubepods-burstable-pod606f504c_0d21_41fc_9747_3e3a7f084fb0.slice. Sep 8 23:52:11.361958 systemd[1]: Created slice kubepods-burstable-poda0d4ea3c_8900_4159_8f81_0edca49f6665.slice - libcontainer container kubepods-burstable-poda0d4ea3c_8900_4159_8f81_0edca49f6665.slice. Sep 8 23:52:11.511458 kubelet[2679]: I0908 23:52:11.511406 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/606f504c-0d21-41fc-9747-3e3a7f084fb0-config-volume\") pod \"coredns-7c65d6cfc9-ntqt7\" (UID: \"606f504c-0d21-41fc-9747-3e3a7f084fb0\") " pod="kube-system/coredns-7c65d6cfc9-ntqt7" Sep 8 23:52:11.511835 kubelet[2679]: I0908 23:52:11.511728 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrb4m\" (UniqueName: \"kubernetes.io/projected/606f504c-0d21-41fc-9747-3e3a7f084fb0-kube-api-access-hrb4m\") pod \"coredns-7c65d6cfc9-ntqt7\" (UID: \"606f504c-0d21-41fc-9747-3e3a7f084fb0\") " pod="kube-system/coredns-7c65d6cfc9-ntqt7" Sep 8 23:52:11.511835 kubelet[2679]: I0908 23:52:11.511761 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d4ea3c-8900-4159-8f81-0edca49f6665-config-volume\") pod \"coredns-7c65d6cfc9-6s7qz\" (UID: \"a0d4ea3c-8900-4159-8f81-0edca49f6665\") " pod="kube-system/coredns-7c65d6cfc9-6s7qz" Sep 8 23:52:11.511835 kubelet[2679]: I0908 23:52:11.511874 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdkn8\" (UniqueName: \"kubernetes.io/projected/a0d4ea3c-8900-4159-8f81-0edca49f6665-kube-api-access-xdkn8\") pod \"coredns-7c65d6cfc9-6s7qz\" (UID: \"a0d4ea3c-8900-4159-8f81-0edca49f6665\") " pod="kube-system/coredns-7c65d6cfc9-6s7qz" Sep 8 23:52:11.660001 containerd[1549]: time="2025-09-08T23:52:11.659890000Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ntqt7,Uid:606f504c-0d21-41fc-9747-3e3a7f084fb0,Namespace:kube-system,Attempt:0,}" Sep 8 23:52:11.665604 containerd[1549]: time="2025-09-08T23:52:11.665564269Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-6s7qz,Uid:a0d4ea3c-8900-4159-8f81-0edca49f6665,Namespace:kube-system,Attempt:0,}" Sep 8 23:52:14.054009 systemd-networkd[1441]: cilium_host: Link UP Sep 8 23:52:14.054153 systemd-networkd[1441]: cilium_net: Link UP Sep 8 23:52:14.054270 systemd-networkd[1441]: cilium_host: Gained carrier Sep 8 23:52:14.054374 systemd-networkd[1441]: cilium_net: Gained carrier Sep 8 23:52:14.136401 systemd-networkd[1441]: cilium_vxlan: Link UP Sep 8 23:52:14.136414 systemd-networkd[1441]: cilium_vxlan: Gained carrier Sep 8 23:52:14.262896 systemd-networkd[1441]: cilium_host: Gained IPv6LL Sep 8 23:52:14.396088 kernel: NET: Registered PF_ALG protocol family Sep 8 23:52:14.886225 systemd-networkd[1441]: cilium_net: Gained IPv6LL Sep 8 23:52:14.958255 systemd-networkd[1441]: lxc_health: Link UP Sep 8 23:52:14.966529 systemd-networkd[1441]: lxc_health: Gained carrier Sep 8 23:52:15.223999 systemd-networkd[1441]: lxc327359cdf4ee: Link UP Sep 8 23:52:15.238184 kernel: eth0: renamed from tmp21731 Sep 8 23:52:15.239063 kernel: eth0: renamed from tmpf6d5a Sep 8 23:52:15.241748 systemd-networkd[1441]: lxce21bbfb23017: Link UP Sep 8 23:52:15.243589 systemd-networkd[1441]: lxc327359cdf4ee: Gained carrier Sep 8 23:52:15.243774 systemd-networkd[1441]: lxce21bbfb23017: Gained carrier Sep 8 23:52:15.590175 systemd-networkd[1441]: cilium_vxlan: Gained IPv6LL Sep 8 23:52:16.102225 systemd-networkd[1441]: lxc_health: Gained IPv6LL Sep 8 23:52:16.147132 systemd[1]: Started sshd@8-10.0.0.130:22-10.0.0.1:57806.service - OpenSSH per-connection server daemon (10.0.0.1:57806). Sep 8 23:52:16.212627 sshd[3869]: Accepted publickey for core from 10.0.0.1 port 57806 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:16.214694 sshd-session[3869]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:16.219686 systemd-logind[1524]: New session 9 of user core. Sep 8 23:52:16.230269 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 8 23:52:16.352325 sshd[3872]: Connection closed by 10.0.0.1 port 57806 Sep 8 23:52:16.351270 sshd-session[3869]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:16.354923 systemd-logind[1524]: Session 9 logged out. Waiting for processes to exit. Sep 8 23:52:16.355477 systemd[1]: session-9.scope: Deactivated successfully. Sep 8 23:52:16.358157 systemd-networkd[1441]: lxce21bbfb23017: Gained IPv6LL Sep 8 23:52:16.358722 systemd[1]: sshd@8-10.0.0.130:22-10.0.0.1:57806.service: Deactivated successfully. Sep 8 23:52:16.362811 systemd-logind[1524]: Removed session 9. Sep 8 23:52:16.868945 kubelet[2679]: I0908 23:52:16.865738 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-zb7sk" podStartSLOduration=11.184758557 podStartE2EDuration="30.865715036s" podCreationTimestamp="2025-09-08 23:51:46 +0000 UTC" firstStartedPulling="2025-09-08 23:51:46.938368177 +0000 UTC m=+5.204267453" lastFinishedPulling="2025-09-08 23:52:06.619324616 +0000 UTC m=+24.885223932" observedRunningTime="2025-09-08 23:52:12.015870888 +0000 UTC m=+30.281770204" watchObservedRunningTime="2025-09-08 23:52:16.865715036 +0000 UTC m=+35.131614312" Sep 8 23:52:17.190207 systemd-networkd[1441]: lxc327359cdf4ee: Gained IPv6LL Sep 8 23:52:18.871420 containerd[1549]: time="2025-09-08T23:52:18.870580995Z" level=info msg="connecting to shim 217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda" address="unix:///run/containerd/s/d1b304b5984570ff1c8d9a6b8f17cf3b148d47e33265e623502849cb42064389" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:52:18.872198 containerd[1549]: time="2025-09-08T23:52:18.872152691Z" level=info msg="connecting to shim f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed" address="unix:///run/containerd/s/8a9ca4f1b26673bb96b079e18758805822acc9208932997f77cd4aaa5e69de93" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:52:18.896234 systemd[1]: Started cri-containerd-f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed.scope - libcontainer container f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed. Sep 8 23:52:18.899601 systemd[1]: Started cri-containerd-217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda.scope - libcontainer container 217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda. Sep 8 23:52:18.908231 systemd-resolved[1355]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 8 23:52:18.914794 systemd-resolved[1355]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 8 23:52:18.934366 containerd[1549]: time="2025-09-08T23:52:18.934323717Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-6s7qz,Uid:a0d4ea3c-8900-4159-8f81-0edca49f6665,Namespace:kube-system,Attempt:0,} returns sandbox id \"f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed\"" Sep 8 23:52:18.937949 containerd[1549]: time="2025-09-08T23:52:18.937915811Z" level=info msg="CreateContainer within sandbox \"f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 8 23:52:18.950083 containerd[1549]: time="2025-09-08T23:52:18.949715193Z" level=info msg="Container b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:18.951278 containerd[1549]: time="2025-09-08T23:52:18.951243443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-ntqt7,Uid:606f504c-0d21-41fc-9747-3e3a7f084fb0,Namespace:kube-system,Attempt:0,} returns sandbox id \"217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda\"" Sep 8 23:52:18.955748 containerd[1549]: time="2025-09-08T23:52:18.955458422Z" level=info msg="CreateContainer within sandbox \"217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 8 23:52:18.956271 containerd[1549]: time="2025-09-08T23:52:18.956234249Z" level=info msg="CreateContainer within sandbox \"f6d5ac25b49b494eb499947a05ca8241f0bb672c2da44e120849a0ba0b3bd0ed\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec\"" Sep 8 23:52:18.956792 containerd[1549]: time="2025-09-08T23:52:18.956766682Z" level=info msg="StartContainer for \"b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec\"" Sep 8 23:52:18.957944 containerd[1549]: time="2025-09-08T23:52:18.957896998Z" level=info msg="connecting to shim b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec" address="unix:///run/containerd/s/8a9ca4f1b26673bb96b079e18758805822acc9208932997f77cd4aaa5e69de93" protocol=ttrpc version=3 Sep 8 23:52:18.970322 containerd[1549]: time="2025-09-08T23:52:18.969695980Z" level=info msg="Container 42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:52:18.976121 containerd[1549]: time="2025-09-08T23:52:18.976081137Z" level=info msg="CreateContainer within sandbox \"217315309df4ffd8f4e2cc4b3e7a5b9e73ea92874208e99507231d110dbebfda\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755\"" Sep 8 23:52:18.976746 containerd[1549]: time="2025-09-08T23:52:18.976725986Z" level=info msg="StartContainer for \"42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755\"" Sep 8 23:52:18.977705 containerd[1549]: time="2025-09-08T23:52:18.977680157Z" level=info msg="connecting to shim 42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755" address="unix:///run/containerd/s/d1b304b5984570ff1c8d9a6b8f17cf3b148d47e33265e623502849cb42064389" protocol=ttrpc version=3 Sep 8 23:52:18.978745 systemd[1]: Started cri-containerd-b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec.scope - libcontainer container b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec. Sep 8 23:52:18.999330 systemd[1]: Started cri-containerd-42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755.scope - libcontainer container 42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755. Sep 8 23:52:19.012764 containerd[1549]: time="2025-09-08T23:52:19.012704044Z" level=info msg="StartContainer for \"b55c513b92dd133d6d90ac4a255ac0c4b80c0c98c51e3f725ec720d204af5aec\" returns successfully" Sep 8 23:52:19.041540 containerd[1549]: time="2025-09-08T23:52:19.041387311Z" level=info msg="StartContainer for \"42c234cad852ab9250f47f8ac612c869e67140c03b2b2b3367bd91b9de1d5755\" returns successfully" Sep 8 23:52:19.851490 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount883120007.mount: Deactivated successfully. Sep 8 23:52:20.058550 kubelet[2679]: I0908 23:52:20.058389 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-6s7qz" podStartSLOduration=34.058372095 podStartE2EDuration="34.058372095s" podCreationTimestamp="2025-09-08 23:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:52:19.054883592 +0000 UTC m=+37.320782908" watchObservedRunningTime="2025-09-08 23:52:20.058372095 +0000 UTC m=+38.324271411" Sep 8 23:52:20.071428 kubelet[2679]: I0908 23:52:20.071369 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-ntqt7" podStartSLOduration=34.071350298 podStartE2EDuration="34.071350298s" podCreationTimestamp="2025-09-08 23:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:52:20.058732302 +0000 UTC m=+38.324631618" watchObservedRunningTime="2025-09-08 23:52:20.071350298 +0000 UTC m=+38.337249614" Sep 8 23:52:21.366417 systemd[1]: Started sshd@9-10.0.0.130:22-10.0.0.1:55714.service - OpenSSH per-connection server daemon (10.0.0.1:55714). Sep 8 23:52:21.432173 sshd[4073]: Accepted publickey for core from 10.0.0.1 port 55714 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:21.433809 sshd-session[4073]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:21.439112 systemd-logind[1524]: New session 10 of user core. Sep 8 23:52:21.455271 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 8 23:52:21.571489 sshd[4076]: Connection closed by 10.0.0.1 port 55714 Sep 8 23:52:21.571843 sshd-session[4073]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:21.575302 systemd[1]: sshd@9-10.0.0.130:22-10.0.0.1:55714.service: Deactivated successfully. Sep 8 23:52:21.577360 systemd[1]: session-10.scope: Deactivated successfully. Sep 8 23:52:21.578229 systemd-logind[1524]: Session 10 logged out. Waiting for processes to exit. Sep 8 23:52:21.579187 systemd-logind[1524]: Removed session 10. Sep 8 23:52:26.591841 systemd[1]: Started sshd@10-10.0.0.130:22-10.0.0.1:55716.service - OpenSSH per-connection server daemon (10.0.0.1:55716). Sep 8 23:52:26.657674 sshd[4092]: Accepted publickey for core from 10.0.0.1 port 55716 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:26.662168 sshd-session[4092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:26.669674 systemd-logind[1524]: New session 11 of user core. Sep 8 23:52:26.675252 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 8 23:52:26.801176 sshd[4095]: Connection closed by 10.0.0.1 port 55716 Sep 8 23:52:26.802675 sshd-session[4092]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:26.811231 systemd[1]: sshd@10-10.0.0.130:22-10.0.0.1:55716.service: Deactivated successfully. Sep 8 23:52:26.812756 systemd[1]: session-11.scope: Deactivated successfully. Sep 8 23:52:26.814544 systemd-logind[1524]: Session 11 logged out. Waiting for processes to exit. Sep 8 23:52:26.816461 systemd[1]: Started sshd@11-10.0.0.130:22-10.0.0.1:55720.service - OpenSSH per-connection server daemon (10.0.0.1:55720). Sep 8 23:52:26.822350 systemd-logind[1524]: Removed session 11. Sep 8 23:52:26.876020 sshd[4109]: Accepted publickey for core from 10.0.0.1 port 55720 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:26.878220 sshd-session[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:26.883197 systemd-logind[1524]: New session 12 of user core. Sep 8 23:52:26.896274 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 8 23:52:27.077157 sshd[4112]: Connection closed by 10.0.0.1 port 55720 Sep 8 23:52:27.078557 sshd-session[4109]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:27.088956 systemd[1]: sshd@11-10.0.0.130:22-10.0.0.1:55720.service: Deactivated successfully. Sep 8 23:52:27.096269 systemd[1]: session-12.scope: Deactivated successfully. Sep 8 23:52:27.098999 systemd-logind[1524]: Session 12 logged out. Waiting for processes to exit. Sep 8 23:52:27.105170 systemd[1]: Started sshd@12-10.0.0.130:22-10.0.0.1:55732.service - OpenSSH per-connection server daemon (10.0.0.1:55732). Sep 8 23:52:27.109056 systemd-logind[1524]: Removed session 12. Sep 8 23:52:27.176646 sshd[4125]: Accepted publickey for core from 10.0.0.1 port 55732 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:27.177902 sshd-session[4125]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:27.182432 systemd-logind[1524]: New session 13 of user core. Sep 8 23:52:27.189232 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 8 23:52:27.308108 sshd[4128]: Connection closed by 10.0.0.1 port 55732 Sep 8 23:52:27.308676 sshd-session[4125]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:27.312065 systemd[1]: sshd@12-10.0.0.130:22-10.0.0.1:55732.service: Deactivated successfully. Sep 8 23:52:27.313644 systemd[1]: session-13.scope: Deactivated successfully. Sep 8 23:52:27.314314 systemd-logind[1524]: Session 13 logged out. Waiting for processes to exit. Sep 8 23:52:27.315638 systemd-logind[1524]: Removed session 13. Sep 8 23:52:32.326629 systemd[1]: Started sshd@13-10.0.0.130:22-10.0.0.1:54986.service - OpenSSH per-connection server daemon (10.0.0.1:54986). Sep 8 23:52:32.391521 sshd[4143]: Accepted publickey for core from 10.0.0.1 port 54986 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:32.392943 sshd-session[4143]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:32.397322 systemd-logind[1524]: New session 14 of user core. Sep 8 23:52:32.407254 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 8 23:52:32.525860 sshd[4146]: Connection closed by 10.0.0.1 port 54986 Sep 8 23:52:32.526240 sshd-session[4143]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:32.529777 systemd[1]: sshd@13-10.0.0.130:22-10.0.0.1:54986.service: Deactivated successfully. Sep 8 23:52:32.531547 systemd[1]: session-14.scope: Deactivated successfully. Sep 8 23:52:32.533576 systemd-logind[1524]: Session 14 logged out. Waiting for processes to exit. Sep 8 23:52:32.534526 systemd-logind[1524]: Removed session 14. Sep 8 23:52:37.541610 systemd[1]: Started sshd@14-10.0.0.130:22-10.0.0.1:54990.service - OpenSSH per-connection server daemon (10.0.0.1:54990). Sep 8 23:52:37.588707 sshd[4160]: Accepted publickey for core from 10.0.0.1 port 54990 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:37.590206 sshd-session[4160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:37.594023 systemd-logind[1524]: New session 15 of user core. Sep 8 23:52:37.604246 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 8 23:52:37.718921 sshd[4163]: Connection closed by 10.0.0.1 port 54990 Sep 8 23:52:37.719466 sshd-session[4160]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:37.740546 systemd[1]: sshd@14-10.0.0.130:22-10.0.0.1:54990.service: Deactivated successfully. Sep 8 23:52:37.742758 systemd[1]: session-15.scope: Deactivated successfully. Sep 8 23:52:37.743744 systemd-logind[1524]: Session 15 logged out. Waiting for processes to exit. Sep 8 23:52:37.746172 systemd[1]: Started sshd@15-10.0.0.130:22-10.0.0.1:55006.service - OpenSSH per-connection server daemon (10.0.0.1:55006). Sep 8 23:52:37.747632 systemd-logind[1524]: Removed session 15. Sep 8 23:52:37.807303 sshd[4177]: Accepted publickey for core from 10.0.0.1 port 55006 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:37.808827 sshd-session[4177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:37.813197 systemd-logind[1524]: New session 16 of user core. Sep 8 23:52:37.824257 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 8 23:52:38.036404 sshd[4180]: Connection closed by 10.0.0.1 port 55006 Sep 8 23:52:38.037463 sshd-session[4177]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:38.049802 systemd[1]: sshd@15-10.0.0.130:22-10.0.0.1:55006.service: Deactivated successfully. Sep 8 23:52:38.052201 systemd[1]: session-16.scope: Deactivated successfully. Sep 8 23:52:38.053758 systemd-logind[1524]: Session 16 logged out. Waiting for processes to exit. Sep 8 23:52:38.055802 systemd-logind[1524]: Removed session 16. Sep 8 23:52:38.057405 systemd[1]: Started sshd@16-10.0.0.130:22-10.0.0.1:55010.service - OpenSSH per-connection server daemon (10.0.0.1:55010). Sep 8 23:52:38.118117 sshd[4191]: Accepted publickey for core from 10.0.0.1 port 55010 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:38.119464 sshd-session[4191]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:38.123421 systemd-logind[1524]: New session 17 of user core. Sep 8 23:52:38.134284 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 8 23:52:39.415183 sshd[4194]: Connection closed by 10.0.0.1 port 55010 Sep 8 23:52:39.416614 sshd-session[4191]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:39.429220 systemd[1]: sshd@16-10.0.0.130:22-10.0.0.1:55010.service: Deactivated successfully. Sep 8 23:52:39.433289 systemd[1]: session-17.scope: Deactivated successfully. Sep 8 23:52:39.434487 systemd-logind[1524]: Session 17 logged out. Waiting for processes to exit. Sep 8 23:52:39.437413 systemd[1]: Started sshd@17-10.0.0.130:22-10.0.0.1:55026.service - OpenSSH per-connection server daemon (10.0.0.1:55026). Sep 8 23:52:39.439260 systemd-logind[1524]: Removed session 17. Sep 8 23:52:39.505592 sshd[4214]: Accepted publickey for core from 10.0.0.1 port 55026 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:39.506971 sshd-session[4214]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:39.511643 systemd-logind[1524]: New session 18 of user core. Sep 8 23:52:39.519269 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 8 23:52:39.745367 sshd[4217]: Connection closed by 10.0.0.1 port 55026 Sep 8 23:52:39.745663 sshd-session[4214]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:39.762029 systemd[1]: sshd@17-10.0.0.130:22-10.0.0.1:55026.service: Deactivated successfully. Sep 8 23:52:39.765751 systemd[1]: session-18.scope: Deactivated successfully. Sep 8 23:52:39.766756 systemd-logind[1524]: Session 18 logged out. Waiting for processes to exit. Sep 8 23:52:39.771613 systemd[1]: Started sshd@18-10.0.0.130:22-10.0.0.1:55032.service - OpenSSH per-connection server daemon (10.0.0.1:55032). Sep 8 23:52:39.772270 systemd-logind[1524]: Removed session 18. Sep 8 23:52:39.828213 sshd[4229]: Accepted publickey for core from 10.0.0.1 port 55032 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:39.829649 sshd-session[4229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:39.834782 systemd-logind[1524]: New session 19 of user core. Sep 8 23:52:39.844306 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 8 23:52:39.959413 sshd[4232]: Connection closed by 10.0.0.1 port 55032 Sep 8 23:52:39.959755 sshd-session[4229]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:39.963418 systemd[1]: sshd@18-10.0.0.130:22-10.0.0.1:55032.service: Deactivated successfully. Sep 8 23:52:39.965337 systemd[1]: session-19.scope: Deactivated successfully. Sep 8 23:52:39.966137 systemd-logind[1524]: Session 19 logged out. Waiting for processes to exit. Sep 8 23:52:39.967268 systemd-logind[1524]: Removed session 19. Sep 8 23:52:44.446611 kernel: hrtimer: interrupt took 7058399 ns Sep 8 23:52:44.977445 systemd[1]: Started sshd@19-10.0.0.130:22-10.0.0.1:60120.service - OpenSSH per-connection server daemon (10.0.0.1:60120). Sep 8 23:52:45.041767 sshd[4251]: Accepted publickey for core from 10.0.0.1 port 60120 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:45.040829 sshd-session[4251]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:45.051548 systemd-logind[1524]: New session 20 of user core. Sep 8 23:52:45.060682 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 8 23:52:45.191350 sshd[4254]: Connection closed by 10.0.0.1 port 60120 Sep 8 23:52:45.192011 sshd-session[4251]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:45.195123 systemd[1]: sshd@19-10.0.0.130:22-10.0.0.1:60120.service: Deactivated successfully. Sep 8 23:52:45.202992 systemd[1]: session-20.scope: Deactivated successfully. Sep 8 23:52:45.203878 systemd-logind[1524]: Session 20 logged out. Waiting for processes to exit. Sep 8 23:52:45.205782 systemd-logind[1524]: Removed session 20. Sep 8 23:52:50.202901 systemd[1]: Started sshd@20-10.0.0.130:22-10.0.0.1:40606.service - OpenSSH per-connection server daemon (10.0.0.1:40606). Sep 8 23:52:50.259312 sshd[4269]: Accepted publickey for core from 10.0.0.1 port 40606 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:50.261387 sshd-session[4269]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:50.270164 systemd-logind[1524]: New session 21 of user core. Sep 8 23:52:50.279453 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 8 23:52:50.396155 sshd[4272]: Connection closed by 10.0.0.1 port 40606 Sep 8 23:52:50.396577 sshd-session[4269]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:50.401705 systemd[1]: sshd@20-10.0.0.130:22-10.0.0.1:40606.service: Deactivated successfully. Sep 8 23:52:50.403291 systemd[1]: session-21.scope: Deactivated successfully. Sep 8 23:52:50.404026 systemd-logind[1524]: Session 21 logged out. Waiting for processes to exit. Sep 8 23:52:50.406942 systemd-logind[1524]: Removed session 21. Sep 8 23:52:55.408453 systemd[1]: Started sshd@21-10.0.0.130:22-10.0.0.1:40622.service - OpenSSH per-connection server daemon (10.0.0.1:40622). Sep 8 23:52:55.474818 sshd[4286]: Accepted publickey for core from 10.0.0.1 port 40622 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:55.476233 sshd-session[4286]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:55.483593 systemd-logind[1524]: New session 22 of user core. Sep 8 23:52:55.492247 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 8 23:52:55.610077 sshd[4289]: Connection closed by 10.0.0.1 port 40622 Sep 8 23:52:55.611642 sshd-session[4286]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:55.628888 systemd[1]: sshd@21-10.0.0.130:22-10.0.0.1:40622.service: Deactivated successfully. Sep 8 23:52:55.630806 systemd[1]: session-22.scope: Deactivated successfully. Sep 8 23:52:55.632097 systemd-logind[1524]: Session 22 logged out. Waiting for processes to exit. Sep 8 23:52:55.634166 systemd[1]: Started sshd@22-10.0.0.130:22-10.0.0.1:40638.service - OpenSSH per-connection server daemon (10.0.0.1:40638). Sep 8 23:52:55.635503 systemd-logind[1524]: Removed session 22. Sep 8 23:52:55.720391 sshd[4302]: Accepted publickey for core from 10.0.0.1 port 40638 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:55.721877 sshd-session[4302]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:55.728859 systemd-logind[1524]: New session 23 of user core. Sep 8 23:52:55.748268 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 8 23:52:57.819339 containerd[1549]: time="2025-09-08T23:52:57.819285460Z" level=info msg="StopContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" with timeout 30 (s)" Sep 8 23:52:57.820460 containerd[1549]: time="2025-09-08T23:52:57.819729158Z" level=info msg="Stop container \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" with signal terminated" Sep 8 23:52:57.831181 systemd[1]: cri-containerd-b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3.scope: Deactivated successfully. Sep 8 23:52:57.835818 containerd[1549]: time="2025-09-08T23:52:57.834017244Z" level=info msg="received exit event container_id:\"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" id:\"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" pid:3303 exited_at:{seconds:1757375577 nanos:833383436}" Sep 8 23:52:57.835818 containerd[1549]: time="2025-09-08T23:52:57.834156077Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" id:\"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" pid:3303 exited_at:{seconds:1757375577 nanos:833383436}" Sep 8 23:52:57.855737 containerd[1549]: time="2025-09-08T23:52:57.855694602Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" id:\"a7a037f80bfa4814e721e2aafa2cbfbcfdd8e3cb5a2af2bdfc134a9e98abe0dd\" pid:4331 exited_at:{seconds:1757375577 nanos:855436655}" Sep 8 23:52:57.858870 containerd[1549]: time="2025-09-08T23:52:57.858831885Z" level=info msg="StopContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" with timeout 2 (s)" Sep 8 23:52:57.859391 containerd[1549]: time="2025-09-08T23:52:57.859364979Z" level=info msg="Stop container \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" with signal terminated" Sep 8 23:52:57.863683 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3-rootfs.mount: Deactivated successfully. Sep 8 23:52:57.866960 systemd-networkd[1441]: lxc_health: Link DOWN Sep 8 23:52:57.866966 systemd-networkd[1441]: lxc_health: Lost carrier Sep 8 23:52:57.873073 containerd[1549]: time="2025-09-08T23:52:57.872947981Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 8 23:52:57.883819 systemd[1]: cri-containerd-c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d.scope: Deactivated successfully. Sep 8 23:52:57.884161 systemd[1]: cri-containerd-c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d.scope: Consumed 6.238s CPU time, 124.4M memory peak, 132K read from disk, 12.9M written to disk. Sep 8 23:52:57.885487 containerd[1549]: time="2025-09-08T23:52:57.885223968Z" level=info msg="received exit event container_id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" pid:3364 exited_at:{seconds:1757375577 nanos:884880665}" Sep 8 23:52:57.885487 containerd[1549]: time="2025-09-08T23:52:57.885297444Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" id:\"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" pid:3364 exited_at:{seconds:1757375577 nanos:884880665}" Sep 8 23:52:57.898371 containerd[1549]: time="2025-09-08T23:52:57.897959012Z" level=info msg="StopContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" returns successfully" Sep 8 23:52:57.901534 containerd[1549]: time="2025-09-08T23:52:57.901457637Z" level=info msg="StopPodSandbox for \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\"" Sep 8 23:52:57.908331 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d-rootfs.mount: Deactivated successfully. Sep 8 23:52:57.911542 containerd[1549]: time="2025-09-08T23:52:57.911468337Z" level=info msg="Container to stop \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:57.918912 systemd[1]: cri-containerd-4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb.scope: Deactivated successfully. Sep 8 23:52:57.920316 containerd[1549]: time="2025-09-08T23:52:57.920238699Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" id:\"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" pid:2919 exit_status:137 exited_at:{seconds:1757375577 nanos:919636649}" Sep 8 23:52:57.950072 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb-rootfs.mount: Deactivated successfully. Sep 8 23:52:58.020916 containerd[1549]: time="2025-09-08T23:52:58.020851736Z" level=info msg="shim disconnected" id=4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb namespace=k8s.io Sep 8 23:52:58.020916 containerd[1549]: time="2025-09-08T23:52:58.020884135Z" level=warning msg="cleaning up after shim disconnected" id=4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb namespace=k8s.io Sep 8 23:52:58.020916 containerd[1549]: time="2025-09-08T23:52:58.020916253Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 8 23:52:58.022021 containerd[1549]: time="2025-09-08T23:52:58.021935566Z" level=info msg="StopContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" returns successfully" Sep 8 23:52:58.023053 containerd[1549]: time="2025-09-08T23:52:58.022984437Z" level=info msg="StopPodSandbox for \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\"" Sep 8 23:52:58.023148 containerd[1549]: time="2025-09-08T23:52:58.023061473Z" level=info msg="Container to stop \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:58.023148 containerd[1549]: time="2025-09-08T23:52:58.023075633Z" level=info msg="Container to stop \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:58.023148 containerd[1549]: time="2025-09-08T23:52:58.023084592Z" level=info msg="Container to stop \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:58.023148 containerd[1549]: time="2025-09-08T23:52:58.023093272Z" level=info msg="Container to stop \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:58.023148 containerd[1549]: time="2025-09-08T23:52:58.023101032Z" level=info msg="Container to stop \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 8 23:52:58.030339 systemd[1]: cri-containerd-a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206.scope: Deactivated successfully. Sep 8 23:52:58.044999 containerd[1549]: time="2025-09-08T23:52:58.043052021Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" id:\"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" pid:2824 exit_status:137 exited_at:{seconds:1757375578 nanos:31931420}" Sep 8 23:52:58.044930 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb-shm.mount: Deactivated successfully. Sep 8 23:52:58.045216 containerd[1549]: time="2025-09-08T23:52:58.045008130Z" level=info msg="received exit event sandbox_id:\"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" exit_status:137 exited_at:{seconds:1757375577 nanos:919636649}" Sep 8 23:52:58.045527 containerd[1549]: time="2025-09-08T23:52:58.045497307Z" level=info msg="TearDown network for sandbox \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" successfully" Sep 8 23:52:58.045527 containerd[1549]: time="2025-09-08T23:52:58.045522346Z" level=info msg="StopPodSandbox for \"4f3ad17e2ccab4a8d0897b6b5413b00c3b6e97e9b93fc732b38427c28d152ecb\" returns successfully" Sep 8 23:52:58.058160 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206-rootfs.mount: Deactivated successfully. Sep 8 23:52:58.079151 containerd[1549]: time="2025-09-08T23:52:58.079003744Z" level=info msg="shim disconnected" id=a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206 namespace=k8s.io Sep 8 23:52:58.079151 containerd[1549]: time="2025-09-08T23:52:58.079063781Z" level=warning msg="cleaning up after shim disconnected" id=a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206 namespace=k8s.io Sep 8 23:52:58.079151 containerd[1549]: time="2025-09-08T23:52:58.079098180Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 8 23:52:58.079927 containerd[1549]: time="2025-09-08T23:52:58.079438444Z" level=info msg="received exit event sandbox_id:\"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" exit_status:137 exited_at:{seconds:1757375578 nanos:31931420}" Sep 8 23:52:58.080532 containerd[1549]: time="2025-09-08T23:52:58.080482595Z" level=info msg="TearDown network for sandbox \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" successfully" Sep 8 23:52:58.080532 containerd[1549]: time="2025-09-08T23:52:58.080513234Z" level=info msg="StopPodSandbox for \"a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206\" returns successfully" Sep 8 23:52:58.126239 kubelet[2679]: I0908 23:52:58.126187 2679 scope.go:117] "RemoveContainer" containerID="b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3" Sep 8 23:52:58.129183 containerd[1549]: time="2025-09-08T23:52:58.129141006Z" level=info msg="RemoveContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\"" Sep 8 23:52:58.137830 containerd[1549]: time="2025-09-08T23:52:58.137793162Z" level=info msg="RemoveContainer for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" returns successfully" Sep 8 23:52:58.138183 kubelet[2679]: I0908 23:52:58.138152 2679 scope.go:117] "RemoveContainer" containerID="b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3" Sep 8 23:52:58.138475 containerd[1549]: time="2025-09-08T23:52:58.138438532Z" level=error msg="ContainerStatus for \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\": not found" Sep 8 23:52:58.141728 kubelet[2679]: E0908 23:52:58.141476 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\": not found" containerID="b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3" Sep 8 23:52:58.141728 kubelet[2679]: I0908 23:52:58.141537 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3"} err="failed to get container status \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\": rpc error: code = NotFound desc = an error occurred when try to find container \"b6e96dd4dd735f8cad9a8f31f90afe7b6bc335319b9ab18a331723aeaf17f3d3\": not found" Sep 8 23:52:58.141728 kubelet[2679]: I0908 23:52:58.141618 2679 scope.go:117] "RemoveContainer" containerID="c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d" Sep 8 23:52:58.143808 containerd[1549]: time="2025-09-08T23:52:58.143781843Z" level=info msg="RemoveContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\"" Sep 8 23:52:58.150644 containerd[1549]: time="2025-09-08T23:52:58.150595805Z" level=info msg="RemoveContainer for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" returns successfully" Sep 8 23:52:58.150881 kubelet[2679]: I0908 23:52:58.150850 2679 scope.go:117] "RemoveContainer" containerID="5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b" Sep 8 23:52:58.156185 containerd[1549]: time="2025-09-08T23:52:58.156097228Z" level=info msg="RemoveContainer for \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\"" Sep 8 23:52:58.160186 containerd[1549]: time="2025-09-08T23:52:58.160144799Z" level=info msg="RemoveContainer for \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" returns successfully" Sep 8 23:52:58.160393 kubelet[2679]: I0908 23:52:58.160348 2679 scope.go:117] "RemoveContainer" containerID="bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33" Sep 8 23:52:58.162763 containerd[1549]: time="2025-09-08T23:52:58.162736759Z" level=info msg="RemoveContainer for \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\"" Sep 8 23:52:58.166919 containerd[1549]: time="2025-09-08T23:52:58.166885445Z" level=info msg="RemoveContainer for \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" returns successfully" Sep 8 23:52:58.167225 kubelet[2679]: I0908 23:52:58.167100 2679 scope.go:117] "RemoveContainer" containerID="ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a" Sep 8 23:52:58.168740 containerd[1549]: time="2025-09-08T23:52:58.168714520Z" level=info msg="RemoveContainer for \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\"" Sep 8 23:52:58.172149 containerd[1549]: time="2025-09-08T23:52:58.172111321Z" level=info msg="RemoveContainer for \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" returns successfully" Sep 8 23:52:58.172354 kubelet[2679]: I0908 23:52:58.172325 2679 scope.go:117] "RemoveContainer" containerID="1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997" Sep 8 23:52:58.173749 containerd[1549]: time="2025-09-08T23:52:58.173691368Z" level=info msg="RemoveContainer for \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\"" Sep 8 23:52:58.177135 containerd[1549]: time="2025-09-08T23:52:58.177083969Z" level=info msg="RemoveContainer for \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" returns successfully" Sep 8 23:52:58.177345 kubelet[2679]: I0908 23:52:58.177321 2679 scope.go:117] "RemoveContainer" containerID="c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d" Sep 8 23:52:58.177650 containerd[1549]: time="2025-09-08T23:52:58.177584186Z" level=error msg="ContainerStatus for \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\": not found" Sep 8 23:52:58.177755 kubelet[2679]: E0908 23:52:58.177729 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\": not found" containerID="c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d" Sep 8 23:52:58.177795 kubelet[2679]: I0908 23:52:58.177765 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d"} err="failed to get container status \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\": rpc error: code = NotFound desc = an error occurred when try to find container \"c9d1e8d99960b1c717ae3aa96408af2a9ac3e93d36f81f01a6a464bd62ab564d\": not found" Sep 8 23:52:58.177795 kubelet[2679]: I0908 23:52:58.177787 2679 scope.go:117] "RemoveContainer" containerID="5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b" Sep 8 23:52:58.177995 containerd[1549]: time="2025-09-08T23:52:58.177964568Z" level=error msg="ContainerStatus for \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\": not found" Sep 8 23:52:58.178159 kubelet[2679]: E0908 23:52:58.178135 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\": not found" containerID="5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b" Sep 8 23:52:58.178217 kubelet[2679]: I0908 23:52:58.178162 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b"} err="failed to get container status \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\": rpc error: code = NotFound desc = an error occurred when try to find container \"5badee8b1a87546f4fac21dd177f6e2340fbaee6c47826f104664e913ebd683b\": not found" Sep 8 23:52:58.178217 kubelet[2679]: I0908 23:52:58.178179 2679 scope.go:117] "RemoveContainer" containerID="bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33" Sep 8 23:52:58.178393 containerd[1549]: time="2025-09-08T23:52:58.178361510Z" level=error msg="ContainerStatus for \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\": not found" Sep 8 23:52:58.178546 kubelet[2679]: E0908 23:52:58.178521 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\": not found" containerID="bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33" Sep 8 23:52:58.178635 kubelet[2679]: I0908 23:52:58.178610 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33"} err="failed to get container status \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\": rpc error: code = NotFound desc = an error occurred when try to find container \"bd3aca0e7f594967bff2aab34c3fa79e3313806dd665a93641af66230a0b0c33\": not found" Sep 8 23:52:58.178693 kubelet[2679]: I0908 23:52:58.178682 2679 scope.go:117] "RemoveContainer" containerID="ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a" Sep 8 23:52:58.178964 containerd[1549]: time="2025-09-08T23:52:58.178934163Z" level=error msg="ContainerStatus for \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\": not found" Sep 8 23:52:58.179118 kubelet[2679]: E0908 23:52:58.179077 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\": not found" containerID="ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a" Sep 8 23:52:58.179170 kubelet[2679]: I0908 23:52:58.179115 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a"} err="failed to get container status \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\": rpc error: code = NotFound desc = an error occurred when try to find container \"ec128a08d529a11c25872d0ffff6433f249afd9246c00de098fbd7b705fa6f4a\": not found" Sep 8 23:52:58.179170 kubelet[2679]: I0908 23:52:58.179135 2679 scope.go:117] "RemoveContainer" containerID="1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997" Sep 8 23:52:58.179352 containerd[1549]: time="2025-09-08T23:52:58.179319865Z" level=error msg="ContainerStatus for \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\": not found" Sep 8 23:52:58.179555 kubelet[2679]: E0908 23:52:58.179531 2679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\": not found" containerID="1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997" Sep 8 23:52:58.179602 kubelet[2679]: I0908 23:52:58.179558 2679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997"} err="failed to get container status \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\": rpc error: code = NotFound desc = an error occurred when try to find container \"1380a5d9736126b5c661474cdd508fab23d70694887ea276829e310c8e8a3997\": not found" Sep 8 23:52:58.208306 kubelet[2679]: I0908 23:52:58.208253 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-hostproc\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208306 kubelet[2679]: I0908 23:52:58.208301 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-net\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208354 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b4e7c630-7500-4e07-bd47-d442efe93d8e-cilium-config-path\") pod \"b4e7c630-7500-4e07-bd47-d442efe93d8e\" (UID: \"b4e7c630-7500-4e07-bd47-d442efe93d8e\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208372 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-kernel\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208390 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn9dh\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-kube-api-access-rn9dh\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208407 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-xtables-lock\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208422 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-bpf-maps\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208472 kubelet[2679]: I0908 23:52:58.208435 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-run\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208601 kubelet[2679]: I0908 23:52:58.208450 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-cgroup\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208601 kubelet[2679]: I0908 23:52:58.208465 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-hubble-tls\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208601 kubelet[2679]: I0908 23:52:58.208526 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-etc-cni-netd\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208601 kubelet[2679]: I0908 23:52:58.208565 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqwf4\" (UniqueName: \"kubernetes.io/projected/b4e7c630-7500-4e07-bd47-d442efe93d8e-kube-api-access-cqwf4\") pod \"b4e7c630-7500-4e07-bd47-d442efe93d8e\" (UID: \"b4e7c630-7500-4e07-bd47-d442efe93d8e\") " Sep 8 23:52:58.208601 kubelet[2679]: I0908 23:52:58.208583 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-lib-modules\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208698 kubelet[2679]: I0908 23:52:58.208598 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cni-path\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208698 kubelet[2679]: I0908 23:52:58.208622 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-config-path\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.208698 kubelet[2679]: I0908 23:52:58.208640 2679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d4d502e4-c0c6-4e11-98be-8a23553b9240-clustermesh-secrets\") pod \"d4d502e4-c0c6-4e11-98be-8a23553b9240\" (UID: \"d4d502e4-c0c6-4e11-98be-8a23553b9240\") " Sep 8 23:52:58.212673 kubelet[2679]: I0908 23:52:58.212624 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4e7c630-7500-4e07-bd47-d442efe93d8e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "b4e7c630-7500-4e07-bd47-d442efe93d8e" (UID: "b4e7c630-7500-4e07-bd47-d442efe93d8e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 8 23:52:58.212773 kubelet[2679]: I0908 23:52:58.212707 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.212773 kubelet[2679]: I0908 23:52:58.212727 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.213795 kubelet[2679]: I0908 23:52:58.213735 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.213795 kubelet[2679]: I0908 23:52:58.213843 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.213795 kubelet[2679]: I0908 23:52:58.213877 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.213795 kubelet[2679]: I0908 23:52:58.213895 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.214877 kubelet[2679]: I0908 23:52:58.214853 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-hostproc" (OuterVolumeSpecName: "hostproc") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.215958 kubelet[2679]: I0908 23:52:58.215178 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cni-path" (OuterVolumeSpecName: "cni-path") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.215958 kubelet[2679]: I0908 23:52:58.215210 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.215958 kubelet[2679]: I0908 23:52:58.214532 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 8 23:52:58.216796 kubelet[2679]: I0908 23:52:58.216755 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e7c630-7500-4e07-bd47-d442efe93d8e-kube-api-access-cqwf4" (OuterVolumeSpecName: "kube-api-access-cqwf4") pod "b4e7c630-7500-4e07-bd47-d442efe93d8e" (UID: "b4e7c630-7500-4e07-bd47-d442efe93d8e"). InnerVolumeSpecName "kube-api-access-cqwf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 8 23:52:58.217063 kubelet[2679]: I0908 23:52:58.216974 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d502e4-c0c6-4e11-98be-8a23553b9240-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 8 23:52:58.217063 kubelet[2679]: I0908 23:52:58.217014 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-kube-api-access-rn9dh" (OuterVolumeSpecName: "kube-api-access-rn9dh") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "kube-api-access-rn9dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 8 23:52:58.219645 kubelet[2679]: I0908 23:52:58.219530 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 8 23:52:58.219645 kubelet[2679]: I0908 23:52:58.219593 2679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "d4d502e4-c0c6-4e11-98be-8a23553b9240" (UID: "d4d502e4-c0c6-4e11-98be-8a23553b9240"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 8 23:52:58.310204 kubelet[2679]: I0908 23:52:58.310153 2679 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310204 kubelet[2679]: I0908 23:52:58.310196 2679 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310204 kubelet[2679]: I0908 23:52:58.310208 2679 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310204 kubelet[2679]: I0908 23:52:58.310218 2679 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d4d502e4-c0c6-4e11-98be-8a23553b9240-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310229 2679 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310237 2679 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310245 2679 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b4e7c630-7500-4e07-bd47-d442efe93d8e-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310253 2679 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310261 2679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn9dh\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-kube-api-access-rn9dh\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310268 2679 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310275 2679 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310398 kubelet[2679]: I0908 23:52:58.310282 2679 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310554 kubelet[2679]: I0908 23:52:58.310289 2679 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310554 kubelet[2679]: I0908 23:52:58.310296 2679 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d4d502e4-c0c6-4e11-98be-8a23553b9240-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310554 kubelet[2679]: I0908 23:52:58.310303 2679 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d4d502e4-c0c6-4e11-98be-8a23553b9240-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.310554 kubelet[2679]: I0908 23:52:58.310315 2679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqwf4\" (UniqueName: \"kubernetes.io/projected/b4e7c630-7500-4e07-bd47-d442efe93d8e-kube-api-access-cqwf4\") on node \"localhost\" DevicePath \"\"" Sep 8 23:52:58.432610 systemd[1]: Removed slice kubepods-besteffort-podb4e7c630_7500_4e07_bd47_d442efe93d8e.slice - libcontainer container kubepods-besteffort-podb4e7c630_7500_4e07_bd47_d442efe93d8e.slice. Sep 8 23:52:58.443952 systemd[1]: Removed slice kubepods-burstable-podd4d502e4_c0c6_4e11_98be_8a23553b9240.slice - libcontainer container kubepods-burstable-podd4d502e4_c0c6_4e11_98be_8a23553b9240.slice. Sep 8 23:52:58.444205 systemd[1]: kubepods-burstable-podd4d502e4_c0c6_4e11_98be_8a23553b9240.slice: Consumed 6.327s CPU time, 124.7M memory peak, 136K read from disk, 12.9M written to disk. Sep 8 23:52:58.863755 systemd[1]: var-lib-kubelet-pods-b4e7c630\x2d7500\x2d4e07\x2dbd47\x2dd442efe93d8e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dcqwf4.mount: Deactivated successfully. Sep 8 23:52:58.863852 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a301de77bcc801171d6a7e9a3c703712f183f4f878d28bb86a104382e083b206-shm.mount: Deactivated successfully. Sep 8 23:52:58.863910 systemd[1]: var-lib-kubelet-pods-d4d502e4\x2dc0c6\x2d4e11\x2d98be\x2d8a23553b9240-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2drn9dh.mount: Deactivated successfully. Sep 8 23:52:58.863960 systemd[1]: var-lib-kubelet-pods-d4d502e4\x2dc0c6\x2d4e11\x2d98be\x2d8a23553b9240-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 8 23:52:58.864011 systemd[1]: var-lib-kubelet-pods-d4d502e4\x2dc0c6\x2d4e11\x2d98be\x2d8a23553b9240-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 8 23:52:59.762355 sshd[4305]: Connection closed by 10.0.0.1 port 40638 Sep 8 23:52:59.762272 sshd-session[4302]: pam_unix(sshd:session): session closed for user core Sep 8 23:52:59.772295 systemd[1]: sshd@22-10.0.0.130:22-10.0.0.1:40638.service: Deactivated successfully. Sep 8 23:52:59.773848 systemd[1]: session-23.scope: Deactivated successfully. Sep 8 23:52:59.774079 systemd[1]: session-23.scope: Consumed 1.331s CPU time, 24.9M memory peak. Sep 8 23:52:59.774676 systemd-logind[1524]: Session 23 logged out. Waiting for processes to exit. Sep 8 23:52:59.776819 systemd[1]: Started sshd@23-10.0.0.130:22-10.0.0.1:40648.service - OpenSSH per-connection server daemon (10.0.0.1:40648). Sep 8 23:52:59.778297 systemd-logind[1524]: Removed session 23. Sep 8 23:52:59.837448 kubelet[2679]: I0908 23:52:59.837406 2679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e7c630-7500-4e07-bd47-d442efe93d8e" path="/var/lib/kubelet/pods/b4e7c630-7500-4e07-bd47-d442efe93d8e/volumes" Sep 8 23:52:59.837777 kubelet[2679]: I0908 23:52:59.837761 2679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" path="/var/lib/kubelet/pods/d4d502e4-c0c6-4e11-98be-8a23553b9240/volumes" Sep 8 23:52:59.839337 sshd[4458]: Accepted publickey for core from 10.0.0.1 port 40648 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:52:59.840940 sshd-session[4458]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:52:59.846547 systemd-logind[1524]: New session 24 of user core. Sep 8 23:52:59.855279 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 8 23:53:01.150924 sshd[4461]: Connection closed by 10.0.0.1 port 40648 Sep 8 23:53:01.150797 sshd-session[4458]: pam_unix(sshd:session): session closed for user core Sep 8 23:53:01.166470 systemd[1]: sshd@23-10.0.0.130:22-10.0.0.1:40648.service: Deactivated successfully. Sep 8 23:53:01.171596 systemd[1]: session-24.scope: Deactivated successfully. Sep 8 23:53:01.172085 systemd[1]: session-24.scope: Consumed 1.187s CPU time, 26.1M memory peak. Sep 8 23:53:01.173788 systemd-logind[1524]: Session 24 logged out. Waiting for processes to exit. Sep 8 23:53:01.175598 kubelet[2679]: E0908 23:53:01.175556 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="mount-bpf-fs" Sep 8 23:53:01.175598 kubelet[2679]: E0908 23:53:01.175590 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="clean-cilium-state" Sep 8 23:53:01.175598 kubelet[2679]: E0908 23:53:01.175598 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="cilium-agent" Sep 8 23:53:01.175598 kubelet[2679]: E0908 23:53:01.175605 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="apply-sysctl-overwrites" Sep 8 23:53:01.175922 kubelet[2679]: E0908 23:53:01.175611 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b4e7c630-7500-4e07-bd47-d442efe93d8e" containerName="cilium-operator" Sep 8 23:53:01.175922 kubelet[2679]: E0908 23:53:01.175618 2679 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="mount-cgroup" Sep 8 23:53:01.175922 kubelet[2679]: I0908 23:53:01.175641 2679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e7c630-7500-4e07-bd47-d442efe93d8e" containerName="cilium-operator" Sep 8 23:53:01.175922 kubelet[2679]: I0908 23:53:01.175648 2679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d502e4-c0c6-4e11-98be-8a23553b9240" containerName="cilium-agent" Sep 8 23:53:01.177319 systemd[1]: Started sshd@24-10.0.0.130:22-10.0.0.1:42404.service - OpenSSH per-connection server daemon (10.0.0.1:42404). Sep 8 23:53:01.181928 systemd-logind[1524]: Removed session 24. Sep 8 23:53:01.192684 systemd[1]: Created slice kubepods-burstable-pod0739c4b8_550f_4633_885a_eec65300c3b3.slice - libcontainer container kubepods-burstable-pod0739c4b8_550f_4633_885a_eec65300c3b3.slice. Sep 8 23:53:01.245429 sshd[4473]: Accepted publickey for core from 10.0.0.1 port 42404 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:53:01.246358 sshd-session[4473]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:53:01.254032 systemd-logind[1524]: New session 25 of user core. Sep 8 23:53:01.259576 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 8 23:53:01.312563 sshd[4476]: Connection closed by 10.0.0.1 port 42404 Sep 8 23:53:01.313088 sshd-session[4473]: pam_unix(sshd:session): session closed for user core Sep 8 23:53:01.327923 kubelet[2679]: I0908 23:53:01.327882 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-bpf-maps\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.327923 kubelet[2679]: I0908 23:53:01.327927 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-xtables-lock\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328086 kubelet[2679]: I0908 23:53:01.327957 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0739c4b8-550f-4633-885a-eec65300c3b3-cilium-config-path\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328086 kubelet[2679]: I0908 23:53:01.327976 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-lib-modules\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328086 kubelet[2679]: I0908 23:53:01.327995 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-host-proc-sys-kernel\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328086 kubelet[2679]: I0908 23:53:01.328012 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj5p7\" (UniqueName: \"kubernetes.io/projected/0739c4b8-550f-4633-885a-eec65300c3b3-kube-api-access-wj5p7\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328086 kubelet[2679]: I0908 23:53:01.328032 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-cni-path\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328202 kubelet[2679]: I0908 23:53:01.328061 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/0739c4b8-550f-4633-885a-eec65300c3b3-cilium-ipsec-secrets\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328202 kubelet[2679]: I0908 23:53:01.328086 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-hostproc\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328202 kubelet[2679]: I0908 23:53:01.328132 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0739c4b8-550f-4633-885a-eec65300c3b3-clustermesh-secrets\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328202 kubelet[2679]: I0908 23:53:01.328193 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-host-proc-sys-net\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328285 kubelet[2679]: I0908 23:53:01.328222 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-cilium-cgroup\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328285 kubelet[2679]: I0908 23:53:01.328244 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-etc-cni-netd\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328285 kubelet[2679]: I0908 23:53:01.328267 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0739c4b8-550f-4633-885a-eec65300c3b3-cilium-run\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.328401 kubelet[2679]: I0908 23:53:01.328305 2679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0739c4b8-550f-4633-885a-eec65300c3b3-hubble-tls\") pod \"cilium-25lcl\" (UID: \"0739c4b8-550f-4633-885a-eec65300c3b3\") " pod="kube-system/cilium-25lcl" Sep 8 23:53:01.329814 systemd[1]: sshd@24-10.0.0.130:22-10.0.0.1:42404.service: Deactivated successfully. Sep 8 23:53:01.332614 systemd[1]: session-25.scope: Deactivated successfully. Sep 8 23:53:01.333806 systemd-logind[1524]: Session 25 logged out. Waiting for processes to exit. Sep 8 23:53:01.336749 systemd[1]: Started sshd@25-10.0.0.130:22-10.0.0.1:42420.service - OpenSSH per-connection server daemon (10.0.0.1:42420). Sep 8 23:53:01.338674 systemd-logind[1524]: Removed session 25. Sep 8 23:53:01.397579 sshd[4483]: Accepted publickey for core from 10.0.0.1 port 42420 ssh2: RSA SHA256:HeCgiWNNKJuNyxF8eI797w9VfyFOv21mB1ET+U9TBic Sep 8 23:53:01.399020 sshd-session[4483]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 8 23:53:01.403087 systemd-logind[1524]: New session 26 of user core. Sep 8 23:53:01.413254 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 8 23:53:01.498430 containerd[1549]: time="2025-09-08T23:53:01.498385271Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-25lcl,Uid:0739c4b8-550f-4633-885a-eec65300c3b3,Namespace:kube-system,Attempt:0,}" Sep 8 23:53:01.527287 containerd[1549]: time="2025-09-08T23:53:01.527200913Z" level=info msg="connecting to shim 18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" namespace=k8s.io protocol=ttrpc version=3 Sep 8 23:53:01.554238 systemd[1]: Started cri-containerd-18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a.scope - libcontainer container 18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a. Sep 8 23:53:01.591132 containerd[1549]: time="2025-09-08T23:53:01.591050526Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-25lcl,Uid:0739c4b8-550f-4633-885a-eec65300c3b3,Namespace:kube-system,Attempt:0,} returns sandbox id \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\"" Sep 8 23:53:01.593971 containerd[1549]: time="2025-09-08T23:53:01.593573872Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 8 23:53:01.601224 containerd[1549]: time="2025-09-08T23:53:01.601185347Z" level=info msg="Container 64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:53:01.607369 containerd[1549]: time="2025-09-08T23:53:01.607327117Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\"" Sep 8 23:53:01.608315 containerd[1549]: time="2025-09-08T23:53:01.608273482Z" level=info msg="StartContainer for \"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\"" Sep 8 23:53:01.609140 containerd[1549]: time="2025-09-08T23:53:01.609080612Z" level=info msg="connecting to shim 64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" protocol=ttrpc version=3 Sep 8 23:53:01.630271 systemd[1]: Started cri-containerd-64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90.scope - libcontainer container 64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90. Sep 8 23:53:01.668671 containerd[1549]: time="2025-09-08T23:53:01.668088286Z" level=info msg="StartContainer for \"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\" returns successfully" Sep 8 23:53:01.677871 systemd[1]: cri-containerd-64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90.scope: Deactivated successfully. Sep 8 23:53:01.680231 containerd[1549]: time="2025-09-08T23:53:01.680186193Z" level=info msg="received exit event container_id:\"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\" id:\"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\" pid:4558 exited_at:{seconds:1757375581 nanos:679903084}" Sep 8 23:53:01.680536 containerd[1549]: time="2025-09-08T23:53:01.680380626Z" level=info msg="TaskExit event in podsandbox handler container_id:\"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\" id:\"64778e81876ff3a38f4731ec79227e18aea4b5f22dc92f4832e5aa9d1485ab90\" pid:4558 exited_at:{seconds:1757375581 nanos:679903084}" Sep 8 23:53:01.881465 kubelet[2679]: E0908 23:53:01.881409 2679 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 8 23:53:02.154573 containerd[1549]: time="2025-09-08T23:53:02.154534342Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 8 23:53:02.161170 containerd[1549]: time="2025-09-08T23:53:02.161123554Z" level=info msg="Container 9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:53:02.168262 containerd[1549]: time="2025-09-08T23:53:02.168213150Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\"" Sep 8 23:53:02.168789 containerd[1549]: time="2025-09-08T23:53:02.168754891Z" level=info msg="StartContainer for \"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\"" Sep 8 23:53:02.170052 containerd[1549]: time="2025-09-08T23:53:02.169967089Z" level=info msg="connecting to shim 9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" protocol=ttrpc version=3 Sep 8 23:53:02.194257 systemd[1]: Started cri-containerd-9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc.scope - libcontainer container 9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc. Sep 8 23:53:02.225319 containerd[1549]: time="2025-09-08T23:53:02.225279461Z" level=info msg="StartContainer for \"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\" returns successfully" Sep 8 23:53:02.228509 systemd[1]: cri-containerd-9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc.scope: Deactivated successfully. Sep 8 23:53:02.229738 containerd[1549]: time="2025-09-08T23:53:02.229382080Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\" id:\"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\" pid:4601 exited_at:{seconds:1757375582 nanos:228788900}" Sep 8 23:53:02.229738 containerd[1549]: time="2025-09-08T23:53:02.229612432Z" level=info msg="received exit event container_id:\"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\" id:\"9fad33495ba3117db6041d7bfb87bf2a060309a3ce17ebd2b434b7bd07535cfc\" pid:4601 exited_at:{seconds:1757375582 nanos:228788900}" Sep 8 23:53:03.154941 containerd[1549]: time="2025-09-08T23:53:03.154853407Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 8 23:53:03.173067 containerd[1549]: time="2025-09-08T23:53:03.172582046Z" level=info msg="Container aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:53:03.178055 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2147728871.mount: Deactivated successfully. Sep 8 23:53:03.182105 containerd[1549]: time="2025-09-08T23:53:03.182064225Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\"" Sep 8 23:53:03.183293 containerd[1549]: time="2025-09-08T23:53:03.183269307Z" level=info msg="StartContainer for \"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\"" Sep 8 23:53:03.185342 containerd[1549]: time="2025-09-08T23:53:03.185302962Z" level=info msg="connecting to shim aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" protocol=ttrpc version=3 Sep 8 23:53:03.211275 systemd[1]: Started cri-containerd-aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31.scope - libcontainer container aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31. Sep 8 23:53:03.253301 systemd[1]: cri-containerd-aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31.scope: Deactivated successfully. Sep 8 23:53:03.256243 containerd[1549]: time="2025-09-08T23:53:03.255573135Z" level=info msg="StartContainer for \"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\" returns successfully" Sep 8 23:53:03.256548 containerd[1549]: time="2025-09-08T23:53:03.255679252Z" level=info msg="received exit event container_id:\"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\" id:\"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\" pid:4647 exited_at:{seconds:1757375583 nanos:255509937}" Sep 8 23:53:03.256548 containerd[1549]: time="2025-09-08T23:53:03.256013481Z" level=info msg="TaskExit event in podsandbox handler container_id:\"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\" id:\"aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31\" pid:4647 exited_at:{seconds:1757375583 nanos:255509937}" Sep 8 23:53:03.286390 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-aa31ad16ff972b7a135d2db70b1b6df48c118cd44db836da665638aec0103b31-rootfs.mount: Deactivated successfully. Sep 8 23:53:03.644970 kubelet[2679]: I0908 23:53:03.644909 2679 setters.go:600] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-08T23:53:03Z","lastTransitionTime":"2025-09-08T23:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 8 23:53:04.157379 containerd[1549]: time="2025-09-08T23:53:04.157341340Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 8 23:53:04.168222 containerd[1549]: time="2025-09-08T23:53:04.168178866Z" level=info msg="Container 8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:53:04.174904 containerd[1549]: time="2025-09-08T23:53:04.174849753Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\"" Sep 8 23:53:04.175485 containerd[1549]: time="2025-09-08T23:53:04.175461095Z" level=info msg="StartContainer for \"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\"" Sep 8 23:53:04.177579 containerd[1549]: time="2025-09-08T23:53:04.177552954Z" level=info msg="connecting to shim 8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" protocol=ttrpc version=3 Sep 8 23:53:04.198219 systemd[1]: Started cri-containerd-8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3.scope - libcontainer container 8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3. Sep 8 23:53:04.225699 systemd[1]: cri-containerd-8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3.scope: Deactivated successfully. Sep 8 23:53:04.227903 containerd[1549]: time="2025-09-08T23:53:04.227865376Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\" id:\"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\" pid:4687 exited_at:{seconds:1757375584 nanos:227518307}" Sep 8 23:53:04.243947 containerd[1549]: time="2025-09-08T23:53:04.243735797Z" level=info msg="received exit event container_id:\"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\" id:\"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\" pid:4687 exited_at:{seconds:1757375584 nanos:227518307}" Sep 8 23:53:04.250964 containerd[1549]: time="2025-09-08T23:53:04.250932188Z" level=info msg="StartContainer for \"8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3\" returns successfully" Sep 8 23:53:04.267364 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8ed8621c3084398c33689b7e248aa26fe3c2d9d29f140e414c857e38c4d992f3-rootfs.mount: Deactivated successfully. Sep 8 23:53:05.164146 containerd[1549]: time="2025-09-08T23:53:05.164106716Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 8 23:53:05.178200 containerd[1549]: time="2025-09-08T23:53:05.178157785Z" level=info msg="Container 19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8: CDI devices from CRI Config.CDIDevices: []" Sep 8 23:53:05.187767 containerd[1549]: time="2025-09-08T23:53:05.187722893Z" level=info msg="CreateContainer within sandbox \"18c14e9ca08f5a8e809c3a21dbf743cfeb4ae056d0b1508dbf2da77aea49978a\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\"" Sep 8 23:53:05.189053 containerd[1549]: time="2025-09-08T23:53:05.188967941Z" level=info msg="StartContainer for \"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\"" Sep 8 23:53:05.190131 containerd[1549]: time="2025-09-08T23:53:05.190082551Z" level=info msg="connecting to shim 19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8" address="unix:///run/containerd/s/ac55c3fcb0f27d5a76b763503065ded887adf98eb6a94fcc7c257f5166270b2d" protocol=ttrpc version=3 Sep 8 23:53:05.214374 systemd[1]: Started cri-containerd-19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8.scope - libcontainer container 19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8. Sep 8 23:53:05.259118 containerd[1549]: time="2025-09-08T23:53:05.259077973Z" level=info msg="StartContainer for \"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" returns successfully" Sep 8 23:53:05.318427 containerd[1549]: time="2025-09-08T23:53:05.318388171Z" level=info msg="TaskExit event in podsandbox handler container_id:\"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" id:\"eb4d2cc2f6ad39f9e5054e4a06f50a215a089514d01eb5e4ef69f3f0107889e9\" pid:4754 exited_at:{seconds:1757375585 nanos:317963382}" Sep 8 23:53:05.567070 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Sep 8 23:53:07.875540 containerd[1549]: time="2025-09-08T23:53:07.874788429Z" level=info msg="TaskExit event in podsandbox handler container_id:\"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" id:\"963f10671d4ffc1aa99758d6d606bbdb1daaf55f37de26a6703ceb32b2acc551\" pid:5083 exit_status:1 exited_at:{seconds:1757375587 nanos:874455596}" Sep 8 23:53:08.561029 systemd-networkd[1441]: lxc_health: Link UP Sep 8 23:53:08.561338 systemd-networkd[1441]: lxc_health: Gained carrier Sep 8 23:53:09.521077 kubelet[2679]: I0908 23:53:09.521000 2679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-25lcl" podStartSLOduration=8.520981836 podStartE2EDuration="8.520981836s" podCreationTimestamp="2025-09-08 23:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-08 23:53:06.201986683 +0000 UTC m=+84.467885999" watchObservedRunningTime="2025-09-08 23:53:09.520981836 +0000 UTC m=+87.786881152" Sep 8 23:53:10.002714 containerd[1549]: time="2025-09-08T23:53:10.002676353Z" level=info msg="TaskExit event in podsandbox handler container_id:\"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" id:\"09128b311462cbd425efa5945a6972f5a4ba094d75f69b7cf308266241fb6dc8\" pid:5297 exited_at:{seconds:1757375590 nanos:2338038}" Sep 8 23:53:10.374255 systemd-networkd[1441]: lxc_health: Gained IPv6LL Sep 8 23:53:12.126764 containerd[1549]: time="2025-09-08T23:53:12.126703527Z" level=info msg="TaskExit event in podsandbox handler container_id:\"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" id:\"33a2cf89121f3056db6dea84969ee18807c4416fde176b72509ff11dc5797f06\" pid:5324 exited_at:{seconds:1757375592 nanos:126323571}" Sep 8 23:53:14.251252 containerd[1549]: time="2025-09-08T23:53:14.250904198Z" level=info msg="TaskExit event in podsandbox handler container_id:\"19eb7f742b976f58aef4b615e2df370e249e04921fcbecc36c71df63e57077d8\" id:\"cd638bad1ab11dcfbc2675d692fc8bc3d2232aadc0cc3877d4878328ed5bac0a\" pid:5357 exited_at:{seconds:1757375594 nanos:250645720}" Sep 8 23:53:14.281022 sshd[4486]: Connection closed by 10.0.0.1 port 42420 Sep 8 23:53:14.281629 sshd-session[4483]: pam_unix(sshd:session): session closed for user core Sep 8 23:53:14.285421 systemd[1]: sshd@25-10.0.0.130:22-10.0.0.1:42420.service: Deactivated successfully. Sep 8 23:53:14.287298 systemd[1]: session-26.scope: Deactivated successfully. Sep 8 23:53:14.288050 systemd-logind[1524]: Session 26 logged out. Waiting for processes to exit. Sep 8 23:53:14.289285 systemd-logind[1524]: Removed session 26.