Sep 13 00:05:19.705825 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 13 00:05:19.705846 kernel: Linux version 5.15.192-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Fri Sep 12 23:05:37 -00 2025 Sep 13 00:05:19.705853 kernel: efi: EFI v2.70 by EDK II Sep 13 00:05:19.706442 kernel: efi: SMBIOS 3.0=0xd9260000 ACPI 2.0=0xd9240000 MEMATTR=0xda32b018 RNG=0xd9220018 MEMRESERVE=0xd9521c18 Sep 13 00:05:19.706452 kernel: random: crng init done Sep 13 00:05:19.706458 kernel: ACPI: Early table checksum verification disabled Sep 13 00:05:19.706464 kernel: ACPI: RSDP 0x00000000D9240000 000024 (v02 BOCHS ) Sep 13 00:05:19.706475 kernel: ACPI: XSDT 0x00000000D9230000 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 13 00:05:19.706481 kernel: ACPI: FACP 0x00000000D91E0000 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706494 kernel: ACPI: DSDT 0x00000000D91F0000 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706500 kernel: ACPI: APIC 0x00000000D91D0000 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706505 kernel: ACPI: PPTT 0x00000000D91C0000 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706511 kernel: ACPI: GTDT 0x00000000D91B0000 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706517 kernel: ACPI: MCFG 0x00000000D91A0000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706525 kernel: ACPI: SPCR 0x00000000D9190000 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706531 kernel: ACPI: DBG2 0x00000000D9180000 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706537 kernel: ACPI: IORT 0x00000000D9170000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:05:19.706542 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 13 00:05:19.706548 kernel: NUMA: Failed to initialise from firmware Sep 13 00:05:19.706555 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 13 00:05:19.706561 kernel: NUMA: NODE_DATA [mem 0xdcb0b900-0xdcb10fff] Sep 13 00:05:19.706566 kernel: Zone ranges: Sep 13 00:05:19.706572 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 13 00:05:19.706579 kernel: DMA32 empty Sep 13 00:05:19.706584 kernel: Normal empty Sep 13 00:05:19.706590 kernel: Movable zone start for each node Sep 13 00:05:19.706595 kernel: Early memory node ranges Sep 13 00:05:19.706602 kernel: node 0: [mem 0x0000000040000000-0x00000000d924ffff] Sep 13 00:05:19.706608 kernel: node 0: [mem 0x00000000d9250000-0x00000000d951ffff] Sep 13 00:05:19.706614 kernel: node 0: [mem 0x00000000d9520000-0x00000000dc7fffff] Sep 13 00:05:19.706619 kernel: node 0: [mem 0x00000000dc800000-0x00000000dc88ffff] Sep 13 00:05:19.706625 kernel: node 0: [mem 0x00000000dc890000-0x00000000dc89ffff] Sep 13 00:05:19.706631 kernel: node 0: [mem 0x00000000dc8a0000-0x00000000dc9bffff] Sep 13 00:05:19.706637 kernel: node 0: [mem 0x00000000dc9c0000-0x00000000dcffffff] Sep 13 00:05:19.706643 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 13 00:05:19.706650 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 13 00:05:19.706655 kernel: psci: probing for conduit method from ACPI. Sep 13 00:05:19.706661 kernel: psci: PSCIv1.1 detected in firmware. Sep 13 00:05:19.706667 kernel: psci: Using standard PSCI v0.2 function IDs Sep 13 00:05:19.706673 kernel: psci: Trusted OS migration not required Sep 13 00:05:19.706681 kernel: psci: SMC Calling Convention v1.1 Sep 13 00:05:19.706687 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 13 00:05:19.706695 kernel: ACPI: SRAT not present Sep 13 00:05:19.706701 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 Sep 13 00:05:19.706707 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 Sep 13 00:05:19.706714 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 13 00:05:19.706754 kernel: Detected PIPT I-cache on CPU0 Sep 13 00:05:19.706763 kernel: CPU features: detected: GIC system register CPU interface Sep 13 00:05:19.706769 kernel: CPU features: detected: Hardware dirty bit management Sep 13 00:05:19.706775 kernel: CPU features: detected: Spectre-v4 Sep 13 00:05:19.706781 kernel: CPU features: detected: Spectre-BHB Sep 13 00:05:19.706789 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 13 00:05:19.706795 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 13 00:05:19.706802 kernel: CPU features: detected: ARM erratum 1418040 Sep 13 00:05:19.706808 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 13 00:05:19.706814 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Sep 13 00:05:19.706820 kernel: Policy zone: DMA Sep 13 00:05:19.706827 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=563df7b8a9b19b8c496587ae06f3c3ec1604a5105c3a3f313c9ccaa21d8055ca Sep 13 00:05:19.706834 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 00:05:19.706840 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 13 00:05:19.706846 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 00:05:19.706852 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 00:05:19.706860 kernel: Memory: 2457340K/2572288K available (9792K kernel code, 2094K rwdata, 7592K rodata, 36416K init, 777K bss, 114948K reserved, 0K cma-reserved) Sep 13 00:05:19.706866 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 13 00:05:19.706872 kernel: trace event string verifier disabled Sep 13 00:05:19.706878 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 13 00:05:19.706885 kernel: rcu: RCU event tracing is enabled. Sep 13 00:05:19.706891 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 13 00:05:19.706897 kernel: Trampoline variant of Tasks RCU enabled. Sep 13 00:05:19.706903 kernel: Tracing variant of Tasks RCU enabled. Sep 13 00:05:19.706910 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 00:05:19.706916 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 13 00:05:19.706922 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 13 00:05:19.706929 kernel: GICv3: 256 SPIs implemented Sep 13 00:05:19.706935 kernel: GICv3: 0 Extended SPIs implemented Sep 13 00:05:19.706941 kernel: GICv3: Distributor has no Range Selector support Sep 13 00:05:19.706947 kernel: Root IRQ handler: gic_handle_irq Sep 13 00:05:19.706953 kernel: GICv3: 16 PPIs implemented Sep 13 00:05:19.706959 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 13 00:05:19.706965 kernel: ACPI: SRAT not present Sep 13 00:05:19.706971 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 13 00:05:19.706977 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400b0000 (indirect, esz 8, psz 64K, shr 1) Sep 13 00:05:19.706983 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400c0000 (flat, esz 8, psz 64K, shr 1) Sep 13 00:05:19.706989 kernel: GICv3: using LPI property table @0x00000000400d0000 Sep 13 00:05:19.706995 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000000400e0000 Sep 13 00:05:19.707003 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 13 00:05:19.707009 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 13 00:05:19.707015 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 13 00:05:19.707021 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 13 00:05:19.707027 kernel: arm-pv: using stolen time PV Sep 13 00:05:19.707034 kernel: Console: colour dummy device 80x25 Sep 13 00:05:19.707040 kernel: ACPI: Core revision 20210730 Sep 13 00:05:19.707046 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 13 00:05:19.707053 kernel: pid_max: default: 32768 minimum: 301 Sep 13 00:05:19.707059 kernel: LSM: Security Framework initializing Sep 13 00:05:19.707066 kernel: SELinux: Initializing. Sep 13 00:05:19.707073 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 13 00:05:19.707079 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 13 00:05:19.707085 kernel: rcu: Hierarchical SRCU implementation. Sep 13 00:05:19.707091 kernel: Platform MSI: ITS@0x8080000 domain created Sep 13 00:05:19.707097 kernel: PCI/MSI: ITS@0x8080000 domain created Sep 13 00:05:19.707104 kernel: Remapping and enabling EFI services. Sep 13 00:05:19.707110 kernel: smp: Bringing up secondary CPUs ... Sep 13 00:05:19.707116 kernel: Detected PIPT I-cache on CPU1 Sep 13 00:05:19.707124 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 13 00:05:19.707130 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000000400f0000 Sep 13 00:05:19.707136 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 13 00:05:19.707143 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 13 00:05:19.707149 kernel: Detected PIPT I-cache on CPU2 Sep 13 00:05:19.707155 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 13 00:05:19.707162 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040100000 Sep 13 00:05:19.707168 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 13 00:05:19.707175 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 13 00:05:19.707181 kernel: Detected PIPT I-cache on CPU3 Sep 13 00:05:19.707188 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 13 00:05:19.707195 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040110000 Sep 13 00:05:19.707201 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 13 00:05:19.707207 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 13 00:05:19.707218 kernel: smp: Brought up 1 node, 4 CPUs Sep 13 00:05:19.707226 kernel: SMP: Total of 4 processors activated. Sep 13 00:05:19.707232 kernel: CPU features: detected: 32-bit EL0 Support Sep 13 00:05:19.707239 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 13 00:05:19.707246 kernel: CPU features: detected: Common not Private translations Sep 13 00:05:19.707252 kernel: CPU features: detected: CRC32 instructions Sep 13 00:05:19.707259 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 13 00:05:19.707265 kernel: CPU features: detected: LSE atomic instructions Sep 13 00:05:19.707273 kernel: CPU features: detected: Privileged Access Never Sep 13 00:05:19.707280 kernel: CPU features: detected: RAS Extension Support Sep 13 00:05:19.707287 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 13 00:05:19.707294 kernel: CPU: All CPU(s) started at EL1 Sep 13 00:05:19.707300 kernel: alternatives: patching kernel code Sep 13 00:05:19.707308 kernel: devtmpfs: initialized Sep 13 00:05:19.707314 kernel: KASLR enabled Sep 13 00:05:19.707321 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 00:05:19.707328 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 13 00:05:19.707335 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 00:05:19.707341 kernel: SMBIOS 3.0.0 present. Sep 13 00:05:19.707348 kernel: DMI: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 Sep 13 00:05:19.707355 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 00:05:19.707361 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 13 00:05:19.707369 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 13 00:05:19.707376 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 13 00:05:19.707383 kernel: audit: initializing netlink subsys (disabled) Sep 13 00:05:19.707390 kernel: audit: type=2000 audit(0.031:1): state=initialized audit_enabled=0 res=1 Sep 13 00:05:19.707396 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 00:05:19.707403 kernel: cpuidle: using governor menu Sep 13 00:05:19.707409 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 13 00:05:19.707416 kernel: ASID allocator initialised with 32768 entries Sep 13 00:05:19.707422 kernel: ACPI: bus type PCI registered Sep 13 00:05:19.707430 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 00:05:19.707436 kernel: Serial: AMBA PL011 UART driver Sep 13 00:05:19.707443 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Sep 13 00:05:19.707450 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Sep 13 00:05:19.707456 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 00:05:19.707463 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Sep 13 00:05:19.707470 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 00:05:19.707477 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 13 00:05:19.707483 kernel: ACPI: Added _OSI(Module Device) Sep 13 00:05:19.707491 kernel: ACPI: Added _OSI(Processor Device) Sep 13 00:05:19.707498 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 00:05:19.707504 kernel: ACPI: Added _OSI(Linux-Dell-Video) Sep 13 00:05:19.707511 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Sep 13 00:05:19.707518 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Sep 13 00:05:19.707524 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 13 00:05:19.707531 kernel: ACPI: Interpreter enabled Sep 13 00:05:19.707537 kernel: ACPI: Using GIC for interrupt routing Sep 13 00:05:19.707544 kernel: ACPI: MCFG table detected, 1 entries Sep 13 00:05:19.707552 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 13 00:05:19.707559 kernel: printk: console [ttyAMA0] enabled Sep 13 00:05:19.707566 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 13 00:05:19.707716 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 13 00:05:19.707811 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 13 00:05:19.707874 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 13 00:05:19.707937 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 13 00:05:19.708009 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 13 00:05:19.708019 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 13 00:05:19.708026 kernel: PCI host bridge to bus 0000:00 Sep 13 00:05:19.708106 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 13 00:05:19.708168 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 13 00:05:19.708224 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 13 00:05:19.708280 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 13 00:05:19.708402 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Sep 13 00:05:19.708503 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Sep 13 00:05:19.708578 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Sep 13 00:05:19.708643 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Sep 13 00:05:19.708706 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Sep 13 00:05:19.708790 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Sep 13 00:05:19.708856 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Sep 13 00:05:19.708922 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Sep 13 00:05:19.708979 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 13 00:05:19.709032 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 13 00:05:19.709086 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 13 00:05:19.709096 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 13 00:05:19.709103 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 13 00:05:19.709109 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 13 00:05:19.709117 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 13 00:05:19.709124 kernel: iommu: Default domain type: Translated Sep 13 00:05:19.709131 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 13 00:05:19.709137 kernel: vgaarb: loaded Sep 13 00:05:19.709187 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 13 00:05:19.709195 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 13 00:05:19.709202 kernel: PTP clock support registered Sep 13 00:05:19.709208 kernel: Registered efivars operations Sep 13 00:05:19.709215 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 13 00:05:19.709222 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 00:05:19.709232 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 00:05:19.709239 kernel: pnp: PnP ACPI init Sep 13 00:05:19.709332 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 13 00:05:19.709343 kernel: pnp: PnP ACPI: found 1 devices Sep 13 00:05:19.709350 kernel: NET: Registered PF_INET protocol family Sep 13 00:05:19.709357 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 13 00:05:19.709364 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 13 00:05:19.709371 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 00:05:19.709379 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 13 00:05:19.709386 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Sep 13 00:05:19.709393 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 13 00:05:19.709400 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 13 00:05:19.709407 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 13 00:05:19.709414 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 00:05:19.709421 kernel: PCI: CLS 0 bytes, default 64 Sep 13 00:05:19.709428 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Sep 13 00:05:19.709435 kernel: kvm [1]: HYP mode not available Sep 13 00:05:19.709443 kernel: Initialise system trusted keyrings Sep 13 00:05:19.709450 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 13 00:05:19.709457 kernel: Key type asymmetric registered Sep 13 00:05:19.709464 kernel: Asymmetric key parser 'x509' registered Sep 13 00:05:19.709472 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 13 00:05:19.709479 kernel: io scheduler mq-deadline registered Sep 13 00:05:19.709486 kernel: io scheduler kyber registered Sep 13 00:05:19.709493 kernel: io scheduler bfq registered Sep 13 00:05:19.709500 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 13 00:05:19.709509 kernel: ACPI: button: Power Button [PWRB] Sep 13 00:05:19.709516 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 13 00:05:19.709587 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 13 00:05:19.709597 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 00:05:19.709604 kernel: thunder_xcv, ver 1.0 Sep 13 00:05:19.709611 kernel: thunder_bgx, ver 1.0 Sep 13 00:05:19.709617 kernel: nicpf, ver 1.0 Sep 13 00:05:19.709624 kernel: nicvf, ver 1.0 Sep 13 00:05:19.709696 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 13 00:05:19.709802 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-09-13T00:05:19 UTC (1757721919) Sep 13 00:05:19.709816 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 13 00:05:19.709824 kernel: NET: Registered PF_INET6 protocol family Sep 13 00:05:19.709830 kernel: Segment Routing with IPv6 Sep 13 00:05:19.709837 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 00:05:19.709844 kernel: NET: Registered PF_PACKET protocol family Sep 13 00:05:19.709851 kernel: Key type dns_resolver registered Sep 13 00:05:19.709860 kernel: registered taskstats version 1 Sep 13 00:05:19.709869 kernel: Loading compiled-in X.509 certificates Sep 13 00:05:19.709876 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.192-flatcar: 47ac98e9306f36eebe4291d409359a5a5d0c2b9c' Sep 13 00:05:19.709883 kernel: Key type .fscrypt registered Sep 13 00:05:19.709889 kernel: Key type fscrypt-provisioning registered Sep 13 00:05:19.709898 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 13 00:05:19.709905 kernel: ima: Allocated hash algorithm: sha1 Sep 13 00:05:19.709912 kernel: ima: No architecture policies found Sep 13 00:05:19.709919 kernel: clk: Disabling unused clocks Sep 13 00:05:19.709925 kernel: Freeing unused kernel memory: 36416K Sep 13 00:05:19.709933 kernel: Run /init as init process Sep 13 00:05:19.709940 kernel: with arguments: Sep 13 00:05:19.709947 kernel: /init Sep 13 00:05:19.709953 kernel: with environment: Sep 13 00:05:19.709960 kernel: HOME=/ Sep 13 00:05:19.709969 kernel: TERM=linux Sep 13 00:05:19.709976 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 00:05:19.709984 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:05:19.709995 systemd[1]: Detected virtualization kvm. Sep 13 00:05:19.710003 systemd[1]: Detected architecture arm64. Sep 13 00:05:19.710009 systemd[1]: Running in initrd. Sep 13 00:05:19.710016 systemd[1]: No hostname configured, using default hostname. Sep 13 00:05:19.710023 systemd[1]: Hostname set to . Sep 13 00:05:19.710042 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:05:19.710050 systemd[1]: Queued start job for default target initrd.target. Sep 13 00:05:19.710059 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:05:19.710068 systemd[1]: Reached target cryptsetup.target. Sep 13 00:05:19.710075 systemd[1]: Reached target paths.target. Sep 13 00:05:19.710082 systemd[1]: Reached target slices.target. Sep 13 00:05:19.710088 systemd[1]: Reached target swap.target. Sep 13 00:05:19.710096 systemd[1]: Reached target timers.target. Sep 13 00:05:19.710103 systemd[1]: Listening on iscsid.socket. Sep 13 00:05:19.710110 systemd[1]: Listening on iscsiuio.socket. Sep 13 00:05:19.710120 systemd[1]: Listening on systemd-journald-audit.socket. Sep 13 00:05:19.710128 systemd[1]: Listening on systemd-journald-dev-log.socket. Sep 13 00:05:19.710135 systemd[1]: Listening on systemd-journald.socket. Sep 13 00:05:19.710142 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:05:19.710150 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:05:19.710157 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:05:19.710164 systemd[1]: Reached target sockets.target. Sep 13 00:05:19.710171 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:05:19.710180 systemd[1]: Finished network-cleanup.service. Sep 13 00:05:19.710189 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 00:05:19.710196 systemd[1]: Starting systemd-journald.service... Sep 13 00:05:19.710204 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:05:19.710212 systemd[1]: Starting systemd-resolved.service... Sep 13 00:05:19.710220 systemd[1]: Starting systemd-vconsole-setup.service... Sep 13 00:05:19.710228 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:05:19.710235 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 00:05:19.710243 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 13 00:05:19.710256 systemd-journald[289]: Journal started Sep 13 00:05:19.710311 systemd-journald[289]: Runtime Journal (/run/log/journal/120c733ea798469392b581a2dcb84651) is 6.0M, max 48.7M, 42.6M free. Sep 13 00:05:19.708531 systemd-modules-load[290]: Inserted module 'overlay' Sep 13 00:05:19.714154 systemd[1]: Started systemd-journald.service. Sep 13 00:05:19.714000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.714550 systemd[1]: Finished systemd-vconsole-setup.service. Sep 13 00:05:19.716909 kernel: audit: type=1130 audit(1757721919.714:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.716000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.717794 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 13 00:05:19.724402 kernel: audit: type=1130 audit(1757721919.716:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.722296 systemd-resolved[291]: Positive Trust Anchors: Sep 13 00:05:19.722304 systemd-resolved[291]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:05:19.730272 kernel: audit: type=1130 audit(1757721919.724:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.730293 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 00:05:19.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.722331 systemd-resolved[291]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:05:19.726909 systemd[1]: Starting dracut-cmdline-ask.service... Sep 13 00:05:19.738740 kernel: Bridge firewalling registered Sep 13 00:05:19.738763 kernel: audit: type=1130 audit(1757721919.735:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.735000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.727184 systemd-resolved[291]: Defaulting to hostname 'linux'. Sep 13 00:05:19.730911 systemd[1]: Started systemd-resolved.service. Sep 13 00:05:19.736120 systemd-modules-load[290]: Inserted module 'br_netfilter' Sep 13 00:05:19.737416 systemd[1]: Reached target nss-lookup.target. Sep 13 00:05:19.747592 systemd[1]: Finished dracut-cmdline-ask.service. Sep 13 00:05:19.748786 kernel: SCSI subsystem initialized Sep 13 00:05:19.748000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.750226 systemd[1]: Starting dracut-cmdline.service... Sep 13 00:05:19.752024 kernel: audit: type=1130 audit(1757721919.748:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.756008 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 00:05:19.756047 kernel: device-mapper: uevent: version 1.0.3 Sep 13 00:05:19.756058 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Sep 13 00:05:19.759061 systemd-modules-load[290]: Inserted module 'dm_multipath' Sep 13 00:05:19.759801 dracut-cmdline[307]: dracut-dracut-053 Sep 13 00:05:19.760011 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:05:19.761000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.763756 kernel: audit: type=1130 audit(1757721919.761:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.762058 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:05:19.765071 dracut-cmdline[307]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=563df7b8a9b19b8c496587ae06f3c3ec1604a5105c3a3f313c9ccaa21d8055ca Sep 13 00:05:19.769951 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:05:19.769000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.773753 kernel: audit: type=1130 audit(1757721919.769:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.825770 kernel: Loading iSCSI transport class v2.0-870. Sep 13 00:05:19.837773 kernel: iscsi: registered transport (tcp) Sep 13 00:05:19.852772 kernel: iscsi: registered transport (qla4xxx) Sep 13 00:05:19.852826 kernel: QLogic iSCSI HBA Driver Sep 13 00:05:19.888451 systemd[1]: Finished dracut-cmdline.service. Sep 13 00:05:19.888000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.889983 systemd[1]: Starting dracut-pre-udev.service... Sep 13 00:05:19.892522 kernel: audit: type=1130 audit(1757721919.888:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:19.937772 kernel: raid6: neonx8 gen() 13735 MB/s Sep 13 00:05:19.954758 kernel: raid6: neonx8 xor() 10732 MB/s Sep 13 00:05:19.971752 kernel: raid6: neonx4 gen() 13441 MB/s Sep 13 00:05:19.988747 kernel: raid6: neonx4 xor() 11121 MB/s Sep 13 00:05:20.005759 kernel: raid6: neonx2 gen() 12896 MB/s Sep 13 00:05:20.022792 kernel: raid6: neonx2 xor() 10234 MB/s Sep 13 00:05:20.039769 kernel: raid6: neonx1 gen() 10583 MB/s Sep 13 00:05:20.056777 kernel: raid6: neonx1 xor() 8764 MB/s Sep 13 00:05:20.073779 kernel: raid6: int64x8 gen() 6257 MB/s Sep 13 00:05:20.090768 kernel: raid6: int64x8 xor() 3533 MB/s Sep 13 00:05:20.107772 kernel: raid6: int64x4 gen() 7167 MB/s Sep 13 00:05:20.124774 kernel: raid6: int64x4 xor() 3842 MB/s Sep 13 00:05:20.141762 kernel: raid6: int64x2 gen() 6130 MB/s Sep 13 00:05:20.158754 kernel: raid6: int64x2 xor() 3301 MB/s Sep 13 00:05:20.175771 kernel: raid6: int64x1 gen() 5021 MB/s Sep 13 00:05:20.193252 kernel: raid6: int64x1 xor() 2631 MB/s Sep 13 00:05:20.193311 kernel: raid6: using algorithm neonx8 gen() 13735 MB/s Sep 13 00:05:20.193320 kernel: raid6: .... xor() 10732 MB/s, rmw enabled Sep 13 00:05:20.193329 kernel: raid6: using neon recovery algorithm Sep 13 00:05:20.203811 kernel: xor: measuring software checksum speed Sep 13 00:05:20.203853 kernel: 8regs : 17056 MB/sec Sep 13 00:05:20.205087 kernel: 32regs : 20728 MB/sec Sep 13 00:05:20.205122 kernel: arm64_neon : 27794 MB/sec Sep 13 00:05:20.205132 kernel: xor: using function: arm64_neon (27794 MB/sec) Sep 13 00:05:20.264785 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Sep 13 00:05:20.283498 systemd[1]: Finished dracut-pre-udev.service. Sep 13 00:05:20.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:20.287000 audit: BPF prog-id=7 op=LOAD Sep 13 00:05:20.287000 audit: BPF prog-id=8 op=LOAD Sep 13 00:05:20.288765 kernel: audit: type=1130 audit(1757721920.283:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:20.289104 systemd[1]: Starting systemd-udevd.service... Sep 13 00:05:20.303176 systemd-udevd[490]: Using default interface naming scheme 'v252'. Sep 13 00:05:20.306542 systemd[1]: Started systemd-udevd.service. Sep 13 00:05:20.307000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:20.312532 systemd[1]: Starting dracut-pre-trigger.service... Sep 13 00:05:20.326456 dracut-pre-trigger[509]: rd.md=0: removing MD RAID activation Sep 13 00:05:20.355960 systemd[1]: Finished dracut-pre-trigger.service. Sep 13 00:05:20.355000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:20.357338 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:05:20.399295 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:05:20.399000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:20.431807 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 13 00:05:20.446618 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 00:05:20.446634 kernel: GPT:9289727 != 19775487 Sep 13 00:05:20.446642 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 00:05:20.446657 kernel: GPT:9289727 != 19775487 Sep 13 00:05:20.446665 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 00:05:20.446673 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:05:20.459773 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (555) Sep 13 00:05:20.462952 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Sep 13 00:05:20.465588 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Sep 13 00:05:20.466480 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Sep 13 00:05:20.472452 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Sep 13 00:05:20.475771 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:05:20.477252 systemd[1]: Starting disk-uuid.service... Sep 13 00:05:20.483993 disk-uuid[564]: Primary Header is updated. Sep 13 00:05:20.483993 disk-uuid[564]: Secondary Entries is updated. Sep 13 00:05:20.483993 disk-uuid[564]: Secondary Header is updated. Sep 13 00:05:20.488750 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:05:20.491760 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:05:20.494751 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:05:21.494366 disk-uuid[565]: The operation has completed successfully. Sep 13 00:05:21.495379 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:05:21.534117 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 00:05:21.534000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.534000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.534217 systemd[1]: Finished disk-uuid.service. Sep 13 00:05:21.549272 systemd[1]: Starting verity-setup.service... Sep 13 00:05:21.564755 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Sep 13 00:05:21.608394 systemd[1]: Found device dev-mapper-usr.device. Sep 13 00:05:21.612764 systemd[1]: Mounting sysusr-usr.mount... Sep 13 00:05:21.615140 systemd[1]: Finished verity-setup.service. Sep 13 00:05:21.615000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.667705 systemd[1]: Mounted sysusr-usr.mount. Sep 13 00:05:21.668769 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Sep 13 00:05:21.668397 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Sep 13 00:05:21.669202 systemd[1]: Starting ignition-setup.service... Sep 13 00:05:21.671541 systemd[1]: Starting parse-ip-for-networkd.service... Sep 13 00:05:21.682939 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 13 00:05:21.683056 kernel: BTRFS info (device vda6): using free space tree Sep 13 00:05:21.683084 kernel: BTRFS info (device vda6): has skinny extents Sep 13 00:05:21.694762 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 13 00:05:21.705166 systemd[1]: Finished ignition-setup.service. Sep 13 00:05:21.705000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.708136 systemd[1]: Starting ignition-fetch-offline.service... Sep 13 00:05:21.787167 systemd[1]: Finished parse-ip-for-networkd.service. Sep 13 00:05:21.787000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.788000 audit: BPF prog-id=9 op=LOAD Sep 13 00:05:21.789658 systemd[1]: Starting systemd-networkd.service... Sep 13 00:05:21.809657 ignition[660]: Ignition 2.14.0 Sep 13 00:05:21.810418 ignition[660]: Stage: fetch-offline Sep 13 00:05:21.811034 ignition[660]: no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:21.811756 ignition[660]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:21.811856 systemd-networkd[741]: lo: Link UP Sep 13 00:05:21.811860 systemd-networkd[741]: lo: Gained carrier Sep 13 00:05:21.813000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.812267 systemd-networkd[741]: Enumeration completed Sep 13 00:05:21.812386 systemd[1]: Started systemd-networkd.service. Sep 13 00:05:21.813425 systemd[1]: Reached target network.target. Sep 13 00:05:21.816796 ignition[660]: parsed url from cmdline: "" Sep 13 00:05:21.815244 systemd[1]: Starting iscsiuio.service... Sep 13 00:05:21.816801 ignition[660]: no config URL provided Sep 13 00:05:21.816418 systemd-networkd[741]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 00:05:21.816809 ignition[660]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 00:05:21.818171 systemd-networkd[741]: eth0: Link UP Sep 13 00:05:21.816898 ignition[660]: no config at "/usr/lib/ignition/user.ign" Sep 13 00:05:21.818175 systemd-networkd[741]: eth0: Gained carrier Sep 13 00:05:21.816928 ignition[660]: op(1): [started] loading QEMU firmware config module Sep 13 00:05:21.816933 ignition[660]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 13 00:05:21.823593 systemd[1]: Started iscsiuio.service. Sep 13 00:05:21.824000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.825705 systemd[1]: Starting iscsid.service... Sep 13 00:05:21.826381 ignition[660]: op(1): [finished] loading QEMU firmware config module Sep 13 00:05:21.829476 iscsid[748]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:05:21.829476 iscsid[748]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Sep 13 00:05:21.829476 iscsid[748]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Sep 13 00:05:21.829476 iscsid[748]: If using hardware iscsi like qla4xxx this message can be ignored. Sep 13 00:05:21.835000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.838746 iscsid[748]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:05:21.838746 iscsid[748]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Sep 13 00:05:21.832776 systemd[1]: Started iscsid.service. Sep 13 00:05:21.837603 systemd[1]: Starting dracut-initqueue.service... Sep 13 00:05:21.841862 systemd-networkd[741]: eth0: DHCPv4 address 10.0.0.20/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 13 00:05:21.850009 systemd[1]: Finished dracut-initqueue.service. Sep 13 00:05:21.850000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.850865 systemd[1]: Reached target remote-fs-pre.target. Sep 13 00:05:21.852556 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:05:21.854018 systemd[1]: Reached target remote-fs.target. Sep 13 00:05:21.856173 systemd[1]: Starting dracut-pre-mount.service... Sep 13 00:05:21.864309 systemd[1]: Finished dracut-pre-mount.service. Sep 13 00:05:21.864000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.894038 ignition[660]: parsing config with SHA512: e1995fdbd005765bee96f9928bb4710dce086132999b393cf3ac6bf2605d9ab5bf8ad8ec825b0662dbfa7ff8c3d07a8dfab87189d2d4a8b04a601df786a6ee77 Sep 13 00:05:21.901617 unknown[660]: fetched base config from "system" Sep 13 00:05:21.901629 unknown[660]: fetched user config from "qemu" Sep 13 00:05:21.902223 ignition[660]: fetch-offline: fetch-offline passed Sep 13 00:05:21.903162 systemd[1]: Finished ignition-fetch-offline.service. Sep 13 00:05:21.903000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.902282 ignition[660]: Ignition finished successfully Sep 13 00:05:21.904526 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 13 00:05:21.905315 systemd[1]: Starting ignition-kargs.service... Sep 13 00:05:21.914928 ignition[762]: Ignition 2.14.0 Sep 13 00:05:21.914941 ignition[762]: Stage: kargs Sep 13 00:05:21.915041 ignition[762]: no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:21.916956 systemd[1]: Finished ignition-kargs.service. Sep 13 00:05:21.918000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.915051 ignition[762]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:21.915901 ignition[762]: kargs: kargs passed Sep 13 00:05:21.919177 systemd[1]: Starting ignition-disks.service... Sep 13 00:05:21.915946 ignition[762]: Ignition finished successfully Sep 13 00:05:21.925500 ignition[768]: Ignition 2.14.0 Sep 13 00:05:21.925510 ignition[768]: Stage: disks Sep 13 00:05:21.925602 ignition[768]: no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:21.925611 ignition[768]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:21.926539 ignition[768]: disks: disks passed Sep 13 00:05:21.926582 ignition[768]: Ignition finished successfully Sep 13 00:05:21.929980 systemd[1]: Finished ignition-disks.service. Sep 13 00:05:21.930000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.930867 systemd[1]: Reached target initrd-root-device.target. Sep 13 00:05:21.931987 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:05:21.933295 systemd[1]: Reached target local-fs.target. Sep 13 00:05:21.934390 systemd[1]: Reached target sysinit.target. Sep 13 00:05:21.935526 systemd[1]: Reached target basic.target. Sep 13 00:05:21.937758 systemd[1]: Starting systemd-fsck-root.service... Sep 13 00:05:21.950774 systemd-fsck[776]: ROOT: clean, 629/553520 files, 56027/553472 blocks Sep 13 00:05:21.955060 systemd[1]: Finished systemd-fsck-root.service. Sep 13 00:05:21.955000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:21.956535 systemd[1]: Mounting sysroot.mount... Sep 13 00:05:21.966799 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Sep 13 00:05:21.967000 systemd[1]: Mounted sysroot.mount. Sep 13 00:05:21.967666 systemd[1]: Reached target initrd-root-fs.target. Sep 13 00:05:21.969843 systemd[1]: Mounting sysroot-usr.mount... Sep 13 00:05:21.970574 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Sep 13 00:05:21.970612 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 00:05:21.970634 systemd[1]: Reached target ignition-diskful.target. Sep 13 00:05:21.972630 systemd[1]: Mounted sysroot-usr.mount. Sep 13 00:05:21.974046 systemd[1]: Starting initrd-setup-root.service... Sep 13 00:05:21.979954 initrd-setup-root[786]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 00:05:21.984550 initrd-setup-root[795]: cut: /sysroot/etc/group: No such file or directory Sep 13 00:05:21.987983 initrd-setup-root[803]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 00:05:21.991990 initrd-setup-root[811]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 00:05:22.024571 systemd[1]: Finished initrd-setup-root.service. Sep 13 00:05:22.024000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:22.026261 systemd[1]: Starting ignition-mount.service... Sep 13 00:05:22.027415 systemd[1]: Starting sysroot-boot.service... Sep 13 00:05:22.032317 bash[828]: umount: /sysroot/usr/share/oem: not mounted. Sep 13 00:05:22.042425 ignition[830]: INFO : Ignition 2.14.0 Sep 13 00:05:22.042425 ignition[830]: INFO : Stage: mount Sep 13 00:05:22.043776 ignition[830]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:22.043776 ignition[830]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:22.043776 ignition[830]: INFO : mount: mount passed Sep 13 00:05:22.044000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:22.047545 ignition[830]: INFO : Ignition finished successfully Sep 13 00:05:22.047000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:22.044533 systemd[1]: Finished ignition-mount.service. Sep 13 00:05:22.046902 systemd[1]: Finished sysroot-boot.service. Sep 13 00:05:22.627097 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 13 00:05:22.635752 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (838) Sep 13 00:05:22.638152 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 13 00:05:22.638201 kernel: BTRFS info (device vda6): using free space tree Sep 13 00:05:22.638211 kernel: BTRFS info (device vda6): has skinny extents Sep 13 00:05:22.641646 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 13 00:05:22.643099 systemd[1]: Starting ignition-files.service... Sep 13 00:05:22.658009 ignition[858]: INFO : Ignition 2.14.0 Sep 13 00:05:22.658009 ignition[858]: INFO : Stage: files Sep 13 00:05:22.659301 ignition[858]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:22.659301 ignition[858]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:22.659301 ignition[858]: DEBUG : files: compiled without relabeling support, skipping Sep 13 00:05:22.661871 ignition[858]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 00:05:22.661871 ignition[858]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 00:05:22.664266 ignition[858]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 00:05:22.664266 ignition[858]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 00:05:22.666271 ignition[858]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 00:05:22.666271 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 13 00:05:22.666271 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Sep 13 00:05:22.664476 unknown[858]: wrote ssh authorized keys file for user: core Sep 13 00:05:22.712637 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 13 00:05:22.992564 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 13 00:05:22.994399 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:05:22.994399 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Sep 13 00:05:23.204280 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 13 00:05:23.347052 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:05:23.347052 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 13 00:05:23.353186 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Sep 13 00:05:23.556246 systemd-networkd[741]: eth0: Gained IPv6LL Sep 13 00:05:23.633444 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 13 00:05:24.048243 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Sep 13 00:05:24.048243 ignition[858]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(11): [started] setting preset to disabled for "coreos-metadata.service" Sep 13 00:05:24.052156 ignition[858]: INFO : files: op(11): op(12): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 13 00:05:24.108783 ignition[858]: INFO : files: op(11): op(12): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 13 00:05:24.111000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.112883 ignition[858]: INFO : files: op(11): [finished] setting preset to disabled for "coreos-metadata.service" Sep 13 00:05:24.112883 ignition[858]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:05:24.112883 ignition[858]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:05:24.112883 ignition[858]: INFO : files: files passed Sep 13 00:05:24.112883 ignition[858]: INFO : Ignition finished successfully Sep 13 00:05:24.133402 kernel: kauditd_printk_skb: 23 callbacks suppressed Sep 13 00:05:24.133425 kernel: audit: type=1130 audit(1757721924.111:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.133436 kernel: audit: type=1130 audit(1757721924.121:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.133445 kernel: audit: type=1131 audit(1757721924.121:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.133454 kernel: audit: type=1130 audit(1757721924.129:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.121000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.121000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.111511 systemd[1]: Finished ignition-files.service. Sep 13 00:05:24.113287 systemd[1]: Starting initrd-setup-root-after-ignition.service... Sep 13 00:05:24.116745 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Sep 13 00:05:24.138614 initrd-setup-root-after-ignition[884]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Sep 13 00:05:24.117465 systemd[1]: Starting ignition-quench.service... Sep 13 00:05:24.141149 initrd-setup-root-after-ignition[886]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 00:05:24.121108 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 00:05:24.121184 systemd[1]: Finished ignition-quench.service. Sep 13 00:05:24.124642 systemd[1]: Finished initrd-setup-root-after-ignition.service. Sep 13 00:05:24.129593 systemd[1]: Reached target ignition-complete.target. Sep 13 00:05:24.135419 systemd[1]: Starting initrd-parse-etc.service... Sep 13 00:05:24.148841 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 00:05:24.148956 systemd[1]: Finished initrd-parse-etc.service. Sep 13 00:05:24.150374 systemd[1]: Reached target initrd-fs.target. Sep 13 00:05:24.155750 kernel: audit: type=1130 audit(1757721924.150:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.155778 kernel: audit: type=1131 audit(1757721924.150:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.150000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.150000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.155350 systemd[1]: Reached target initrd.target. Sep 13 00:05:24.156655 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Sep 13 00:05:24.157936 systemd[1]: Starting dracut-pre-pivot.service... Sep 13 00:05:24.168163 systemd[1]: Finished dracut-pre-pivot.service. Sep 13 00:05:24.168000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.169529 systemd[1]: Starting initrd-cleanup.service... Sep 13 00:05:24.172200 kernel: audit: type=1130 audit(1757721924.168:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.178513 systemd[1]: Stopped target nss-lookup.target. Sep 13 00:05:24.179839 systemd[1]: Stopped target remote-cryptsetup.target. Sep 13 00:05:24.180540 systemd[1]: Stopped target timers.target. Sep 13 00:05:24.181792 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 00:05:24.182000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.181985 systemd[1]: Stopped dracut-pre-pivot.service. Sep 13 00:05:24.186490 kernel: audit: type=1131 audit(1757721924.182:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.183192 systemd[1]: Stopped target initrd.target. Sep 13 00:05:24.186170 systemd[1]: Stopped target basic.target. Sep 13 00:05:24.187312 systemd[1]: Stopped target ignition-complete.target. Sep 13 00:05:24.188591 systemd[1]: Stopped target ignition-diskful.target. Sep 13 00:05:24.189570 systemd[1]: Stopped target initrd-root-device.target. Sep 13 00:05:24.190973 systemd[1]: Stopped target remote-fs.target. Sep 13 00:05:24.192039 systemd[1]: Stopped target remote-fs-pre.target. Sep 13 00:05:24.193340 systemd[1]: Stopped target sysinit.target. Sep 13 00:05:24.194612 systemd[1]: Stopped target local-fs.target. Sep 13 00:05:24.195802 systemd[1]: Stopped target local-fs-pre.target. Sep 13 00:05:24.197340 systemd[1]: Stopped target swap.target. Sep 13 00:05:24.198000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.198505 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 00:05:24.202764 kernel: audit: type=1131 audit(1757721924.198:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.198700 systemd[1]: Stopped dracut-pre-mount.service. Sep 13 00:05:24.199914 systemd[1]: Stopped target cryptsetup.target. Sep 13 00:05:24.203000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.203357 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 00:05:24.208079 kernel: audit: type=1131 audit(1757721924.203:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.206000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.203460 systemd[1]: Stopped dracut-initqueue.service. Sep 13 00:05:24.204542 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 00:05:24.204629 systemd[1]: Stopped ignition-fetch-offline.service. Sep 13 00:05:24.207551 systemd[1]: Stopped target paths.target. Sep 13 00:05:24.208602 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 00:05:24.212958 systemd[1]: Stopped systemd-ask-password-console.path. Sep 13 00:05:24.213749 systemd[1]: Stopped target slices.target. Sep 13 00:05:24.215077 systemd[1]: Stopped target sockets.target. Sep 13 00:05:24.216064 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 00:05:24.217000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.216184 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Sep 13 00:05:24.218000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.217259 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 00:05:24.217349 systemd[1]: Stopped ignition-files.service. Sep 13 00:05:24.221372 iscsid[748]: iscsid shutting down. Sep 13 00:05:24.219615 systemd[1]: Stopping ignition-mount.service... Sep 13 00:05:24.222000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.220775 systemd[1]: Stopping iscsid.service... Sep 13 00:05:24.221727 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 00:05:24.221837 systemd[1]: Stopped kmod-static-nodes.service. Sep 13 00:05:24.223773 systemd[1]: Stopping sysroot-boot.service... Sep 13 00:05:24.226000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.226000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.224754 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 00:05:24.224898 systemd[1]: Stopped systemd-udev-trigger.service. Sep 13 00:05:24.230632 ignition[899]: INFO : Ignition 2.14.0 Sep 13 00:05:24.230632 ignition[899]: INFO : Stage: umount Sep 13 00:05:24.230632 ignition[899]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 00:05:24.230632 ignition[899]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 13 00:05:24.230632 ignition[899]: INFO : umount: umount passed Sep 13 00:05:24.230632 ignition[899]: INFO : Ignition finished successfully Sep 13 00:05:24.231000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.234000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.226373 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 00:05:24.237000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.237000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.226458 systemd[1]: Stopped dracut-pre-trigger.service. Sep 13 00:05:24.229870 systemd[1]: iscsid.service: Deactivated successfully. Sep 13 00:05:24.241000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.241000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.229955 systemd[1]: Stopped iscsid.service. Sep 13 00:05:24.242000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.232543 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 00:05:24.232622 systemd[1]: Stopped ignition-mount.service. Sep 13 00:05:24.236770 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 00:05:24.237166 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 00:05:24.246000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.237241 systemd[1]: Finished initrd-cleanup.service. Sep 13 00:05:24.238599 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 00:05:24.238631 systemd[1]: Closed iscsid.socket. Sep 13 00:05:24.239956 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 00:05:24.239996 systemd[1]: Stopped ignition-disks.service. Sep 13 00:05:24.241173 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 00:05:24.241209 systemd[1]: Stopped ignition-kargs.service. Sep 13 00:05:24.241908 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 00:05:24.260000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.241942 systemd[1]: Stopped ignition-setup.service. Sep 13 00:05:24.243920 systemd[1]: Stopping iscsiuio.service... Sep 13 00:05:24.246444 systemd[1]: iscsiuio.service: Deactivated successfully. Sep 13 00:05:24.246543 systemd[1]: Stopped iscsiuio.service. Sep 13 00:05:24.247593 systemd[1]: Stopped target network.target. Sep 13 00:05:24.265000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.248776 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 00:05:24.266000 audit: BPF prog-id=6 op=UNLOAD Sep 13 00:05:24.248807 systemd[1]: Closed iscsiuio.socket. Sep 13 00:05:24.250278 systemd[1]: Stopping systemd-networkd.service... Sep 13 00:05:24.251758 systemd[1]: Stopping systemd-resolved.service... Sep 13 00:05:24.270000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.259783 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 00:05:24.272000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.259884 systemd[1]: Stopped systemd-resolved.service. Sep 13 00:05:24.263755 systemd-networkd[741]: eth0: DHCPv6 lease lost Sep 13 00:05:24.273000 audit: BPF prog-id=9 op=UNLOAD Sep 13 00:05:24.274000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.264800 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 00:05:24.264898 systemd[1]: Stopped systemd-networkd.service. Sep 13 00:05:24.266154 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 00:05:24.266184 systemd[1]: Closed systemd-networkd.socket. Sep 13 00:05:24.267920 systemd[1]: Stopping network-cleanup.service... Sep 13 00:05:24.269505 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 00:05:24.269564 systemd[1]: Stopped parse-ip-for-networkd.service. Sep 13 00:05:24.284000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.270820 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:05:24.270868 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:05:24.273652 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 00:05:24.273736 systemd[1]: Stopped systemd-modules-load.service. Sep 13 00:05:24.288000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.278983 systemd[1]: Stopping systemd-udevd.service... Sep 13 00:05:24.280762 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 13 00:05:24.283557 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 00:05:24.291000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.283657 systemd[1]: Stopped network-cleanup.service. Sep 13 00:05:24.292000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.285609 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 00:05:24.293000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.285772 systemd[1]: Stopped systemd-udevd.service. Sep 13 00:05:24.289318 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 00:05:24.289352 systemd[1]: Closed systemd-udevd-control.socket. Sep 13 00:05:24.290186 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 00:05:24.290216 systemd[1]: Closed systemd-udevd-kernel.socket. Sep 13 00:05:24.291246 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 00:05:24.297000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.298000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.299000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.291284 systemd[1]: Stopped dracut-pre-udev.service. Sep 13 00:05:24.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.301000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.292564 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 00:05:24.292597 systemd[1]: Stopped dracut-cmdline.service. Sep 13 00:05:24.293703 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 00:05:24.293747 systemd[1]: Stopped dracut-cmdline-ask.service. Sep 13 00:05:24.295534 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Sep 13 00:05:24.296757 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 00:05:24.296811 systemd[1]: Stopped systemd-vconsole-setup.service. Sep 13 00:05:24.298504 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 00:05:24.298594 systemd[1]: Stopped sysroot-boot.service. Sep 13 00:05:24.299414 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 00:05:24.299452 systemd[1]: Stopped initrd-setup-root.service. Sep 13 00:05:24.301181 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 00:05:24.301262 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Sep 13 00:05:24.302520 systemd[1]: Reached target initrd-switch-root.target. Sep 13 00:05:24.304227 systemd[1]: Starting initrd-switch-root.service... Sep 13 00:05:24.311027 systemd[1]: Switching root. Sep 13 00:05:24.328115 systemd-journald[289]: Journal stopped Sep 13 00:05:26.487049 systemd-journald[289]: Received SIGTERM from PID 1 (systemd). Sep 13 00:05:26.487107 kernel: SELinux: Class mctp_socket not defined in policy. Sep 13 00:05:26.487120 kernel: SELinux: Class anon_inode not defined in policy. Sep 13 00:05:26.487130 kernel: SELinux: the above unknown classes and permissions will be allowed Sep 13 00:05:26.487140 kernel: SELinux: policy capability network_peer_controls=1 Sep 13 00:05:26.487151 kernel: SELinux: policy capability open_perms=1 Sep 13 00:05:26.487161 kernel: SELinux: policy capability extended_socket_class=1 Sep 13 00:05:26.487175 kernel: SELinux: policy capability always_check_network=0 Sep 13 00:05:26.487186 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 13 00:05:26.487217 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 13 00:05:26.487229 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 13 00:05:26.487239 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 13 00:05:26.487252 systemd[1]: Successfully loaded SELinux policy in 36.782ms. Sep 13 00:05:26.487269 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.922ms. Sep 13 00:05:26.487283 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:05:26.487295 systemd[1]: Detected virtualization kvm. Sep 13 00:05:26.487306 systemd[1]: Detected architecture arm64. Sep 13 00:05:26.487319 systemd[1]: Detected first boot. Sep 13 00:05:26.487330 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:05:26.487344 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Sep 13 00:05:26.487354 systemd[1]: Populated /etc with preset unit settings. Sep 13 00:05:26.487365 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:05:26.487377 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:05:26.487392 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:05:26.487407 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 13 00:05:26.487418 systemd[1]: Stopped initrd-switch-root.service. Sep 13 00:05:26.487429 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 13 00:05:26.487439 systemd[1]: Created slice system-addon\x2dconfig.slice. Sep 13 00:05:26.487450 systemd[1]: Created slice system-addon\x2drun.slice. Sep 13 00:05:26.487460 systemd[1]: Created slice system-getty.slice. Sep 13 00:05:26.487473 systemd[1]: Created slice system-modprobe.slice. Sep 13 00:05:26.487483 systemd[1]: Created slice system-serial\x2dgetty.slice. Sep 13 00:05:26.487494 systemd[1]: Created slice system-system\x2dcloudinit.slice. Sep 13 00:05:26.487505 systemd[1]: Created slice system-systemd\x2dfsck.slice. Sep 13 00:05:26.487515 systemd[1]: Created slice user.slice. Sep 13 00:05:26.487526 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:05:26.487536 systemd[1]: Started systemd-ask-password-wall.path. Sep 13 00:05:26.487546 systemd[1]: Set up automount boot.automount. Sep 13 00:05:26.487557 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Sep 13 00:05:26.487568 systemd[1]: Stopped target initrd-switch-root.target. Sep 13 00:05:26.487579 systemd[1]: Stopped target initrd-fs.target. Sep 13 00:05:26.487589 systemd[1]: Stopped target initrd-root-fs.target. Sep 13 00:05:26.487600 systemd[1]: Reached target integritysetup.target. Sep 13 00:05:26.487611 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:05:26.487622 systemd[1]: Reached target remote-fs.target. Sep 13 00:05:26.487632 systemd[1]: Reached target slices.target. Sep 13 00:05:26.487643 systemd[1]: Reached target swap.target. Sep 13 00:05:26.487663 systemd[1]: Reached target torcx.target. Sep 13 00:05:26.487676 systemd[1]: Reached target veritysetup.target. Sep 13 00:05:26.487686 systemd[1]: Listening on systemd-coredump.socket. Sep 13 00:05:26.487701 systemd[1]: Listening on systemd-initctl.socket. Sep 13 00:05:26.487712 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:05:26.487730 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:05:26.487742 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:05:26.487753 systemd[1]: Listening on systemd-userdbd.socket. Sep 13 00:05:26.487764 systemd[1]: Mounting dev-hugepages.mount... Sep 13 00:05:26.487775 systemd[1]: Mounting dev-mqueue.mount... Sep 13 00:05:26.487786 systemd[1]: Mounting media.mount... Sep 13 00:05:26.487796 systemd[1]: Mounting sys-kernel-debug.mount... Sep 13 00:05:26.487807 systemd[1]: Mounting sys-kernel-tracing.mount... Sep 13 00:05:26.487817 systemd[1]: Mounting tmp.mount... Sep 13 00:05:26.487827 systemd[1]: Starting flatcar-tmpfiles.service... Sep 13 00:05:26.487837 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:05:26.487848 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:05:26.487858 systemd[1]: Starting modprobe@configfs.service... Sep 13 00:05:26.487870 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:05:26.487882 systemd[1]: Starting modprobe@drm.service... Sep 13 00:05:26.487897 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:05:26.487910 systemd[1]: Starting modprobe@fuse.service... Sep 13 00:05:26.487920 systemd[1]: Starting modprobe@loop.service... Sep 13 00:05:26.487932 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 00:05:26.487942 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 13 00:05:26.487952 systemd[1]: Stopped systemd-fsck-root.service. Sep 13 00:05:26.487962 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 13 00:05:26.487975 systemd[1]: Stopped systemd-fsck-usr.service. Sep 13 00:05:26.487985 systemd[1]: Stopped systemd-journald.service. Sep 13 00:05:26.487995 systemd[1]: Starting systemd-journald.service... Sep 13 00:05:26.488005 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:05:26.488015 kernel: fuse: init (API version 7.34) Sep 13 00:05:26.488025 systemd[1]: Starting systemd-network-generator.service... Sep 13 00:05:26.488035 kernel: loop: module loaded Sep 13 00:05:26.488046 systemd[1]: Starting systemd-remount-fs.service... Sep 13 00:05:26.488056 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:05:26.488067 systemd[1]: verity-setup.service: Deactivated successfully. Sep 13 00:05:26.488079 systemd[1]: Stopped verity-setup.service. Sep 13 00:05:26.488093 systemd[1]: Mounted dev-hugepages.mount. Sep 13 00:05:26.488107 systemd[1]: Mounted dev-mqueue.mount. Sep 13 00:05:26.488128 systemd[1]: Mounted media.mount. Sep 13 00:05:26.488139 systemd[1]: Mounted sys-kernel-debug.mount. Sep 13 00:05:26.488149 systemd[1]: Mounted sys-kernel-tracing.mount. Sep 13 00:05:26.488159 systemd[1]: Mounted tmp.mount. Sep 13 00:05:26.488170 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:05:26.488180 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 13 00:05:26.488192 systemd[1]: Finished modprobe@configfs.service. Sep 13 00:05:26.488202 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:05:26.488215 systemd-journald[994]: Journal started Sep 13 00:05:26.488261 systemd-journald[994]: Runtime Journal (/run/log/journal/120c733ea798469392b581a2dcb84651) is 6.0M, max 48.7M, 42.6M free. Sep 13 00:05:24.392000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 13 00:05:24.504000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:05:24.504000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:05:24.504000 audit: BPF prog-id=10 op=LOAD Sep 13 00:05:24.504000 audit: BPF prog-id=10 op=UNLOAD Sep 13 00:05:24.504000 audit: BPF prog-id=11 op=LOAD Sep 13 00:05:24.504000 audit: BPF prog-id=11 op=UNLOAD Sep 13 00:05:24.561000 audit[932]: AVC avc: denied { associate } for pid=932 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Sep 13 00:05:24.561000 audit[932]: SYSCALL arch=c00000b7 syscall=5 success=yes exit=0 a0=40001c58b4 a1=40000c8de0 a2=40000cf0c0 a3=32 items=0 ppid=915 pid=932 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:05:24.561000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 13 00:05:24.563000 audit[932]: AVC avc: denied { associate } for pid=932 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Sep 13 00:05:24.563000 audit[932]: SYSCALL arch=c00000b7 syscall=34 success=yes exit=0 a0=ffffffffffffff9c a1=40001c5989 a2=1ed a3=0 items=2 ppid=915 pid=932 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:05:24.563000 audit: CWD cwd="/" Sep 13 00:05:24.563000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:05:24.563000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:05:24.563000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 13 00:05:26.372000 audit: BPF prog-id=12 op=LOAD Sep 13 00:05:26.372000 audit: BPF prog-id=3 op=UNLOAD Sep 13 00:05:26.373000 audit: BPF prog-id=13 op=LOAD Sep 13 00:05:26.373000 audit: BPF prog-id=14 op=LOAD Sep 13 00:05:26.373000 audit: BPF prog-id=4 op=UNLOAD Sep 13 00:05:26.373000 audit: BPF prog-id=5 op=UNLOAD Sep 13 00:05:26.374000 audit: BPF prog-id=15 op=LOAD Sep 13 00:05:26.374000 audit: BPF prog-id=12 op=UNLOAD Sep 13 00:05:26.374000 audit: BPF prog-id=16 op=LOAD Sep 13 00:05:26.374000 audit: BPF prog-id=17 op=LOAD Sep 13 00:05:26.374000 audit: BPF prog-id=13 op=UNLOAD Sep 13 00:05:26.374000 audit: BPF prog-id=14 op=UNLOAD Sep 13 00:05:26.375000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.489074 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:05:26.377000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.377000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.385000 audit: BPF prog-id=15 op=UNLOAD Sep 13 00:05:26.455000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.457000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.457000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.457000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.458000 audit: BPF prog-id=18 op=LOAD Sep 13 00:05:26.459000 audit: BPF prog-id=19 op=LOAD Sep 13 00:05:26.459000 audit: BPF prog-id=20 op=LOAD Sep 13 00:05:26.459000 audit: BPF prog-id=16 op=UNLOAD Sep 13 00:05:26.459000 audit: BPF prog-id=17 op=UNLOAD Sep 13 00:05:26.474000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.484000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.485000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Sep 13 00:05:26.485000 audit[994]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=6 a1=ffffc4d2cd60 a2=4000 a3=1 items=0 ppid=1 pid=994 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:05:26.485000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Sep 13 00:05:26.487000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.559577 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:05:26.372093 systemd[1]: Queued start job for default target multi-user.target. Sep 13 00:05:24.559877 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 13 00:05:26.372106 systemd[1]: Unnecessary job was removed for dev-vda6.device. Sep 13 00:05:26.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.489000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:24.559896 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 13 00:05:26.375712 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 13 00:05:24.559926 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Sep 13 00:05:24.559954 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="skipped missing lower profile" missing profile=oem Sep 13 00:05:24.560035 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Sep 13 00:05:24.560050 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Sep 13 00:05:24.560304 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Sep 13 00:05:24.560341 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 13 00:05:24.560353 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 13 00:05:24.561936 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Sep 13 00:05:24.561980 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Sep 13 00:05:24.562000 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.8: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.8 Sep 13 00:05:26.490915 systemd[1]: Started systemd-journald.service. Sep 13 00:05:24.562015 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Sep 13 00:05:24.562035 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.8: no such file or directory" path=/var/lib/torcx/store/3510.3.8 Sep 13 00:05:24.562048 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:24Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Sep 13 00:05:26.490000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.096888 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:05:26.097153 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:05:26.097261 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:05:26.097428 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:05:26.097484 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Sep 13 00:05:26.097542 /usr/lib/systemd/system-generators/torcx-generator[932]: time="2025-09-13T00:05:26Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Sep 13 00:05:26.491691 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:05:26.491997 systemd[1]: Finished modprobe@drm.service. Sep 13 00:05:26.491000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.491000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.492874 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:05:26.493037 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:05:26.492000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.492000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.493936 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 13 00:05:26.496002 systemd[1]: Finished modprobe@fuse.service. Sep 13 00:05:26.495000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.495000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.496884 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:05:26.497031 systemd[1]: Finished modprobe@loop.service. Sep 13 00:05:26.496000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.496000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.498043 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:05:26.497000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.499082 systemd[1]: Finished systemd-network-generator.service. Sep 13 00:05:26.499000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.500218 systemd[1]: Finished systemd-remount-fs.service. Sep 13 00:05:26.500000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.501305 systemd[1]: Reached target network-pre.target. Sep 13 00:05:26.503071 systemd[1]: Mounting sys-fs-fuse-connections.mount... Sep 13 00:05:26.504996 systemd[1]: Mounting sys-kernel-config.mount... Sep 13 00:05:26.505630 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 00:05:26.507871 systemd[1]: Starting systemd-hwdb-update.service... Sep 13 00:05:26.509805 systemd[1]: Starting systemd-journal-flush.service... Sep 13 00:05:26.510459 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:05:26.511492 systemd[1]: Starting systemd-random-seed.service... Sep 13 00:05:26.512446 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:05:26.513440 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:05:26.517000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.516507 systemd[1]: Finished flatcar-tmpfiles.service. Sep 13 00:05:26.517393 systemd[1]: Mounted sys-fs-fuse-connections.mount. Sep 13 00:05:26.521317 systemd-journald[994]: Time spent on flushing to /var/log/journal/120c733ea798469392b581a2dcb84651 is 19.453ms for 998 entries. Sep 13 00:05:26.521317 systemd-journald[994]: System Journal (/var/log/journal/120c733ea798469392b581a2dcb84651) is 8.0M, max 195.6M, 187.6M free. Sep 13 00:05:26.556047 systemd-journald[994]: Received client request to flush runtime journal. Sep 13 00:05:26.529000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.530000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.542000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.546000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.518453 systemd[1]: Mounted sys-kernel-config.mount. Sep 13 00:05:26.520228 systemd[1]: Starting systemd-sysusers.service... Sep 13 00:05:26.528962 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:05:26.556703 udevadm[1034]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 13 00:05:26.530046 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:05:26.531907 systemd[1]: Starting systemd-udev-settle.service... Sep 13 00:05:26.541125 systemd[1]: Finished systemd-random-seed.service. Sep 13 00:05:26.543915 systemd[1]: Reached target first-boot-complete.target. Sep 13 00:05:26.545987 systemd[1]: Finished systemd-sysusers.service. Sep 13 00:05:26.557007 systemd[1]: Finished systemd-journal-flush.service. Sep 13 00:05:26.557000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.928554 systemd[1]: Finished systemd-hwdb-update.service. Sep 13 00:05:26.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.929000 audit: BPF prog-id=21 op=LOAD Sep 13 00:05:26.929000 audit: BPF prog-id=22 op=LOAD Sep 13 00:05:26.929000 audit: BPF prog-id=7 op=UNLOAD Sep 13 00:05:26.929000 audit: BPF prog-id=8 op=UNLOAD Sep 13 00:05:26.930905 systemd[1]: Starting systemd-udevd.service... Sep 13 00:05:26.948258 systemd-udevd[1036]: Using default interface naming scheme 'v252'. Sep 13 00:05:26.969186 systemd[1]: Started systemd-udevd.service. Sep 13 00:05:26.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:26.974000 audit: BPF prog-id=23 op=LOAD Sep 13 00:05:26.975061 systemd[1]: Starting systemd-networkd.service... Sep 13 00:05:26.980000 audit: BPF prog-id=24 op=LOAD Sep 13 00:05:26.981000 audit: BPF prog-id=25 op=LOAD Sep 13 00:05:26.981000 audit: BPF prog-id=26 op=LOAD Sep 13 00:05:26.982927 systemd[1]: Starting systemd-userdbd.service... Sep 13 00:05:27.007563 systemd[1]: Started systemd-userdbd.service. Sep 13 00:05:27.007000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.020125 systemd[1]: Condition check resulted in dev-ttyAMA0.device being skipped. Sep 13 00:05:27.060927 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:05:27.062138 systemd-networkd[1044]: lo: Link UP Sep 13 00:05:27.062143 systemd-networkd[1044]: lo: Gained carrier Sep 13 00:05:27.062566 systemd-networkd[1044]: Enumeration completed Sep 13 00:05:27.062679 systemd[1]: Started systemd-networkd.service. Sep 13 00:05:27.062000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.063477 systemd-networkd[1044]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 00:05:27.065017 systemd-networkd[1044]: eth0: Link UP Sep 13 00:05:27.065023 systemd-networkd[1044]: eth0: Gained carrier Sep 13 00:05:27.092891 systemd-networkd[1044]: eth0: DHCPv4 address 10.0.0.20/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 13 00:05:27.101197 systemd[1]: Finished systemd-udev-settle.service. Sep 13 00:05:27.101000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.103996 systemd[1]: Starting lvm2-activation-early.service... Sep 13 00:05:27.114246 lvm[1069]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:05:27.138583 systemd[1]: Finished lvm2-activation-early.service. Sep 13 00:05:27.139504 systemd[1]: Reached target cryptsetup.target. Sep 13 00:05:27.138000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.141541 systemd[1]: Starting lvm2-activation.service... Sep 13 00:05:27.144941 lvm[1070]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:05:27.176639 systemd[1]: Finished lvm2-activation.service. Sep 13 00:05:27.176000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.177529 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:05:27.178431 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 13 00:05:27.178461 systemd[1]: Reached target local-fs.target. Sep 13 00:05:27.179137 systemd[1]: Reached target machines.target. Sep 13 00:05:27.181031 systemd[1]: Starting ldconfig.service... Sep 13 00:05:27.182146 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.182206 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.183313 systemd[1]: Starting systemd-boot-update.service... Sep 13 00:05:27.185285 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Sep 13 00:05:27.187675 systemd[1]: Starting systemd-machine-id-commit.service... Sep 13 00:05:27.190867 systemd[1]: Starting systemd-sysext.service... Sep 13 00:05:27.192088 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1072 (bootctl) Sep 13 00:05:27.195835 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Sep 13 00:05:27.201514 systemd[1]: Unmounting usr-share-oem.mount... Sep 13 00:05:27.207306 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Sep 13 00:05:27.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.209967 systemd[1]: usr-share-oem.mount: Deactivated successfully. Sep 13 00:05:27.210167 systemd[1]: Unmounted usr-share-oem.mount. Sep 13 00:05:27.224804 kernel: loop0: detected capacity change from 0 to 203944 Sep 13 00:05:27.405751 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 13 00:05:27.423760 kernel: loop1: detected capacity change from 0 to 203944 Sep 13 00:05:27.440851 (sd-sysext)[1085]: Using extensions 'kubernetes'. Sep 13 00:05:27.441202 (sd-sysext)[1085]: Merged extensions into '/usr'. Sep 13 00:05:27.443370 systemd-fsck[1082]: fsck.fat 4.2 (2021-01-31) Sep 13 00:05:27.443370 systemd-fsck[1082]: /dev/vda1: 236 files, 117310/258078 clusters Sep 13 00:05:27.451041 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Sep 13 00:05:27.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.455272 systemd[1]: Finished systemd-machine-id-commit.service. Sep 13 00:05:27.455000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.462610 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.464107 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:05:27.466253 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:05:27.468042 systemd[1]: Starting modprobe@loop.service... Sep 13 00:05:27.468709 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.468850 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.469610 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:05:27.469881 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:05:27.470000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.470000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.471238 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:05:27.471349 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:05:27.471000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.471000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.472757 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:05:27.472880 systemd[1]: Finished modprobe@loop.service. Sep 13 00:05:27.473000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.473000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.474119 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:05:27.474221 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.476178 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 13 00:05:27.477958 systemd[1]: Mounting boot.mount... Sep 13 00:05:27.480070 systemd[1]: Mounting usr-share-oem.mount... Sep 13 00:05:27.485586 systemd[1]: Mounted usr-share-oem.mount. Sep 13 00:05:27.489107 systemd[1]: Finished systemd-sysext.service. Sep 13 00:05:27.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.490197 systemd[1]: Mounted boot.mount. Sep 13 00:05:27.492321 systemd[1]: Starting ensure-sysext.service... Sep 13 00:05:27.494439 systemd[1]: Starting systemd-tmpfiles-setup.service... Sep 13 00:05:27.498716 systemd[1]: Finished systemd-boot-update.service. Sep 13 00:05:27.498000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.503432 systemd[1]: Reloading. Sep 13 00:05:27.510502 systemd-tmpfiles[1093]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Sep 13 00:05:27.517240 systemd-tmpfiles[1093]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 13 00:05:27.530077 systemd-tmpfiles[1093]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 13 00:05:27.547610 /usr/lib/systemd/system-generators/torcx-generator[1115]: time="2025-09-13T00:05:27Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:05:27.547652 /usr/lib/systemd/system-generators/torcx-generator[1115]: time="2025-09-13T00:05:27Z" level=info msg="torcx already run" Sep 13 00:05:27.577613 ldconfig[1071]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 13 00:05:27.622498 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:05:27.622520 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:05:27.638675 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:05:27.682000 audit: BPF prog-id=27 op=LOAD Sep 13 00:05:27.682000 audit: BPF prog-id=18 op=UNLOAD Sep 13 00:05:27.682000 audit: BPF prog-id=28 op=LOAD Sep 13 00:05:27.682000 audit: BPF prog-id=29 op=LOAD Sep 13 00:05:27.682000 audit: BPF prog-id=19 op=UNLOAD Sep 13 00:05:27.682000 audit: BPF prog-id=20 op=UNLOAD Sep 13 00:05:27.682000 audit: BPF prog-id=30 op=LOAD Sep 13 00:05:27.682000 audit: BPF prog-id=24 op=UNLOAD Sep 13 00:05:27.683000 audit: BPF prog-id=31 op=LOAD Sep 13 00:05:27.683000 audit: BPF prog-id=32 op=LOAD Sep 13 00:05:27.683000 audit: BPF prog-id=25 op=UNLOAD Sep 13 00:05:27.683000 audit: BPF prog-id=26 op=UNLOAD Sep 13 00:05:27.683000 audit: BPF prog-id=33 op=LOAD Sep 13 00:05:27.683000 audit: BPF prog-id=34 op=LOAD Sep 13 00:05:27.683000 audit: BPF prog-id=21 op=UNLOAD Sep 13 00:05:27.683000 audit: BPF prog-id=22 op=UNLOAD Sep 13 00:05:27.684000 audit: BPF prog-id=35 op=LOAD Sep 13 00:05:27.684000 audit: BPF prog-id=23 op=UNLOAD Sep 13 00:05:27.688853 systemd[1]: Finished ldconfig.service. Sep 13 00:05:27.688000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.690858 systemd[1]: Finished systemd-tmpfiles-setup.service. Sep 13 00:05:27.690000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.694609 systemd[1]: Starting audit-rules.service... Sep 13 00:05:27.696621 systemd[1]: Starting clean-ca-certificates.service... Sep 13 00:05:27.698751 systemd[1]: Starting systemd-journal-catalog-update.service... Sep 13 00:05:27.699000 audit: BPF prog-id=36 op=LOAD Sep 13 00:05:27.701445 systemd[1]: Starting systemd-resolved.service... Sep 13 00:05:27.702000 audit: BPF prog-id=37 op=LOAD Sep 13 00:05:27.704145 systemd[1]: Starting systemd-timesyncd.service... Sep 13 00:05:27.706927 systemd[1]: Starting systemd-update-utmp.service... Sep 13 00:05:27.709000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.708558 systemd[1]: Finished clean-ca-certificates.service. Sep 13 00:05:27.711749 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:05:27.714533 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.716089 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:05:27.718168 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:05:27.717000 audit[1163]: SYSTEM_BOOT pid=1163 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.720109 systemd[1]: Starting modprobe@loop.service... Sep 13 00:05:27.720869 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.721006 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.721098 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:05:27.721969 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:05:27.722131 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:05:27.722000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.722000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.723302 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:05:27.723433 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:05:27.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.724000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.724681 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:05:27.724815 systemd[1]: Finished modprobe@loop.service. Sep 13 00:05:27.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.724000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.727502 systemd[1]: Finished systemd-journal-catalog-update.service. Sep 13 00:05:27.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.730509 systemd[1]: Finished systemd-update-utmp.service. Sep 13 00:05:27.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.732897 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.734450 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:05:27.736964 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:05:27.739178 systemd[1]: Starting modprobe@loop.service... Sep 13 00:05:27.739858 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.739986 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.741447 systemd[1]: Starting systemd-update-done.service... Sep 13 00:05:27.742310 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:05:27.743603 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:05:27.743766 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:05:27.743000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.743000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:05:27.744000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Sep 13 00:05:27.744000 audit[1177]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffecdd8cd0 a2=420 a3=0 items=0 ppid=1152 pid=1177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:05:27.744000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Sep 13 00:05:27.745015 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:05:27.745147 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:05:27.745354 augenrules[1177]: No rules Sep 13 00:05:27.746256 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:05:27.746385 systemd[1]: Finished modprobe@loop.service. Sep 13 00:05:27.749555 systemd[1]: Finished systemd-update-done.service. Sep 13 00:05:27.750913 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.752328 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:05:27.754351 systemd[1]: Starting modprobe@drm.service... Sep 13 00:05:27.756281 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:05:27.758132 systemd[1]: Starting modprobe@loop.service... Sep 13 00:05:27.759051 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.759187 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.760608 systemd[1]: Starting systemd-networkd-wait-online.service... Sep 13 00:05:27.761663 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:05:27.762900 systemd[1]: Finished audit-rules.service. Sep 13 00:05:27.763959 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:05:27.764091 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:05:27.765199 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:05:27.765334 systemd[1]: Finished modprobe@drm.service. Sep 13 00:05:27.766612 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:05:27.766769 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:05:27.766866 systemd-resolved[1156]: Positive Trust Anchors: Sep 13 00:05:27.766877 systemd-resolved[1156]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:05:27.766908 systemd-resolved[1156]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:05:27.768036 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:05:27.768161 systemd[1]: Finished modprobe@loop.service. Sep 13 00:05:27.769559 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:05:27.769649 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.771017 systemd[1]: Finished ensure-sysext.service. Sep 13 00:05:27.780532 systemd-resolved[1156]: Defaulting to hostname 'linux'. Sep 13 00:05:27.781556 systemd[1]: Started systemd-timesyncd.service. Sep 13 00:05:27.782342 systemd-timesyncd[1160]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 13 00:05:27.782392 systemd-timesyncd[1160]: Initial clock synchronization to Sat 2025-09-13 00:05:27.915010 UTC. Sep 13 00:05:27.782706 systemd[1]: Started systemd-resolved.service. Sep 13 00:05:27.783384 systemd[1]: Reached target network.target. Sep 13 00:05:27.784088 systemd[1]: Reached target nss-lookup.target. Sep 13 00:05:27.784697 systemd[1]: Reached target sysinit.target. Sep 13 00:05:27.785479 systemd[1]: Started motdgen.path. Sep 13 00:05:27.786077 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Sep 13 00:05:27.786903 systemd[1]: Started systemd-tmpfiles-clean.timer. Sep 13 00:05:27.787563 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 13 00:05:27.787593 systemd[1]: Reached target paths.target. Sep 13 00:05:27.788227 systemd[1]: Reached target time-set.target. Sep 13 00:05:27.789035 systemd[1]: Started logrotate.timer. Sep 13 00:05:27.789681 systemd[1]: Started mdadm.timer. Sep 13 00:05:27.790256 systemd[1]: Reached target timers.target. Sep 13 00:05:27.791158 systemd[1]: Listening on dbus.socket. Sep 13 00:05:27.792970 systemd[1]: Starting docker.socket... Sep 13 00:05:27.796567 systemd[1]: Listening on sshd.socket. Sep 13 00:05:27.797398 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.797927 systemd[1]: Listening on docker.socket. Sep 13 00:05:27.798596 systemd[1]: Reached target sockets.target. Sep 13 00:05:27.799279 systemd[1]: Reached target basic.target. Sep 13 00:05:27.799926 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.799956 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:05:27.801073 systemd[1]: Starting containerd.service... Sep 13 00:05:27.802795 systemd[1]: Starting dbus.service... Sep 13 00:05:27.804584 systemd[1]: Starting enable-oem-cloudinit.service... Sep 13 00:05:27.806872 systemd[1]: Starting extend-filesystems.service... Sep 13 00:05:27.807767 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Sep 13 00:05:27.808998 systemd[1]: Starting motdgen.service... Sep 13 00:05:27.810661 jq[1195]: false Sep 13 00:05:27.810948 systemd[1]: Starting prepare-helm.service... Sep 13 00:05:27.812914 systemd[1]: Starting ssh-key-proc-cmdline.service... Sep 13 00:05:27.815119 systemd[1]: Starting sshd-keygen.service... Sep 13 00:05:27.818155 systemd[1]: Starting systemd-logind.service... Sep 13 00:05:27.818825 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:05:27.818909 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 13 00:05:27.819327 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 13 00:05:27.820146 systemd[1]: Starting update-engine.service... Sep 13 00:05:27.821913 systemd[1]: Starting update-ssh-keys-after-ignition.service... Sep 13 00:05:27.825835 jq[1209]: true Sep 13 00:05:27.827715 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 13 00:05:27.827934 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Sep 13 00:05:27.828969 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 13 00:05:27.829133 systemd[1]: Finished ssh-key-proc-cmdline.service. Sep 13 00:05:27.835778 extend-filesystems[1196]: Found loop1 Sep 13 00:05:27.836931 jq[1217]: true Sep 13 00:05:27.837844 dbus-daemon[1194]: [system] SELinux support is enabled Sep 13 00:05:27.838081 systemd[1]: Started dbus.service. Sep 13 00:05:27.840662 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 13 00:05:27.840709 systemd[1]: Reached target system-config.target. Sep 13 00:05:27.841403 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 13 00:05:27.841420 systemd[1]: Reached target user-config.target. Sep 13 00:05:27.842429 extend-filesystems[1196]: Found vda Sep 13 00:05:27.842442 systemd[1]: motdgen.service: Deactivated successfully. Sep 13 00:05:27.842610 systemd[1]: Finished motdgen.service. Sep 13 00:05:27.843422 extend-filesystems[1196]: Found vda1 Sep 13 00:05:27.848874 extend-filesystems[1196]: Found vda2 Sep 13 00:05:27.849526 extend-filesystems[1196]: Found vda3 Sep 13 00:05:27.850110 tar[1215]: linux-arm64/helm Sep 13 00:05:27.850368 extend-filesystems[1196]: Found usr Sep 13 00:05:27.851101 extend-filesystems[1196]: Found vda4 Sep 13 00:05:27.851928 extend-filesystems[1196]: Found vda6 Sep 13 00:05:27.851928 extend-filesystems[1196]: Found vda7 Sep 13 00:05:27.851928 extend-filesystems[1196]: Found vda9 Sep 13 00:05:27.851928 extend-filesystems[1196]: Checking size of /dev/vda9 Sep 13 00:05:27.869747 systemd-logind[1206]: Watching system buttons on /dev/input/event0 (Power Button) Sep 13 00:05:27.869936 systemd-logind[1206]: New seat seat0. Sep 13 00:05:27.872471 extend-filesystems[1196]: Resized partition /dev/vda9 Sep 13 00:05:27.882148 extend-filesystems[1245]: resize2fs 1.46.5 (30-Dec-2021) Sep 13 00:05:27.881946 systemd[1]: Finished update-ssh-keys-after-ignition.service. Sep 13 00:05:27.883707 bash[1241]: Updated "/home/core/.ssh/authorized_keys" Sep 13 00:05:27.886099 systemd[1]: Started systemd-logind.service. Sep 13 00:05:27.888754 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 13 00:05:27.899376 update_engine[1208]: I0913 00:05:27.899041 1208 main.cc:92] Flatcar Update Engine starting Sep 13 00:05:27.902478 systemd[1]: Started update-engine.service. Sep 13 00:05:27.902806 update_engine[1208]: I0913 00:05:27.902771 1208 update_check_scheduler.cc:74] Next update check in 4m23s Sep 13 00:05:27.905523 systemd[1]: Started locksmithd.service. Sep 13 00:05:27.909080 env[1216]: time="2025-09-13T00:05:27.909024640Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Sep 13 00:05:27.922759 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 13 00:05:27.927938 env[1216]: time="2025-09-13T00:05:27.927890880Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 13 00:05:27.937898 env[1216]: time="2025-09-13T00:05:27.937858720Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.938153 extend-filesystems[1245]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 13 00:05:27.938153 extend-filesystems[1245]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 13 00:05:27.938153 extend-filesystems[1245]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 13 00:05:27.943366 extend-filesystems[1196]: Resized filesystem in /dev/vda9 Sep 13 00:05:27.940476 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942280560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.192-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942324640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942620400Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942649960Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942663760Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942674520Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.942776440Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.943101400Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.943251960Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:05:27.944802 env[1216]: time="2025-09-13T00:05:27.943269280Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 13 00:05:27.940666 systemd[1]: Finished extend-filesystems.service. Sep 13 00:05:27.945075 env[1216]: time="2025-09-13T00:05:27.943338840Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Sep 13 00:05:27.945075 env[1216]: time="2025-09-13T00:05:27.943361760Z" level=info msg="metadata content store policy set" policy=shared Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948555040Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948603920Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948618640Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948660560Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948675760Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948693200Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.948706720Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949064200Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949119240Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949136240Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949150720Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949163640Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949308920Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 13 00:05:27.951206 env[1216]: time="2025-09-13T00:05:27.949419840Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949674080Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949702000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949715920Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949845120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949859200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949871840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949883040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949910480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949924360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949935360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949946080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.949957960Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.950076440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.950095480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951544 env[1216]: time="2025-09-13T00:05:27.950107960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951852 env[1216]: time="2025-09-13T00:05:27.950119760Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 13 00:05:27.951852 env[1216]: time="2025-09-13T00:05:27.950133880Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Sep 13 00:05:27.951852 env[1216]: time="2025-09-13T00:05:27.950144720Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 13 00:05:27.951852 env[1216]: time="2025-09-13T00:05:27.950161480Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Sep 13 00:05:27.951852 env[1216]: time="2025-09-13T00:05:27.950196280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.950381880Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.950437960Z" level=info msg="Connect containerd service" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.950466640Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951055840Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951242600Z" level=info msg="Start subscribing containerd event" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951294200Z" level=info msg="Start recovering state" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951360240Z" level=info msg="Start event monitor" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951378720Z" level=info msg="Start snapshots syncer" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951390000Z" level=info msg="Start cni network conf syncer for default" Sep 13 00:05:27.951948 env[1216]: time="2025-09-13T00:05:27.951398280Z" level=info msg="Start streaming server" Sep 13 00:05:27.952755 env[1216]: time="2025-09-13T00:05:27.952397800Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 13 00:05:27.952837 env[1216]: time="2025-09-13T00:05:27.952821400Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 13 00:05:27.952940 env[1216]: time="2025-09-13T00:05:27.952927160Z" level=info msg="containerd successfully booted in 0.058538s" Sep 13 00:05:27.953011 systemd[1]: Started containerd.service. Sep 13 00:05:27.968171 locksmithd[1248]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 13 00:05:28.272421 tar[1215]: linux-arm64/LICENSE Sep 13 00:05:28.272421 tar[1215]: linux-arm64/README.md Sep 13 00:05:28.279929 systemd[1]: Finished prepare-helm.service. Sep 13 00:05:28.867965 systemd-networkd[1044]: eth0: Gained IPv6LL Sep 13 00:05:28.869612 systemd[1]: Finished systemd-networkd-wait-online.service. Sep 13 00:05:28.870787 systemd[1]: Reached target network-online.target. Sep 13 00:05:28.873380 systemd[1]: Starting kubelet.service... Sep 13 00:05:29.531003 systemd[1]: Started kubelet.service. Sep 13 00:05:29.949432 kubelet[1262]: E0913 00:05:29.949282 1262 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:05:29.951180 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:05:29.951309 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:05:30.460027 sshd_keygen[1220]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 13 00:05:30.480307 systemd[1]: Finished sshd-keygen.service. Sep 13 00:05:30.483200 systemd[1]: Starting issuegen.service... Sep 13 00:05:30.488112 systemd[1]: issuegen.service: Deactivated successfully. Sep 13 00:05:30.488391 systemd[1]: Finished issuegen.service. Sep 13 00:05:30.491339 systemd[1]: Starting systemd-user-sessions.service... Sep 13 00:05:30.497715 systemd[1]: Finished systemd-user-sessions.service. Sep 13 00:05:30.500485 systemd[1]: Started getty@tty1.service. Sep 13 00:05:30.503949 systemd[1]: Started serial-getty@ttyAMA0.service. Sep 13 00:05:30.505080 systemd[1]: Reached target getty.target. Sep 13 00:05:30.506901 systemd[1]: Reached target multi-user.target. Sep 13 00:05:30.509886 systemd[1]: Starting systemd-update-utmp-runlevel.service... Sep 13 00:05:30.516964 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Sep 13 00:05:30.517436 systemd[1]: Finished systemd-update-utmp-runlevel.service. Sep 13 00:05:30.518669 systemd[1]: Startup finished in 579ms (kernel) + 4.792s (initrd) + 6.166s (userspace) = 11.537s. Sep 13 00:05:31.997200 systemd[1]: Created slice system-sshd.slice. Sep 13 00:05:31.998353 systemd[1]: Started sshd@0-10.0.0.20:22-10.0.0.1:54324.service. Sep 13 00:05:32.054020 sshd[1284]: Accepted publickey for core from 10.0.0.1 port 54324 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:05:32.056010 sshd[1284]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.070242 systemd-logind[1206]: New session 1 of user core. Sep 13 00:05:32.071463 systemd[1]: Created slice user-500.slice. Sep 13 00:05:32.072956 systemd[1]: Starting user-runtime-dir@500.service... Sep 13 00:05:32.082812 systemd[1]: Finished user-runtime-dir@500.service. Sep 13 00:05:32.084080 systemd[1]: Starting user@500.service... Sep 13 00:05:32.088422 (systemd)[1287]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.159483 systemd[1287]: Queued start job for default target default.target. Sep 13 00:05:32.159990 systemd[1287]: Reached target paths.target. Sep 13 00:05:32.160022 systemd[1287]: Reached target sockets.target. Sep 13 00:05:32.160033 systemd[1287]: Reached target timers.target. Sep 13 00:05:32.160042 systemd[1287]: Reached target basic.target. Sep 13 00:05:32.160084 systemd[1287]: Reached target default.target. Sep 13 00:05:32.160107 systemd[1287]: Startup finished in 65ms. Sep 13 00:05:32.160320 systemd[1]: Started user@500.service. Sep 13 00:05:32.167253 systemd[1]: Started session-1.scope. Sep 13 00:05:32.229566 systemd[1]: Started sshd@1-10.0.0.20:22-10.0.0.1:54336.service. Sep 13 00:05:32.288102 sshd[1296]: Accepted publickey for core from 10.0.0.1 port 54336 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:05:32.288676 sshd[1296]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.293189 systemd-logind[1206]: New session 2 of user core. Sep 13 00:05:32.294160 systemd[1]: Started session-2.scope. Sep 13 00:05:32.354715 sshd[1296]: pam_unix(sshd:session): session closed for user core Sep 13 00:05:32.358497 systemd[1]: Started sshd@2-10.0.0.20:22-10.0.0.1:54346.service. Sep 13 00:05:32.358924 systemd[1]: sshd@1-10.0.0.20:22-10.0.0.1:54336.service: Deactivated successfully. Sep 13 00:05:32.359865 systemd-logind[1206]: Session 2 logged out. Waiting for processes to exit. Sep 13 00:05:32.359913 systemd[1]: session-2.scope: Deactivated successfully. Sep 13 00:05:32.360762 systemd-logind[1206]: Removed session 2. Sep 13 00:05:32.405167 sshd[1301]: Accepted publickey for core from 10.0.0.1 port 54346 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:05:32.405589 sshd[1301]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.408907 systemd-logind[1206]: New session 3 of user core. Sep 13 00:05:32.409694 systemd[1]: Started session-3.scope. Sep 13 00:05:32.460992 sshd[1301]: pam_unix(sshd:session): session closed for user core Sep 13 00:05:32.464455 systemd[1]: Started sshd@3-10.0.0.20:22-10.0.0.1:54358.service. Sep 13 00:05:32.464951 systemd[1]: sshd@2-10.0.0.20:22-10.0.0.1:54346.service: Deactivated successfully. Sep 13 00:05:32.465539 systemd[1]: session-3.scope: Deactivated successfully. Sep 13 00:05:32.466545 systemd-logind[1206]: Session 3 logged out. Waiting for processes to exit. Sep 13 00:05:32.468002 systemd-logind[1206]: Removed session 3. Sep 13 00:05:32.507991 sshd[1307]: Accepted publickey for core from 10.0.0.1 port 54358 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:05:32.509234 sshd[1307]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.512958 systemd-logind[1206]: New session 4 of user core. Sep 13 00:05:32.513645 systemd[1]: Started session-4.scope. Sep 13 00:05:32.572415 sshd[1307]: pam_unix(sshd:session): session closed for user core Sep 13 00:05:32.576398 systemd[1]: Started sshd@4-10.0.0.20:22-10.0.0.1:54368.service. Sep 13 00:05:32.576843 systemd[1]: sshd@3-10.0.0.20:22-10.0.0.1:54358.service: Deactivated successfully. Sep 13 00:05:32.577770 systemd-logind[1206]: Session 4 logged out. Waiting for processes to exit. Sep 13 00:05:32.577820 systemd[1]: session-4.scope: Deactivated successfully. Sep 13 00:05:32.578528 systemd-logind[1206]: Removed session 4. Sep 13 00:05:32.622023 sshd[1313]: Accepted publickey for core from 10.0.0.1 port 54368 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:05:32.622515 sshd[1313]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:05:32.626716 systemd-logind[1206]: New session 5 of user core. Sep 13 00:05:32.627542 systemd[1]: Started session-5.scope. Sep 13 00:05:32.687226 sudo[1318]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 13 00:05:32.687466 sudo[1318]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 13 00:05:32.730255 systemd[1]: Starting docker.service... Sep 13 00:05:32.795153 env[1329]: time="2025-09-13T00:05:32.795097472Z" level=info msg="Starting up" Sep 13 00:05:32.796663 env[1329]: time="2025-09-13T00:05:32.796617999Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:05:32.796663 env[1329]: time="2025-09-13T00:05:32.796644858Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:05:32.796663 env[1329]: time="2025-09-13T00:05:32.796665173Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:05:32.796804 env[1329]: time="2025-09-13T00:05:32.796676644Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:05:32.801591 env[1329]: time="2025-09-13T00:05:32.801562023Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:05:32.801685 env[1329]: time="2025-09-13T00:05:32.801670629Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:05:32.801772 env[1329]: time="2025-09-13T00:05:32.801755244Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:05:32.801828 env[1329]: time="2025-09-13T00:05:32.801814737Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:05:32.810746 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2017011766-merged.mount: Deactivated successfully. Sep 13 00:05:33.052949 env[1329]: time="2025-09-13T00:05:33.052898507Z" level=info msg="Loading containers: start." Sep 13 00:05:33.206765 kernel: Initializing XFRM netlink socket Sep 13 00:05:33.233346 env[1329]: time="2025-09-13T00:05:33.233299901Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Sep 13 00:05:33.306273 systemd-networkd[1044]: docker0: Link UP Sep 13 00:05:33.329534 env[1329]: time="2025-09-13T00:05:33.329489375Z" level=info msg="Loading containers: done." Sep 13 00:05:33.347123 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3056200549-merged.mount: Deactivated successfully. Sep 13 00:05:33.352809 env[1329]: time="2025-09-13T00:05:33.352762248Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 13 00:05:33.352977 env[1329]: time="2025-09-13T00:05:33.352956971Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Sep 13 00:05:33.353081 env[1329]: time="2025-09-13T00:05:33.353066576Z" level=info msg="Daemon has completed initialization" Sep 13 00:05:33.371936 env[1329]: time="2025-09-13T00:05:33.371815392Z" level=info msg="API listen on /run/docker.sock" Sep 13 00:05:33.372081 systemd[1]: Started docker.service. Sep 13 00:05:34.043529 env[1216]: time="2025-09-13T00:05:34.043474833Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\"" Sep 13 00:05:34.648231 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1042710886.mount: Deactivated successfully. Sep 13 00:05:35.919040 env[1216]: time="2025-09-13T00:05:35.918965187Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:35.920637 env[1216]: time="2025-09-13T00:05:35.920600453Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:0b1c07d8fd4a3526d5c44502e682df3627a3b01c1e608e5e24c3519c8fb337b6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:35.922592 env[1216]: time="2025-09-13T00:05:35.922567104Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:35.924995 env[1216]: time="2025-09-13T00:05:35.924963095Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:35.925799 env[1216]: time="2025-09-13T00:05:35.925773904Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\" returns image reference \"sha256:0b1c07d8fd4a3526d5c44502e682df3627a3b01c1e608e5e24c3519c8fb337b6\"" Sep 13 00:05:35.928608 env[1216]: time="2025-09-13T00:05:35.928342090Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\"" Sep 13 00:05:37.253938 env[1216]: time="2025-09-13T00:05:37.253888744Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:37.255369 env[1216]: time="2025-09-13T00:05:37.255337777Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c359cb88f3d2147f2cb4c5ada4fbdeadc4b1c009d66c8f33f3856efaf04ee6ef,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:37.257195 env[1216]: time="2025-09-13T00:05:37.257164806Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:37.259533 env[1216]: time="2025-09-13T00:05:37.259507837Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:37.260281 env[1216]: time="2025-09-13T00:05:37.260253458Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\" returns image reference \"sha256:c359cb88f3d2147f2cb4c5ada4fbdeadc4b1c009d66c8f33f3856efaf04ee6ef\"" Sep 13 00:05:37.260745 env[1216]: time="2025-09-13T00:05:37.260711009Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\"" Sep 13 00:05:38.488517 env[1216]: time="2025-09-13T00:05:38.488463523Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:38.490796 env[1216]: time="2025-09-13T00:05:38.490750865Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5e3cbe2ba7db787c6aebfcf4484156dd4ebd7ede811ef72e8929593e59a5fa27,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:38.492717 env[1216]: time="2025-09-13T00:05:38.492683906Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:38.495343 env[1216]: time="2025-09-13T00:05:38.495314901Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:38.496111 env[1216]: time="2025-09-13T00:05:38.496077053Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\" returns image reference \"sha256:5e3cbe2ba7db787c6aebfcf4484156dd4ebd7ede811ef72e8929593e59a5fa27\"" Sep 13 00:05:38.496721 env[1216]: time="2025-09-13T00:05:38.496693613Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\"" Sep 13 00:05:39.556994 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount482454202.mount: Deactivated successfully. Sep 13 00:05:40.103254 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 13 00:05:40.103436 systemd[1]: Stopped kubelet.service. Sep 13 00:05:40.104853 systemd[1]: Starting kubelet.service... Sep 13 00:05:40.195248 systemd[1]: Started kubelet.service. Sep 13 00:05:40.303439 env[1216]: time="2025-09-13T00:05:40.303325821Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:40.305738 env[1216]: time="2025-09-13T00:05:40.305688491Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c15699f0b7002450249485b10f20211982dfd2bec4d61c86c35acebc659e794e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:40.307485 env[1216]: time="2025-09-13T00:05:40.307437557Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:40.310492 env[1216]: time="2025-09-13T00:05:40.310084895Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:40.311036 env[1216]: time="2025-09-13T00:05:40.310984170Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\" returns image reference \"sha256:c15699f0b7002450249485b10f20211982dfd2bec4d61c86c35acebc659e794e\"" Sep 13 00:05:40.313715 env[1216]: time="2025-09-13T00:05:40.313123175Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 13 00:05:40.330244 kubelet[1464]: E0913 00:05:40.330185 1464 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:05:40.332859 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:05:40.332996 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:05:40.928087 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount431347001.mount: Deactivated successfully. Sep 13 00:05:41.906072 env[1216]: time="2025-09-13T00:05:41.906016251Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:41.908075 env[1216]: time="2025-09-13T00:05:41.908038748Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:41.910217 env[1216]: time="2025-09-13T00:05:41.910179590Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:41.912770 env[1216]: time="2025-09-13T00:05:41.912742824Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:41.913640 env[1216]: time="2025-09-13T00:05:41.913609030Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Sep 13 00:05:41.914257 env[1216]: time="2025-09-13T00:05:41.914231364Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 13 00:05:42.420469 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2595441026.mount: Deactivated successfully. Sep 13 00:05:42.425772 env[1216]: time="2025-09-13T00:05:42.425737543Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:42.427998 env[1216]: time="2025-09-13T00:05:42.427968035Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:42.429692 env[1216]: time="2025-09-13T00:05:42.429648604Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:42.431786 env[1216]: time="2025-09-13T00:05:42.431701444Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:42.432282 env[1216]: time="2025-09-13T00:05:42.432247959Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Sep 13 00:05:42.433480 env[1216]: time="2025-09-13T00:05:42.433422155Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 13 00:05:42.977373 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1858037263.mount: Deactivated successfully. Sep 13 00:05:45.206873 env[1216]: time="2025-09-13T00:05:45.206787218Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:45.209091 env[1216]: time="2025-09-13T00:05:45.209064395Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:45.211389 env[1216]: time="2025-09-13T00:05:45.211359402Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:45.213805 env[1216]: time="2025-09-13T00:05:45.213778702Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:45.214614 env[1216]: time="2025-09-13T00:05:45.214583639Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Sep 13 00:05:50.353092 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 13 00:05:50.353263 systemd[1]: Stopped kubelet.service. Sep 13 00:05:50.354698 systemd[1]: Starting kubelet.service... Sep 13 00:05:50.468127 systemd[1]: Started kubelet.service. Sep 13 00:05:50.511639 kubelet[1495]: E0913 00:05:50.511551 1495 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:05:50.514436 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:05:50.514852 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:05:51.353982 systemd[1]: Stopped kubelet.service. Sep 13 00:05:51.357478 systemd[1]: Starting kubelet.service... Sep 13 00:05:51.385576 systemd[1]: Reloading. Sep 13 00:05:51.442625 /usr/lib/systemd/system-generators/torcx-generator[1532]: time="2025-09-13T00:05:51Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:05:51.442657 /usr/lib/systemd/system-generators/torcx-generator[1532]: time="2025-09-13T00:05:51Z" level=info msg="torcx already run" Sep 13 00:05:51.594433 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:05:51.594452 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:05:51.611173 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:05:51.687059 systemd[1]: Started kubelet.service. Sep 13 00:05:51.692491 systemd[1]: Stopping kubelet.service... Sep 13 00:05:51.693197 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:05:51.693484 systemd[1]: Stopped kubelet.service. Sep 13 00:05:51.695461 systemd[1]: Starting kubelet.service... Sep 13 00:05:51.790044 systemd[1]: Started kubelet.service. Sep 13 00:05:51.841182 kubelet[1580]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:05:51.841182 kubelet[1580]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 00:05:51.841182 kubelet[1580]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:05:51.841596 kubelet[1580]: I0913 00:05:51.841237 1580 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:05:52.934572 kubelet[1580]: I0913 00:05:52.934514 1580 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 00:05:52.934572 kubelet[1580]: I0913 00:05:52.934553 1580 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:05:52.934922 kubelet[1580]: I0913 00:05:52.934805 1580 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 00:05:52.962149 kubelet[1580]: E0913 00:05:52.962111 1580 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.20:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:52.964327 kubelet[1580]: I0913 00:05:52.964290 1580 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:05:52.974385 kubelet[1580]: E0913 00:05:52.974344 1580 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:05:52.974385 kubelet[1580]: I0913 00:05:52.974386 1580 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:05:52.978098 kubelet[1580]: I0913 00:05:52.978065 1580 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:05:52.978932 kubelet[1580]: I0913 00:05:52.978901 1580 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 00:05:52.979093 kubelet[1580]: I0913 00:05:52.979051 1580 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:05:52.979271 kubelet[1580]: I0913 00:05:52.979086 1580 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 00:05:52.979353 kubelet[1580]: I0913 00:05:52.979333 1580 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:05:52.979353 kubelet[1580]: I0913 00:05:52.979343 1580 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 00:05:52.979596 kubelet[1580]: I0913 00:05:52.979568 1580 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:05:52.983156 kubelet[1580]: I0913 00:05:52.983124 1580 kubelet.go:408] "Attempting to sync node with API server" Sep 13 00:05:52.983156 kubelet[1580]: I0913 00:05:52.983160 1580 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:05:52.983287 kubelet[1580]: I0913 00:05:52.983189 1580 kubelet.go:314] "Adding apiserver pod source" Sep 13 00:05:52.983287 kubelet[1580]: I0913 00:05:52.983200 1580 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:05:53.013103 kubelet[1580]: W0913 00:05:53.012975 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.20:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:53.013319 kubelet[1580]: E0913 00:05:53.013294 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.20:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:53.018700 kubelet[1580]: W0913 00:05:53.018655 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.20:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:53.018868 kubelet[1580]: E0913 00:05:53.018845 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.20:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:53.019077 kubelet[1580]: I0913 00:05:53.019045 1580 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:05:53.019754 kubelet[1580]: I0913 00:05:53.019735 1580 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 00:05:53.019915 kubelet[1580]: W0913 00:05:53.019903 1580 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 13 00:05:53.020810 kubelet[1580]: I0913 00:05:53.020793 1580 server.go:1274] "Started kubelet" Sep 13 00:05:53.021641 kubelet[1580]: I0913 00:05:53.021340 1580 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:05:53.021861 kubelet[1580]: I0913 00:05:53.021838 1580 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:05:53.021903 kubelet[1580]: I0913 00:05:53.021555 1580 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:05:53.023017 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Sep 13 00:05:53.023152 kubelet[1580]: I0913 00:05:53.023131 1580 server.go:449] "Adding debug handlers to kubelet server" Sep 13 00:05:53.023152 kubelet[1580]: I0913 00:05:53.023146 1580 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:05:53.024352 kubelet[1580]: I0913 00:05:53.024323 1580 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:05:53.027022 kubelet[1580]: E0913 00:05:53.026976 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:53.027022 kubelet[1580]: I0913 00:05:53.027018 1580 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 00:05:53.027199 kubelet[1580]: I0913 00:05:53.027177 1580 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 00:05:53.027276 kubelet[1580]: I0913 00:05:53.027253 1580 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:05:53.027509 kubelet[1580]: E0913 00:05:53.027478 1580 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.20:6443: connect: connection refused" interval="200ms" Sep 13 00:05:53.027679 kubelet[1580]: I0913 00:05:53.027654 1580 factory.go:221] Registration of the systemd container factory successfully Sep 13 00:05:53.027826 kubelet[1580]: E0913 00:05:53.027751 1580 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:05:53.027826 kubelet[1580]: I0913 00:05:53.027758 1580 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:05:53.027959 kubelet[1580]: W0913 00:05:53.027644 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.20:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:53.028009 kubelet[1580]: E0913 00:05:53.027971 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.20:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:53.028095 kubelet[1580]: E0913 00:05:53.026385 1580 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.20:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.20:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1864aecdafb570f2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-09-13 00:05:53.020768498 +0000 UTC m=+1.225781130,LastTimestamp:2025-09-13 00:05:53.020768498 +0000 UTC m=+1.225781130,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 13 00:05:53.029534 kubelet[1580]: I0913 00:05:53.029511 1580 factory.go:221] Registration of the containerd container factory successfully Sep 13 00:05:53.039626 kubelet[1580]: I0913 00:05:53.039610 1580 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 00:05:53.039713 kubelet[1580]: I0913 00:05:53.039702 1580 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 00:05:53.039838 kubelet[1580]: I0913 00:05:53.039827 1580 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:05:53.040873 kubelet[1580]: I0913 00:05:53.040788 1580 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 00:05:53.041622 kubelet[1580]: I0913 00:05:53.041594 1580 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 00:05:53.041622 kubelet[1580]: I0913 00:05:53.041615 1580 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 00:05:53.041682 kubelet[1580]: I0913 00:05:53.041631 1580 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 00:05:53.041682 kubelet[1580]: E0913 00:05:53.041667 1580 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:05:53.046863 kubelet[1580]: W0913 00:05:53.046823 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.20:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:53.047015 kubelet[1580]: E0913 00:05:53.046995 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.20:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:53.127529 kubelet[1580]: E0913 00:05:53.127486 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:53.141889 kubelet[1580]: E0913 00:05:53.141858 1580 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 00:05:53.191266 kubelet[1580]: I0913 00:05:53.190064 1580 policy_none.go:49] "None policy: Start" Sep 13 00:05:53.191266 kubelet[1580]: I0913 00:05:53.190882 1580 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 00:05:53.191266 kubelet[1580]: I0913 00:05:53.190907 1580 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:05:53.196963 systemd[1]: Created slice kubepods.slice. Sep 13 00:05:53.202818 systemd[1]: Created slice kubepods-burstable.slice. Sep 13 00:05:53.206275 systemd[1]: Created slice kubepods-besteffort.slice. Sep 13 00:05:53.223974 kubelet[1580]: I0913 00:05:53.223874 1580 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 00:05:53.224102 kubelet[1580]: I0913 00:05:53.224049 1580 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:05:53.224102 kubelet[1580]: I0913 00:05:53.224061 1580 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:05:53.224290 kubelet[1580]: I0913 00:05:53.224270 1580 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:05:53.226365 kubelet[1580]: E0913 00:05:53.226333 1580 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 13 00:05:53.228458 kubelet[1580]: E0913 00:05:53.228420 1580 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.20:6443: connect: connection refused" interval="400ms" Sep 13 00:05:53.325617 kubelet[1580]: I0913 00:05:53.325590 1580 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 13 00:05:53.326231 kubelet[1580]: E0913 00:05:53.326203 1580 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.20:6443/api/v1/nodes\": dial tcp 10.0.0.20:6443: connect: connection refused" node="localhost" Sep 13 00:05:53.349298 systemd[1]: Created slice kubepods-burstable-podf876ab7e9d10ea8070215097afbf3bbf.slice. Sep 13 00:05:53.363675 systemd[1]: Created slice kubepods-burstable-pod71d8bf7bd9b7c7432927bee9d50592b5.slice. Sep 13 00:05:53.366362 systemd[1]: Created slice kubepods-burstable-podfe5e332fba00ba0b5b33a25fe2e8fd7b.slice. Sep 13 00:05:53.527745 kubelet[1580]: I0913 00:05:53.527613 1580 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 13 00:05:53.528051 kubelet[1580]: E0913 00:05:53.527994 1580 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.20:6443/api/v1/nodes\": dial tcp 10.0.0.20:6443: connect: connection refused" node="localhost" Sep 13 00:05:53.529209 kubelet[1580]: I0913 00:05:53.529166 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:53.529264 kubelet[1580]: I0913 00:05:53.529219 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:53.529264 kubelet[1580]: I0913 00:05:53.529240 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:53.529326 kubelet[1580]: I0913 00:05:53.529277 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:53.529326 kubelet[1580]: I0913 00:05:53.529297 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fe5e332fba00ba0b5b33a25fe2e8fd7b-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"fe5e332fba00ba0b5b33a25fe2e8fd7b\") " pod="kube-system/kube-scheduler-localhost" Sep 13 00:05:53.529326 kubelet[1580]: I0913 00:05:53.529312 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:53.529393 kubelet[1580]: I0913 00:05:53.529340 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:53.529393 kubelet[1580]: I0913 00:05:53.529359 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:53.529393 kubelet[1580]: I0913 00:05:53.529376 1580 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:53.629966 kubelet[1580]: E0913 00:05:53.629908 1580 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.20:6443: connect: connection refused" interval="800ms" Sep 13 00:05:53.662240 kubelet[1580]: E0913 00:05:53.662184 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:53.663025 env[1216]: time="2025-09-13T00:05:53.662960728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:f876ab7e9d10ea8070215097afbf3bbf,Namespace:kube-system,Attempt:0,}" Sep 13 00:05:53.666009 kubelet[1580]: E0913 00:05:53.665984 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:53.666548 env[1216]: time="2025-09-13T00:05:53.666408117Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:71d8bf7bd9b7c7432927bee9d50592b5,Namespace:kube-system,Attempt:0,}" Sep 13 00:05:53.668061 kubelet[1580]: E0913 00:05:53.668039 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:53.668436 env[1216]: time="2025-09-13T00:05:53.668406094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:fe5e332fba00ba0b5b33a25fe2e8fd7b,Namespace:kube-system,Attempt:0,}" Sep 13 00:05:53.929035 kubelet[1580]: I0913 00:05:53.928989 1580 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 13 00:05:53.929317 kubelet[1580]: E0913 00:05:53.929280 1580 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.20:6443/api/v1/nodes\": dial tcp 10.0.0.20:6443: connect: connection refused" node="localhost" Sep 13 00:05:54.209342 kubelet[1580]: W0913 00:05:54.209218 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.20:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:54.209837 kubelet[1580]: E0913 00:05:54.209812 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.20:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:54.387745 kubelet[1580]: W0913 00:05:54.387589 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.20:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:54.387745 kubelet[1580]: E0913 00:05:54.387661 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.20:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:54.396678 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1544754353.mount: Deactivated successfully. Sep 13 00:05:54.400096 env[1216]: time="2025-09-13T00:05:54.400053294Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.401526 env[1216]: time="2025-09-13T00:05:54.401495717Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.404342 env[1216]: time="2025-09-13T00:05:54.404305684Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.405862 env[1216]: time="2025-09-13T00:05:54.405834592Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.407477 env[1216]: time="2025-09-13T00:05:54.407444781Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.409194 env[1216]: time="2025-09-13T00:05:54.409163307Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.411444 env[1216]: time="2025-09-13T00:05:54.411412145Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.412544 env[1216]: time="2025-09-13T00:05:54.412502467Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.415430 env[1216]: time="2025-09-13T00:05:54.415392515Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.417169 env[1216]: time="2025-09-13T00:05:54.417136534Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.419099 env[1216]: time="2025-09-13T00:05:54.419056283Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.420336 env[1216]: time="2025-09-13T00:05:54.420297722Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:05:54.430678 kubelet[1580]: E0913 00:05:54.430586 1580 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.20:6443: connect: connection refused" interval="1.6s" Sep 13 00:05:54.443613 env[1216]: time="2025-09-13T00:05:54.443535212Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:05:54.443923 env[1216]: time="2025-09-13T00:05:54.443586478Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:05:54.443923 env[1216]: time="2025-09-13T00:05:54.443597884Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:05:54.444133 env[1216]: time="2025-09-13T00:05:54.443996930Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c95eaff291b059f58acdb16c3da42df92a0cd10232efafa9e4b98fd65e02c48f pid=1625 runtime=io.containerd.runc.v2 Sep 13 00:05:54.446189 env[1216]: time="2025-09-13T00:05:54.446128708Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:05:54.446293 env[1216]: time="2025-09-13T00:05:54.446168929Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:05:54.446293 env[1216]: time="2025-09-13T00:05:54.446179134Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:05:54.446410 env[1216]: time="2025-09-13T00:05:54.446317405Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6f875492ed67c6ed57919bdb949c72fe99b7f403aa6fa98fa3bcdd1deb34087b pid=1631 runtime=io.containerd.runc.v2 Sep 13 00:05:54.448857 env[1216]: time="2025-09-13T00:05:54.448790039Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:05:54.448857 env[1216]: time="2025-09-13T00:05:54.448831140Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:05:54.449018 env[1216]: time="2025-09-13T00:05:54.448841826Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:05:54.449285 env[1216]: time="2025-09-13T00:05:54.449013794Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/96cb7fc72c40204e5aec0386347e0990ff326bb0e2a01d2c66ca590283d1f62f pid=1656 runtime=io.containerd.runc.v2 Sep 13 00:05:54.459238 systemd[1]: Started cri-containerd-6f875492ed67c6ed57919bdb949c72fe99b7f403aa6fa98fa3bcdd1deb34087b.scope. Sep 13 00:05:54.467282 systemd[1]: Started cri-containerd-96cb7fc72c40204e5aec0386347e0990ff326bb0e2a01d2c66ca590283d1f62f.scope. Sep 13 00:05:54.468906 systemd[1]: Started cri-containerd-c95eaff291b059f58acdb16c3da42df92a0cd10232efafa9e4b98fd65e02c48f.scope. Sep 13 00:05:54.480210 kubelet[1580]: W0913 00:05:54.480145 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.20:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:54.480338 kubelet[1580]: E0913 00:05:54.480218 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.20:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:54.490961 kubelet[1580]: W0913 00:05:54.490899 1580 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.20:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.20:6443: connect: connection refused Sep 13 00:05:54.491066 kubelet[1580]: E0913 00:05:54.490968 1580 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.20:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.20:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:05:54.516239 env[1216]: time="2025-09-13T00:05:54.516187956Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:fe5e332fba00ba0b5b33a25fe2e8fd7b,Namespace:kube-system,Attempt:0,} returns sandbox id \"6f875492ed67c6ed57919bdb949c72fe99b7f403aa6fa98fa3bcdd1deb34087b\"" Sep 13 00:05:54.516557 env[1216]: time="2025-09-13T00:05:54.516417595Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:71d8bf7bd9b7c7432927bee9d50592b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"c95eaff291b059f58acdb16c3da42df92a0cd10232efafa9e4b98fd65e02c48f\"" Sep 13 00:05:54.517666 kubelet[1580]: E0913 00:05:54.517470 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:54.517666 kubelet[1580]: E0913 00:05:54.517506 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:54.519668 env[1216]: time="2025-09-13T00:05:54.519632411Z" level=info msg="CreateContainer within sandbox \"c95eaff291b059f58acdb16c3da42df92a0cd10232efafa9e4b98fd65e02c48f\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 13 00:05:54.519982 env[1216]: time="2025-09-13T00:05:54.519955137Z" level=info msg="CreateContainer within sandbox \"6f875492ed67c6ed57919bdb949c72fe99b7f403aa6fa98fa3bcdd1deb34087b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 13 00:05:54.524540 env[1216]: time="2025-09-13T00:05:54.524509443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:f876ab7e9d10ea8070215097afbf3bbf,Namespace:kube-system,Attempt:0,} returns sandbox id \"96cb7fc72c40204e5aec0386347e0990ff326bb0e2a01d2c66ca590283d1f62f\"" Sep 13 00:05:54.525278 kubelet[1580]: E0913 00:05:54.525252 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:54.527010 env[1216]: time="2025-09-13T00:05:54.526979155Z" level=info msg="CreateContainer within sandbox \"96cb7fc72c40204e5aec0386347e0990ff326bb0e2a01d2c66ca590283d1f62f\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 13 00:05:54.536227 env[1216]: time="2025-09-13T00:05:54.536175092Z" level=info msg="CreateContainer within sandbox \"6f875492ed67c6ed57919bdb949c72fe99b7f403aa6fa98fa3bcdd1deb34087b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"043894d063fbd3cf0d95bf240382beb28ea8e825accff2f1d310347135b665fe\"" Sep 13 00:05:54.536874 env[1216]: time="2025-09-13T00:05:54.536846798Z" level=info msg="StartContainer for \"043894d063fbd3cf0d95bf240382beb28ea8e825accff2f1d310347135b665fe\"" Sep 13 00:05:54.539261 env[1216]: time="2025-09-13T00:05:54.539216338Z" level=info msg="CreateContainer within sandbox \"c95eaff291b059f58acdb16c3da42df92a0cd10232efafa9e4b98fd65e02c48f\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"ebdad319c88fed86ea23090ec64b9c3216c529a0b258f068977ccf52213b09d7\"" Sep 13 00:05:54.539632 env[1216]: time="2025-09-13T00:05:54.539607780Z" level=info msg="StartContainer for \"ebdad319c88fed86ea23090ec64b9c3216c529a0b258f068977ccf52213b09d7\"" Sep 13 00:05:54.547259 env[1216]: time="2025-09-13T00:05:54.547215459Z" level=info msg="CreateContainer within sandbox \"96cb7fc72c40204e5aec0386347e0990ff326bb0e2a01d2c66ca590283d1f62f\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b73d50961c001f76c1e41c97eedbe6df2352dc06f0fea2e3195a8cc13498e27a\"" Sep 13 00:05:54.547671 env[1216]: time="2025-09-13T00:05:54.547646641Z" level=info msg="StartContainer for \"b73d50961c001f76c1e41c97eedbe6df2352dc06f0fea2e3195a8cc13498e27a\"" Sep 13 00:05:54.554049 systemd[1]: Started cri-containerd-043894d063fbd3cf0d95bf240382beb28ea8e825accff2f1d310347135b665fe.scope. Sep 13 00:05:54.557574 systemd[1]: Started cri-containerd-ebdad319c88fed86ea23090ec64b9c3216c529a0b258f068977ccf52213b09d7.scope. Sep 13 00:05:54.572387 systemd[1]: Started cri-containerd-b73d50961c001f76c1e41c97eedbe6df2352dc06f0fea2e3195a8cc13498e27a.scope. Sep 13 00:05:54.600542 env[1216]: time="2025-09-13T00:05:54.600470531Z" level=info msg="StartContainer for \"ebdad319c88fed86ea23090ec64b9c3216c529a0b258f068977ccf52213b09d7\" returns successfully" Sep 13 00:05:54.610182 env[1216]: time="2025-09-13T00:05:54.610120982Z" level=info msg="StartContainer for \"043894d063fbd3cf0d95bf240382beb28ea8e825accff2f1d310347135b665fe\" returns successfully" Sep 13 00:05:54.627641 env[1216]: time="2025-09-13T00:05:54.627589981Z" level=info msg="StartContainer for \"b73d50961c001f76c1e41c97eedbe6df2352dc06f0fea2e3195a8cc13498e27a\" returns successfully" Sep 13 00:05:54.730927 kubelet[1580]: I0913 00:05:54.730827 1580 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 13 00:05:55.053349 kubelet[1580]: E0913 00:05:55.053221 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:55.054668 kubelet[1580]: E0913 00:05:55.054646 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:55.056178 kubelet[1580]: E0913 00:05:55.056158 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:56.039039 kubelet[1580]: E0913 00:05:56.038987 1580 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 13 00:05:56.058711 kubelet[1580]: E0913 00:05:56.058680 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:56.088108 kubelet[1580]: I0913 00:05:56.088057 1580 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 13 00:05:56.088108 kubelet[1580]: E0913 00:05:56.088101 1580 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Sep 13 00:05:56.102688 kubelet[1580]: E0913 00:05:56.102651 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.203229 kubelet[1580]: E0913 00:05:56.203149 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.303927 kubelet[1580]: E0913 00:05:56.303780 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.404586 kubelet[1580]: E0913 00:05:56.404539 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.505269 kubelet[1580]: E0913 00:05:56.505213 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.606074 kubelet[1580]: E0913 00:05:56.606029 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:56.706676 kubelet[1580]: E0913 00:05:56.706605 1580 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:57.014404 kubelet[1580]: I0913 00:05:57.014312 1580 apiserver.go:52] "Watching apiserver" Sep 13 00:05:57.028239 kubelet[1580]: I0913 00:05:57.028200 1580 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 00:05:57.067075 kubelet[1580]: E0913 00:05:57.067042 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:58.060545 kubelet[1580]: E0913 00:05:58.060490 1580 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:58.242712 systemd[1]: Reloading. Sep 13 00:05:58.311232 /usr/lib/systemd/system-generators/torcx-generator[1878]: time="2025-09-13T00:05:58Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:05:58.311283 /usr/lib/systemd/system-generators/torcx-generator[1878]: time="2025-09-13T00:05:58Z" level=info msg="torcx already run" Sep 13 00:05:58.386393 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:05:58.386412 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:05:58.405758 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:05:58.489020 systemd[1]: Stopping kubelet.service... Sep 13 00:05:58.506131 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:05:58.506483 systemd[1]: Stopped kubelet.service. Sep 13 00:05:58.506623 systemd[1]: kubelet.service: Consumed 1.553s CPU time. Sep 13 00:05:58.509102 systemd[1]: Starting kubelet.service... Sep 13 00:05:58.611042 systemd[1]: Started kubelet.service. Sep 13 00:05:58.666343 kubelet[1920]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:05:58.666343 kubelet[1920]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 00:05:58.666343 kubelet[1920]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:05:58.666343 kubelet[1920]: I0913 00:05:58.664711 1920 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:05:58.672257 kubelet[1920]: I0913 00:05:58.672209 1920 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 00:05:58.672257 kubelet[1920]: I0913 00:05:58.672245 1920 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:05:58.672490 kubelet[1920]: I0913 00:05:58.672473 1920 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 00:05:58.673905 kubelet[1920]: I0913 00:05:58.673884 1920 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 13 00:05:58.676908 kubelet[1920]: I0913 00:05:58.676874 1920 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:05:58.680959 kubelet[1920]: E0913 00:05:58.680924 1920 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:05:58.680959 kubelet[1920]: I0913 00:05:58.680961 1920 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:05:58.683542 kubelet[1920]: I0913 00:05:58.683508 1920 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:05:58.683667 kubelet[1920]: I0913 00:05:58.683653 1920 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 00:05:58.683848 kubelet[1920]: I0913 00:05:58.683815 1920 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:05:58.684016 kubelet[1920]: I0913 00:05:58.683851 1920 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 00:05:58.684100 kubelet[1920]: I0913 00:05:58.684019 1920 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:05:58.684100 kubelet[1920]: I0913 00:05:58.684028 1920 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 00:05:58.684100 kubelet[1920]: I0913 00:05:58.684064 1920 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:05:58.684170 kubelet[1920]: I0913 00:05:58.684165 1920 kubelet.go:408] "Attempting to sync node with API server" Sep 13 00:05:58.684200 kubelet[1920]: I0913 00:05:58.684178 1920 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:05:58.684226 kubelet[1920]: I0913 00:05:58.684200 1920 kubelet.go:314] "Adding apiserver pod source" Sep 13 00:05:58.684226 kubelet[1920]: I0913 00:05:58.684213 1920 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:05:58.684850 kubelet[1920]: I0913 00:05:58.684807 1920 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.685395 1920 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.685903 1920 server.go:1274] "Started kubelet" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.686678 1920 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.687459 1920 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.687520 1920 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.688043 1920 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.688642 1920 server.go:449] "Adding debug handlers to kubelet server" Sep 13 00:05:58.691777 kubelet[1920]: I0913 00:05:58.690115 1920 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:05:58.692047 kubelet[1920]: I0913 00:05:58.691946 1920 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 00:05:58.692047 kubelet[1920]: I0913 00:05:58.692034 1920 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 00:05:58.692458 kubelet[1920]: I0913 00:05:58.692140 1920 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:05:58.693314 kubelet[1920]: E0913 00:05:58.693267 1920 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 13 00:05:58.693830 kubelet[1920]: E0913 00:05:58.693676 1920 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:05:58.707735 kubelet[1920]: I0913 00:05:58.707677 1920 factory.go:221] Registration of the containerd container factory successfully Sep 13 00:05:58.707735 kubelet[1920]: I0913 00:05:58.707715 1920 factory.go:221] Registration of the systemd container factory successfully Sep 13 00:05:58.707967 kubelet[1920]: I0913 00:05:58.707935 1920 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:05:58.731954 kubelet[1920]: I0913 00:05:58.731830 1920 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 00:05:58.733389 kubelet[1920]: I0913 00:05:58.733356 1920 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 00:05:58.733957 kubelet[1920]: I0913 00:05:58.733943 1920 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 00:05:58.735091 kubelet[1920]: I0913 00:05:58.735074 1920 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 00:05:58.738137 kubelet[1920]: E0913 00:05:58.735679 1920 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:05:58.747140 kubelet[1920]: I0913 00:05:58.747114 1920 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 00:05:58.747140 kubelet[1920]: I0913 00:05:58.747131 1920 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 00:05:58.747299 kubelet[1920]: I0913 00:05:58.747151 1920 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:05:58.747324 kubelet[1920]: I0913 00:05:58.747307 1920 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 13 00:05:58.747349 kubelet[1920]: I0913 00:05:58.747324 1920 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 13 00:05:58.747349 kubelet[1920]: I0913 00:05:58.747344 1920 policy_none.go:49] "None policy: Start" Sep 13 00:05:58.747987 kubelet[1920]: I0913 00:05:58.747966 1920 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 00:05:58.748074 kubelet[1920]: I0913 00:05:58.747994 1920 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:05:58.748168 kubelet[1920]: I0913 00:05:58.748151 1920 state_mem.go:75] "Updated machine memory state" Sep 13 00:05:58.752239 kubelet[1920]: I0913 00:05:58.752205 1920 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 00:05:58.752564 kubelet[1920]: I0913 00:05:58.752552 1920 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:05:58.752628 kubelet[1920]: I0913 00:05:58.752565 1920 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:05:58.753073 kubelet[1920]: I0913 00:05:58.752808 1920 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:05:58.848519 kubelet[1920]: E0913 00:05:58.848422 1920 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:58.858388 kubelet[1920]: I0913 00:05:58.858353 1920 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Sep 13 00:05:58.867588 kubelet[1920]: I0913 00:05:58.867426 1920 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Sep 13 00:05:58.867933 kubelet[1920]: I0913 00:05:58.867915 1920 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Sep 13 00:05:58.994085 kubelet[1920]: I0913 00:05:58.994046 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:58.994318 kubelet[1920]: I0913 00:05:58.994297 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:58.994404 kubelet[1920]: I0913 00:05:58.994390 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:58.994488 kubelet[1920]: I0913 00:05:58.994475 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:58.994562 kubelet[1920]: I0913 00:05:58.994550 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fe5e332fba00ba0b5b33a25fe2e8fd7b-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"fe5e332fba00ba0b5b33a25fe2e8fd7b\") " pod="kube-system/kube-scheduler-localhost" Sep 13 00:05:58.994637 kubelet[1920]: I0913 00:05:58.994625 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:58.994714 kubelet[1920]: I0913 00:05:58.994702 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f876ab7e9d10ea8070215097afbf3bbf-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"f876ab7e9d10ea8070215097afbf3bbf\") " pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:58.994838 kubelet[1920]: I0913 00:05:58.994823 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:58.995030 kubelet[1920]: I0913 00:05:58.995012 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/71d8bf7bd9b7c7432927bee9d50592b5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"71d8bf7bd9b7c7432927bee9d50592b5\") " pod="kube-system/kube-controller-manager-localhost" Sep 13 00:05:59.145842 kubelet[1920]: E0913 00:05:59.145690 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.147966 kubelet[1920]: E0913 00:05:59.147939 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.149372 kubelet[1920]: E0913 00:05:59.149349 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.238512 sudo[1956]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 13 00:05:59.238797 sudo[1956]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Sep 13 00:05:59.685476 kubelet[1920]: I0913 00:05:59.685441 1920 apiserver.go:52] "Watching apiserver" Sep 13 00:05:59.692695 kubelet[1920]: I0913 00:05:59.692653 1920 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 00:05:59.738864 sudo[1956]: pam_unix(sudo:session): session closed for user root Sep 13 00:05:59.744558 kubelet[1920]: E0913 00:05:59.744532 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.745110 kubelet[1920]: E0913 00:05:59.745094 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.753365 kubelet[1920]: E0913 00:05:59.753338 1920 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 13 00:05:59.753660 kubelet[1920]: E0913 00:05:59.753646 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:05:59.785637 kubelet[1920]: I0913 00:05:59.785575 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.785556695 podStartE2EDuration="1.785556695s" podCreationTimestamp="2025-09-13 00:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:05:59.764931885 +0000 UTC m=+1.148949557" watchObservedRunningTime="2025-09-13 00:05:59.785556695 +0000 UTC m=+1.169574367" Sep 13 00:05:59.796865 kubelet[1920]: I0913 00:05:59.795218 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.795195603 podStartE2EDuration="2.795195603s" podCreationTimestamp="2025-09-13 00:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:05:59.785863656 +0000 UTC m=+1.169881328" watchObservedRunningTime="2025-09-13 00:05:59.795195603 +0000 UTC m=+1.179213275" Sep 13 00:06:00.745863 kubelet[1920]: E0913 00:06:00.745833 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:01.270302 sudo[1318]: pam_unix(sudo:session): session closed for user root Sep 13 00:06:01.271704 sshd[1313]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:01.274157 systemd[1]: sshd@4-10.0.0.20:22-10.0.0.1:54368.service: Deactivated successfully. Sep 13 00:06:01.274878 systemd[1]: session-5.scope: Deactivated successfully. Sep 13 00:06:01.275032 systemd[1]: session-5.scope: Consumed 7.808s CPU time. Sep 13 00:06:01.275656 systemd-logind[1206]: Session 5 logged out. Waiting for processes to exit. Sep 13 00:06:01.276371 systemd-logind[1206]: Removed session 5. Sep 13 00:06:01.747610 kubelet[1920]: E0913 00:06:01.747569 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:02.861296 kubelet[1920]: I0913 00:06:02.861262 1920 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 13 00:06:02.861629 env[1216]: time="2025-09-13T00:06:02.861559020Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 13 00:06:02.862085 kubelet[1920]: I0913 00:06:02.862069 1920 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 13 00:06:03.816144 kubelet[1920]: I0913 00:06:03.816080 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=5.816063862 podStartE2EDuration="5.816063862s" podCreationTimestamp="2025-09-13 00:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:05:59.798846567 +0000 UTC m=+1.182864239" watchObservedRunningTime="2025-09-13 00:06:03.816063862 +0000 UTC m=+5.200081534" Sep 13 00:06:03.822950 systemd[1]: Created slice kubepods-besteffort-pod1baf9ac8_ee06_4adb_9114_bb30c12318aa.slice. Sep 13 00:06:03.847451 systemd[1]: Created slice kubepods-burstable-pod0498e092_a3ef_4562_8c1f_832bdfc5c680.slice. Sep 13 00:06:03.926815 kubelet[1920]: I0913 00:06:03.926750 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1baf9ac8-ee06-4adb-9114-bb30c12318aa-xtables-lock\") pod \"kube-proxy-rb8kl\" (UID: \"1baf9ac8-ee06-4adb-9114-bb30c12318aa\") " pod="kube-system/kube-proxy-rb8kl" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926798 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cni-path\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926865 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-kernel\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926905 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9xs\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-kube-api-access-jl9xs\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926924 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-run\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926939 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-hostproc\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927169 kubelet[1920]: I0913 00:06:03.926953 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-cgroup\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927032 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-hubble-tls\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927074 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-bpf-maps\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927093 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/1baf9ac8-ee06-4adb-9114-bb30c12318aa-kube-proxy\") pod \"kube-proxy-rb8kl\" (UID: \"1baf9ac8-ee06-4adb-9114-bb30c12318aa\") " pod="kube-system/kube-proxy-rb8kl" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927108 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1baf9ac8-ee06-4adb-9114-bb30c12318aa-lib-modules\") pod \"kube-proxy-rb8kl\" (UID: \"1baf9ac8-ee06-4adb-9114-bb30c12318aa\") " pod="kube-system/kube-proxy-rb8kl" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927124 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-xtables-lock\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927319 kubelet[1920]: I0913 00:06:03.927139 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0498e092-a3ef-4562-8c1f-832bdfc5c680-clustermesh-secrets\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.927894 kubelet[1920]: I0913 00:06:03.927857 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-config-path\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.928065 kubelet[1920]: I0913 00:06:03.928039 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrjw\" (UniqueName: \"kubernetes.io/projected/1baf9ac8-ee06-4adb-9114-bb30c12318aa-kube-api-access-zxrjw\") pod \"kube-proxy-rb8kl\" (UID: \"1baf9ac8-ee06-4adb-9114-bb30c12318aa\") " pod="kube-system/kube-proxy-rb8kl" Sep 13 00:06:03.928115 kubelet[1920]: I0913 00:06:03.928081 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-lib-modules\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.928115 kubelet[1920]: I0913 00:06:03.928108 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-net\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.928181 kubelet[1920]: I0913 00:06:03.928133 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-etc-cni-netd\") pod \"cilium-zxhk9\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " pod="kube-system/cilium-zxhk9" Sep 13 00:06:03.968025 systemd[1]: Created slice kubepods-besteffort-pod3fa19538_2cb2_498e_b692_f575000fce93.slice. Sep 13 00:06:04.029076 kubelet[1920]: I0913 00:06:04.029031 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zsd\" (UniqueName: \"kubernetes.io/projected/3fa19538-2cb2-498e-b692-f575000fce93-kube-api-access-97zsd\") pod \"cilium-operator-5d85765b45-8lt7r\" (UID: \"3fa19538-2cb2-498e-b692-f575000fce93\") " pod="kube-system/cilium-operator-5d85765b45-8lt7r" Sep 13 00:06:04.029326 kubelet[1920]: I0913 00:06:04.029208 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3fa19538-2cb2-498e-b692-f575000fce93-cilium-config-path\") pod \"cilium-operator-5d85765b45-8lt7r\" (UID: \"3fa19538-2cb2-498e-b692-f575000fce93\") " pod="kube-system/cilium-operator-5d85765b45-8lt7r" Sep 13 00:06:04.029515 kubelet[1920]: I0913 00:06:04.029483 1920 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 13 00:06:04.146290 kubelet[1920]: E0913 00:06:04.145961 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.146584 env[1216]: time="2025-09-13T00:06:04.146546261Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-rb8kl,Uid:1baf9ac8-ee06-4adb-9114-bb30c12318aa,Namespace:kube-system,Attempt:0,}" Sep 13 00:06:04.150176 kubelet[1920]: E0913 00:06:04.150151 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.150891 env[1216]: time="2025-09-13T00:06:04.150635707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zxhk9,Uid:0498e092-a3ef-4562-8c1f-832bdfc5c680,Namespace:kube-system,Attempt:0,}" Sep 13 00:06:04.247187 env[1216]: time="2025-09-13T00:06:04.246979101Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:06:04.247187 env[1216]: time="2025-09-13T00:06:04.247065078Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:06:04.247187 env[1216]: time="2025-09-13T00:06:04.247076680Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:06:04.247366 env[1216]: time="2025-09-13T00:06:04.247256153Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d pid=2024 runtime=io.containerd.runc.v2 Sep 13 00:06:04.249965 env[1216]: time="2025-09-13T00:06:04.249884885Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:06:04.250081 env[1216]: time="2025-09-13T00:06:04.249978583Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:06:04.250081 env[1216]: time="2025-09-13T00:06:04.250006388Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:06:04.250237 env[1216]: time="2025-09-13T00:06:04.250206906Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/cf6ef502db8e0442603a9bc81ae65bf52cb17f4a3595459b2dca308febd8244d pid=2033 runtime=io.containerd.runc.v2 Sep 13 00:06:04.260842 systemd[1]: Started cri-containerd-e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d.scope. Sep 13 00:06:04.270059 systemd[1]: Started cri-containerd-cf6ef502db8e0442603a9bc81ae65bf52cb17f4a3595459b2dca308febd8244d.scope. Sep 13 00:06:04.272043 kubelet[1920]: E0913 00:06:04.270867 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.274348 env[1216]: time="2025-09-13T00:06:04.274305617Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-8lt7r,Uid:3fa19538-2cb2-498e-b692-f575000fce93,Namespace:kube-system,Attempt:0,}" Sep 13 00:06:04.303284 env[1216]: time="2025-09-13T00:06:04.300586136Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:06:04.303284 env[1216]: time="2025-09-13T00:06:04.302965742Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:06:04.303284 env[1216]: time="2025-09-13T00:06:04.302977704Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:06:04.303513 env[1216]: time="2025-09-13T00:06:04.303316887Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zxhk9,Uid:0498e092-a3ef-4562-8c1f-832bdfc5c680,Namespace:kube-system,Attempt:0,} returns sandbox id \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\"" Sep 13 00:06:04.303544 env[1216]: time="2025-09-13T00:06:04.303500962Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801 pid=2097 runtime=io.containerd.runc.v2 Sep 13 00:06:04.304104 kubelet[1920]: E0913 00:06:04.304077 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.305093 env[1216]: time="2025-09-13T00:06:04.304699466Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-rb8kl,Uid:1baf9ac8-ee06-4adb-9114-bb30c12318aa,Namespace:kube-system,Attempt:0,} returns sandbox id \"cf6ef502db8e0442603a9bc81ae65bf52cb17f4a3595459b2dca308febd8244d\"" Sep 13 00:06:04.305200 env[1216]: time="2025-09-13T00:06:04.305171275Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 13 00:06:04.305248 kubelet[1920]: E0913 00:06:04.305145 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.307133 env[1216]: time="2025-09-13T00:06:04.307095875Z" level=info msg="CreateContainer within sandbox \"cf6ef502db8e0442603a9bc81ae65bf52cb17f4a3595459b2dca308febd8244d\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 13 00:06:04.322558 systemd[1]: Started cri-containerd-50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801.scope. Sep 13 00:06:04.333378 env[1216]: time="2025-09-13T00:06:04.333316583Z" level=info msg="CreateContainer within sandbox \"cf6ef502db8e0442603a9bc81ae65bf52cb17f4a3595459b2dca308febd8244d\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"7b3c626b220448b932ca0c7c2edbabc2c7800a064ee1ba162bf1cfb170bbfa10\"" Sep 13 00:06:04.335675 env[1216]: time="2025-09-13T00:06:04.335591849Z" level=info msg="StartContainer for \"7b3c626b220448b932ca0c7c2edbabc2c7800a064ee1ba162bf1cfb170bbfa10\"" Sep 13 00:06:04.359175 env[1216]: time="2025-09-13T00:06:04.359124174Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-8lt7r,Uid:3fa19538-2cb2-498e-b692-f575000fce93,Namespace:kube-system,Attempt:0,} returns sandbox id \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\"" Sep 13 00:06:04.360739 kubelet[1920]: E0913 00:06:04.360312 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:04.362998 systemd[1]: Started cri-containerd-7b3c626b220448b932ca0c7c2edbabc2c7800a064ee1ba162bf1cfb170bbfa10.scope. Sep 13 00:06:04.401848 env[1216]: time="2025-09-13T00:06:04.397231188Z" level=info msg="StartContainer for \"7b3c626b220448b932ca0c7c2edbabc2c7800a064ee1ba162bf1cfb170bbfa10\" returns successfully" Sep 13 00:06:04.756829 kubelet[1920]: E0913 00:06:04.755946 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:05.798748 kubelet[1920]: E0913 00:06:05.797163 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:05.818707 kubelet[1920]: I0913 00:06:05.818577 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-rb8kl" podStartSLOduration=2.8182452590000002 podStartE2EDuration="2.818245259s" podCreationTimestamp="2025-09-13 00:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:06:04.7687835 +0000 UTC m=+6.152801172" watchObservedRunningTime="2025-09-13 00:06:05.818245259 +0000 UTC m=+7.202262931" Sep 13 00:06:06.759891 kubelet[1920]: E0913 00:06:06.759862 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:09.097888 kubelet[1920]: E0913 00:06:09.097846 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:09.765110 kubelet[1920]: E0913 00:06:09.765075 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:10.755042 kubelet[1920]: E0913 00:06:10.754757 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:12.936673 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3572178031.mount: Deactivated successfully. Sep 13 00:06:12.983819 update_engine[1208]: I0913 00:06:12.983770 1208 update_attempter.cc:509] Updating boot flags... Sep 13 00:06:15.261643 env[1216]: time="2025-09-13T00:06:15.261593800Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:15.263415 env[1216]: time="2025-09-13T00:06:15.263377828Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:15.265062 env[1216]: time="2025-09-13T00:06:15.265030082Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:15.265672 env[1216]: time="2025-09-13T00:06:15.265638826Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Sep 13 00:06:15.271539 env[1216]: time="2025-09-13T00:06:15.271502563Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 13 00:06:15.277779 env[1216]: time="2025-09-13T00:06:15.277751261Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:06:15.304902 env[1216]: time="2025-09-13T00:06:15.304852995Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\"" Sep 13 00:06:15.305402 env[1216]: time="2025-09-13T00:06:15.305376010Z" level=info msg="StartContainer for \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\"" Sep 13 00:06:15.341457 systemd[1]: Started cri-containerd-b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240.scope. Sep 13 00:06:15.384600 systemd[1]: cri-containerd-b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240.scope: Deactivated successfully. Sep 13 00:06:15.411847 env[1216]: time="2025-09-13T00:06:15.411790976Z" level=info msg="StartContainer for \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\" returns successfully" Sep 13 00:06:15.449741 env[1216]: time="2025-09-13T00:06:15.449678765Z" level=info msg="shim disconnected" id=b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240 Sep 13 00:06:15.449926 env[1216]: time="2025-09-13T00:06:15.449750933Z" level=warning msg="cleaning up after shim disconnected" id=b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240 namespace=k8s.io Sep 13 00:06:15.449926 env[1216]: time="2025-09-13T00:06:15.449763574Z" level=info msg="cleaning up dead shim" Sep 13 00:06:15.456402 env[1216]: time="2025-09-13T00:06:15.456351788Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:06:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2370 runtime=io.containerd.runc.v2\n" Sep 13 00:06:15.782338 kubelet[1920]: E0913 00:06:15.782300 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:15.786050 env[1216]: time="2025-09-13T00:06:15.785982178Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:06:15.810931 env[1216]: time="2025-09-13T00:06:15.810863358Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\"" Sep 13 00:06:15.811661 env[1216]: time="2025-09-13T00:06:15.811628039Z" level=info msg="StartContainer for \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\"" Sep 13 00:06:15.833067 systemd[1]: Started cri-containerd-168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05.scope. Sep 13 00:06:15.869918 env[1216]: time="2025-09-13T00:06:15.869858051Z" level=info msg="StartContainer for \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\" returns successfully" Sep 13 00:06:15.881552 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:06:15.881807 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:06:15.881976 systemd[1]: Stopping systemd-sysctl.service... Sep 13 00:06:15.883586 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:06:15.887086 systemd[1]: cri-containerd-168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05.scope: Deactivated successfully. Sep 13 00:06:15.892256 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:06:15.909890 env[1216]: time="2025-09-13T00:06:15.909824379Z" level=info msg="shim disconnected" id=168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05 Sep 13 00:06:15.909890 env[1216]: time="2025-09-13T00:06:15.909888266Z" level=warning msg="cleaning up after shim disconnected" id=168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05 namespace=k8s.io Sep 13 00:06:15.910084 env[1216]: time="2025-09-13T00:06:15.909899107Z" level=info msg="cleaning up dead shim" Sep 13 00:06:15.922382 env[1216]: time="2025-09-13T00:06:15.922331256Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:06:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2436 runtime=io.containerd.runc.v2\n" Sep 13 00:06:16.297582 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240-rootfs.mount: Deactivated successfully. Sep 13 00:06:16.363609 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount274929898.mount: Deactivated successfully. Sep 13 00:06:16.787281 kubelet[1920]: E0913 00:06:16.787242 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:16.801822 env[1216]: time="2025-09-13T00:06:16.799705173Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:06:16.813168 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1855465081.mount: Deactivated successfully. Sep 13 00:06:16.819945 env[1216]: time="2025-09-13T00:06:16.819887878Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\"" Sep 13 00:06:16.820685 env[1216]: time="2025-09-13T00:06:16.820658835Z" level=info msg="StartContainer for \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\"" Sep 13 00:06:16.842386 systemd[1]: Started cri-containerd-f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544.scope. Sep 13 00:06:16.876843 systemd[1]: cri-containerd-f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544.scope: Deactivated successfully. Sep 13 00:06:16.887286 env[1216]: time="2025-09-13T00:06:16.887228316Z" level=info msg="StartContainer for \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\" returns successfully" Sep 13 00:06:16.914331 env[1216]: time="2025-09-13T00:06:16.914282951Z" level=info msg="shim disconnected" id=f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544 Sep 13 00:06:16.914331 env[1216]: time="2025-09-13T00:06:16.914331315Z" level=warning msg="cleaning up after shim disconnected" id=f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544 namespace=k8s.io Sep 13 00:06:16.914568 env[1216]: time="2025-09-13T00:06:16.914343757Z" level=info msg="cleaning up dead shim" Sep 13 00:06:16.925103 env[1216]: time="2025-09-13T00:06:16.925051311Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:06:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2490 runtime=io.containerd.runc.v2\n" Sep 13 00:06:17.516581 env[1216]: time="2025-09-13T00:06:17.516503272Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:17.518800 env[1216]: time="2025-09-13T00:06:17.518768569Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:17.521194 env[1216]: time="2025-09-13T00:06:17.521162438Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:06:17.521734 env[1216]: time="2025-09-13T00:06:17.521690128Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Sep 13 00:06:17.527441 env[1216]: time="2025-09-13T00:06:17.527411596Z" level=info msg="CreateContainer within sandbox \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 13 00:06:17.549886 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2448727038.mount: Deactivated successfully. Sep 13 00:06:17.554419 env[1216]: time="2025-09-13T00:06:17.554363455Z" level=info msg="CreateContainer within sandbox \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\"" Sep 13 00:06:17.555353 env[1216]: time="2025-09-13T00:06:17.555307506Z" level=info msg="StartContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\"" Sep 13 00:06:17.580128 systemd[1]: Started cri-containerd-17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653.scope. Sep 13 00:06:17.657321 env[1216]: time="2025-09-13T00:06:17.657268864Z" level=info msg="StartContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" returns successfully" Sep 13 00:06:17.790958 kubelet[1920]: E0913 00:06:17.790858 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:17.794613 kubelet[1920]: E0913 00:06:17.794579 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:17.796974 env[1216]: time="2025-09-13T00:06:17.796915190Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:06:17.811474 kubelet[1920]: I0913 00:06:17.811412 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-8lt7r" podStartSLOduration=1.650030724 podStartE2EDuration="14.811394135s" podCreationTimestamp="2025-09-13 00:06:03 +0000 UTC" firstStartedPulling="2025-09-13 00:06:04.362486124 +0000 UTC m=+5.746503796" lastFinishedPulling="2025-09-13 00:06:17.523849535 +0000 UTC m=+18.907867207" observedRunningTime="2025-09-13 00:06:17.810437084 +0000 UTC m=+19.194454756" watchObservedRunningTime="2025-09-13 00:06:17.811394135 +0000 UTC m=+19.195411767" Sep 13 00:06:17.817056 env[1216]: time="2025-09-13T00:06:17.816969069Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\"" Sep 13 00:06:17.817988 env[1216]: time="2025-09-13T00:06:17.817956923Z" level=info msg="StartContainer for \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\"" Sep 13 00:06:17.836744 systemd[1]: Started cri-containerd-1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032.scope. Sep 13 00:06:17.866593 systemd[1]: cri-containerd-1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032.scope: Deactivated successfully. Sep 13 00:06:17.869632 env[1216]: time="2025-09-13T00:06:17.869585865Z" level=info msg="StartContainer for \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\" returns successfully" Sep 13 00:06:17.889228 env[1216]: time="2025-09-13T00:06:17.889181300Z" level=info msg="shim disconnected" id=1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032 Sep 13 00:06:17.889228 env[1216]: time="2025-09-13T00:06:17.889225824Z" level=warning msg="cleaning up after shim disconnected" id=1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032 namespace=k8s.io Sep 13 00:06:17.889228 env[1216]: time="2025-09-13T00:06:17.889235825Z" level=info msg="cleaning up dead shim" Sep 13 00:06:17.908272 env[1216]: time="2025-09-13T00:06:17.907597503Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:06:17Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2584 runtime=io.containerd.runc.v2\n" Sep 13 00:06:18.809297 kubelet[1920]: E0913 00:06:18.809268 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:18.812050 kubelet[1920]: E0913 00:06:18.810143 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:18.827600 env[1216]: time="2025-09-13T00:06:18.827402298Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:06:18.851024 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1958835001.mount: Deactivated successfully. Sep 13 00:06:18.856410 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount557076351.mount: Deactivated successfully. Sep 13 00:06:18.862659 env[1216]: time="2025-09-13T00:06:18.862518266Z" level=info msg="CreateContainer within sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\"" Sep 13 00:06:18.863160 env[1216]: time="2025-09-13T00:06:18.863090438Z" level=info msg="StartContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\"" Sep 13 00:06:18.881799 systemd[1]: Started cri-containerd-c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980.scope. Sep 13 00:06:18.929862 env[1216]: time="2025-09-13T00:06:18.929812054Z" level=info msg="StartContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" returns successfully" Sep 13 00:06:19.103131 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Sep 13 00:06:19.111581 kubelet[1920]: I0913 00:06:19.111541 1920 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 13 00:06:19.151228 kubelet[1920]: W0913 00:06:19.151177 1920 reflector.go:561] object-"kube-system"/"coredns": failed to list *v1.ConfigMap: configmaps "coredns" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object Sep 13 00:06:19.152006 kubelet[1920]: E0913 00:06:19.151959 1920 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"coredns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"coredns\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Sep 13 00:06:19.158129 systemd[1]: Created slice kubepods-burstable-pod275dceb3_0f60_4269_9f06_0709b7ab40b9.slice. Sep 13 00:06:19.162603 systemd[1]: Created slice kubepods-burstable-pod258c8c50_882e_4189_b148_a427542985f0.slice. Sep 13 00:06:19.337529 kubelet[1920]: I0913 00:06:19.337489 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/258c8c50-882e-4189-b148-a427542985f0-config-volume\") pod \"coredns-7c65d6cfc9-btkcw\" (UID: \"258c8c50-882e-4189-b148-a427542985f0\") " pod="kube-system/coredns-7c65d6cfc9-btkcw" Sep 13 00:06:19.337641 kubelet[1920]: I0913 00:06:19.337543 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/275dceb3-0f60-4269-9f06-0709b7ab40b9-config-volume\") pod \"coredns-7c65d6cfc9-f95rk\" (UID: \"275dceb3-0f60-4269-9f06-0709b7ab40b9\") " pod="kube-system/coredns-7c65d6cfc9-f95rk" Sep 13 00:06:19.337641 kubelet[1920]: I0913 00:06:19.337562 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lllf7\" (UniqueName: \"kubernetes.io/projected/275dceb3-0f60-4269-9f06-0709b7ab40b9-kube-api-access-lllf7\") pod \"coredns-7c65d6cfc9-f95rk\" (UID: \"275dceb3-0f60-4269-9f06-0709b7ab40b9\") " pod="kube-system/coredns-7c65d6cfc9-f95rk" Sep 13 00:06:19.337641 kubelet[1920]: I0913 00:06:19.337608 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnrlr\" (UniqueName: \"kubernetes.io/projected/258c8c50-882e-4189-b148-a427542985f0-kube-api-access-lnrlr\") pod \"coredns-7c65d6cfc9-btkcw\" (UID: \"258c8c50-882e-4189-b148-a427542985f0\") " pod="kube-system/coredns-7c65d6cfc9-btkcw" Sep 13 00:06:19.369744 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Sep 13 00:06:19.814669 kubelet[1920]: E0913 00:06:19.814637 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:19.843860 kubelet[1920]: I0913 00:06:19.843751 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-zxhk9" podStartSLOduration=5.87764097 podStartE2EDuration="16.843710425s" podCreationTimestamp="2025-09-13 00:06:03 +0000 UTC" firstStartedPulling="2025-09-13 00:06:04.304751796 +0000 UTC m=+5.688769468" lastFinishedPulling="2025-09-13 00:06:15.270821251 +0000 UTC m=+16.654838923" observedRunningTime="2025-09-13 00:06:19.84342612 +0000 UTC m=+21.227443792" watchObservedRunningTime="2025-09-13 00:06:19.843710425 +0000 UTC m=+21.227728097" Sep 13 00:06:20.442309 kubelet[1920]: E0913 00:06:20.442216 1920 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Sep 13 00:06:20.442458 kubelet[1920]: E0913 00:06:20.442331 1920 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/258c8c50-882e-4189-b148-a427542985f0-config-volume podName:258c8c50-882e-4189-b148-a427542985f0 nodeName:}" failed. No retries permitted until 2025-09-13 00:06:20.942303388 +0000 UTC m=+22.326321020 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/258c8c50-882e-4189-b148-a427542985f0-config-volume") pod "coredns-7c65d6cfc9-btkcw" (UID: "258c8c50-882e-4189-b148-a427542985f0") : failed to sync configmap cache: timed out waiting for the condition Sep 13 00:06:20.442458 kubelet[1920]: E0913 00:06:20.442224 1920 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Sep 13 00:06:20.442458 kubelet[1920]: E0913 00:06:20.442398 1920 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/275dceb3-0f60-4269-9f06-0709b7ab40b9-config-volume podName:275dceb3-0f60-4269-9f06-0709b7ab40b9 nodeName:}" failed. No retries permitted until 2025-09-13 00:06:20.942382115 +0000 UTC m=+22.326399747 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/275dceb3-0f60-4269-9f06-0709b7ab40b9-config-volume") pod "coredns-7c65d6cfc9-f95rk" (UID: "275dceb3-0f60-4269-9f06-0709b7ab40b9") : failed to sync configmap cache: timed out waiting for the condition Sep 13 00:06:20.817579 kubelet[1920]: E0913 00:06:20.816683 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:20.961985 kubelet[1920]: E0913 00:06:20.961943 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:20.963009 env[1216]: time="2025-09-13T00:06:20.962964238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-f95rk,Uid:275dceb3-0f60-4269-9f06-0709b7ab40b9,Namespace:kube-system,Attempt:0,}" Sep 13 00:06:20.964568 kubelet[1920]: E0913 00:06:20.964539 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:20.965156 env[1216]: time="2025-09-13T00:06:20.965113338Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-btkcw,Uid:258c8c50-882e-4189-b148-a427542985f0,Namespace:kube-system,Attempt:0,}" Sep 13 00:06:21.012144 systemd-networkd[1044]: cilium_host: Link UP Sep 13 00:06:21.013201 systemd-networkd[1044]: cilium_net: Link UP Sep 13 00:06:21.014038 systemd-networkd[1044]: cilium_net: Gained carrier Sep 13 00:06:21.015592 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Sep 13 00:06:21.015680 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Sep 13 00:06:21.015820 systemd-networkd[1044]: cilium_host: Gained carrier Sep 13 00:06:21.100549 systemd-networkd[1044]: cilium_vxlan: Link UP Sep 13 00:06:21.100556 systemd-networkd[1044]: cilium_vxlan: Gained carrier Sep 13 00:06:21.379750 kernel: NET: Registered PF_ALG protocol family Sep 13 00:06:21.516109 systemd-networkd[1044]: cilium_net: Gained IPv6LL Sep 13 00:06:21.818160 kubelet[1920]: E0913 00:06:21.818056 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:21.924666 systemd-networkd[1044]: cilium_host: Gained IPv6LL Sep 13 00:06:22.023596 systemd-networkd[1044]: lxc_health: Link UP Sep 13 00:06:22.034500 systemd-networkd[1044]: lxc_health: Gained carrier Sep 13 00:06:22.034842 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:06:22.535347 systemd-networkd[1044]: lxc002854b94ca0: Link UP Sep 13 00:06:22.535745 kernel: eth0: renamed from tmp50573 Sep 13 00:06:22.543173 systemd-networkd[1044]: lxc002854b94ca0: Gained carrier Sep 13 00:06:22.543558 systemd-networkd[1044]: lxcc412f24d89f2: Link UP Sep 13 00:06:22.543859 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc002854b94ca0: link becomes ready Sep 13 00:06:22.557764 kernel: eth0: renamed from tmpda999 Sep 13 00:06:22.562760 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcc412f24d89f2: link becomes ready Sep 13 00:06:22.562701 systemd-networkd[1044]: lxcc412f24d89f2: Gained carrier Sep 13 00:06:22.821076 kubelet[1920]: E0913 00:06:22.820952 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:22.884919 systemd-networkd[1044]: cilium_vxlan: Gained IPv6LL Sep 13 00:06:23.972027 systemd-networkd[1044]: lxc_health: Gained IPv6LL Sep 13 00:06:24.035935 systemd-networkd[1044]: lxcc412f24d89f2: Gained IPv6LL Sep 13 00:06:24.419966 systemd-networkd[1044]: lxc002854b94ca0: Gained IPv6LL Sep 13 00:06:26.274025 env[1216]: time="2025-09-13T00:06:26.273950755Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:06:26.274367 env[1216]: time="2025-09-13T00:06:26.273993598Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:06:26.274367 env[1216]: time="2025-09-13T00:06:26.274004119Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:06:26.274367 env[1216]: time="2025-09-13T00:06:26.274140928Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/5057361c5223ee4feded93665fed03010afd620021acb6a82b58f0156346e712 pid=3144 runtime=io.containerd.runc.v2 Sep 13 00:06:26.278858 env[1216]: time="2025-09-13T00:06:26.277514467Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:06:26.278858 env[1216]: time="2025-09-13T00:06:26.277559910Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:06:26.278858 env[1216]: time="2025-09-13T00:06:26.277572831Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:06:26.278858 env[1216]: time="2025-09-13T00:06:26.277682078Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/da999fe449dc16fa9a9dab98166ed77e936f7c6308f21b76447674fa6125206a pid=3161 runtime=io.containerd.runc.v2 Sep 13 00:06:26.291781 systemd[1]: Started cri-containerd-da999fe449dc16fa9a9dab98166ed77e936f7c6308f21b76447674fa6125206a.scope. Sep 13 00:06:26.296249 systemd[1]: Started cri-containerd-5057361c5223ee4feded93665fed03010afd620021acb6a82b58f0156346e712.scope. Sep 13 00:06:26.315178 systemd-resolved[1156]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 13 00:06:26.318462 systemd-resolved[1156]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 13 00:06:26.336972 env[1216]: time="2025-09-13T00:06:26.336617471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-f95rk,Uid:275dceb3-0f60-4269-9f06-0709b7ab40b9,Namespace:kube-system,Attempt:0,} returns sandbox id \"5057361c5223ee4feded93665fed03010afd620021acb6a82b58f0156346e712\"" Sep 13 00:06:26.337552 kubelet[1920]: E0913 00:06:26.337370 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:26.339376 env[1216]: time="2025-09-13T00:06:26.339327007Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-btkcw,Uid:258c8c50-882e-4189-b148-a427542985f0,Namespace:kube-system,Attempt:0,} returns sandbox id \"da999fe449dc16fa9a9dab98166ed77e936f7c6308f21b76447674fa6125206a\"" Sep 13 00:06:26.339902 env[1216]: time="2025-09-13T00:06:26.339868962Z" level=info msg="CreateContainer within sandbox \"5057361c5223ee4feded93665fed03010afd620021acb6a82b58f0156346e712\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:06:26.341221 kubelet[1920]: E0913 00:06:26.341197 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:26.344173 env[1216]: time="2025-09-13T00:06:26.344128959Z" level=info msg="CreateContainer within sandbox \"da999fe449dc16fa9a9dab98166ed77e936f7c6308f21b76447674fa6125206a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:06:26.356001 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2395535189.mount: Deactivated successfully. Sep 13 00:06:26.360771 env[1216]: time="2025-09-13T00:06:26.360702517Z" level=info msg="CreateContainer within sandbox \"5057361c5223ee4feded93665fed03010afd620021acb6a82b58f0156346e712\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"48423e334f51b7a235fb5286c56e01441f4df83e82e9e2254c0e1b3a9019038f\"" Sep 13 00:06:26.361401 env[1216]: time="2025-09-13T00:06:26.361355959Z" level=info msg="StartContainer for \"48423e334f51b7a235fb5286c56e01441f4df83e82e9e2254c0e1b3a9019038f\"" Sep 13 00:06:26.366149 env[1216]: time="2025-09-13T00:06:26.366100068Z" level=info msg="CreateContainer within sandbox \"da999fe449dc16fa9a9dab98166ed77e936f7c6308f21b76447674fa6125206a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b05442e26bcd04543bff217604d9c8ef28460aa93aaf5640d772c516290587f7\"" Sep 13 00:06:26.366851 env[1216]: time="2025-09-13T00:06:26.366697827Z" level=info msg="StartContainer for \"b05442e26bcd04543bff217604d9c8ef28460aa93aaf5640d772c516290587f7\"" Sep 13 00:06:26.376693 systemd[1]: Started cri-containerd-48423e334f51b7a235fb5286c56e01441f4df83e82e9e2254c0e1b3a9019038f.scope. Sep 13 00:06:26.381522 systemd[1]: Started cri-containerd-b05442e26bcd04543bff217604d9c8ef28460aa93aaf5640d772c516290587f7.scope. Sep 13 00:06:26.415216 env[1216]: time="2025-09-13T00:06:26.415169059Z" level=info msg="StartContainer for \"48423e334f51b7a235fb5286c56e01441f4df83e82e9e2254c0e1b3a9019038f\" returns successfully" Sep 13 00:06:26.423214 env[1216]: time="2025-09-13T00:06:26.423169939Z" level=info msg="StartContainer for \"b05442e26bcd04543bff217604d9c8ef28460aa93aaf5640d772c516290587f7\" returns successfully" Sep 13 00:06:26.829573 kubelet[1920]: E0913 00:06:26.829524 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:26.832420 kubelet[1920]: E0913 00:06:26.832388 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:26.842895 kubelet[1920]: I0913 00:06:26.842853 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-f95rk" podStartSLOduration=23.84283199 podStartE2EDuration="23.84283199s" podCreationTimestamp="2025-09-13 00:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:06:26.84221895 +0000 UTC m=+28.226236622" watchObservedRunningTime="2025-09-13 00:06:26.84283199 +0000 UTC m=+28.226849662" Sep 13 00:06:27.833785 kubelet[1920]: E0913 00:06:27.833755 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:27.834095 kubelet[1920]: E0913 00:06:27.833808 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:28.835909 kubelet[1920]: E0913 00:06:28.835873 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:28.836327 kubelet[1920]: E0913 00:06:28.836310 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:31.316329 systemd[1]: Started sshd@5-10.0.0.20:22-10.0.0.1:39186.service. Sep 13 00:06:31.364501 sshd[3304]: Accepted publickey for core from 10.0.0.1 port 39186 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:31.366315 sshd[3304]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:31.369747 systemd-logind[1206]: New session 6 of user core. Sep 13 00:06:31.370927 systemd[1]: Started session-6.scope. Sep 13 00:06:31.486935 sshd[3304]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:31.489776 systemd[1]: sshd@5-10.0.0.20:22-10.0.0.1:39186.service: Deactivated successfully. Sep 13 00:06:31.490490 systemd[1]: session-6.scope: Deactivated successfully. Sep 13 00:06:31.491040 systemd-logind[1206]: Session 6 logged out. Waiting for processes to exit. Sep 13 00:06:31.491695 systemd-logind[1206]: Removed session 6. Sep 13 00:06:32.577030 kubelet[1920]: I0913 00:06:32.576969 1920 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 00:06:32.577544 kubelet[1920]: E0913 00:06:32.577399 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:32.608104 kubelet[1920]: I0913 00:06:32.608039 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-btkcw" podStartSLOduration=29.608022531 podStartE2EDuration="29.608022531s" podCreationTimestamp="2025-09-13 00:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:06:26.86650729 +0000 UTC m=+28.250524922" watchObservedRunningTime="2025-09-13 00:06:32.608022531 +0000 UTC m=+33.992040203" Sep 13 00:06:32.842310 kubelet[1920]: E0913 00:06:32.842185 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:06:36.494538 systemd[1]: Started sshd@6-10.0.0.20:22-10.0.0.1:39188.service. Sep 13 00:06:36.545469 sshd[3322]: Accepted publickey for core from 10.0.0.1 port 39188 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:36.547176 sshd[3322]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:36.552336 systemd[1]: Started session-7.scope. Sep 13 00:06:36.552618 systemd-logind[1206]: New session 7 of user core. Sep 13 00:06:36.669998 sshd[3322]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:36.672774 systemd[1]: sshd@6-10.0.0.20:22-10.0.0.1:39188.service: Deactivated successfully. Sep 13 00:06:36.673438 systemd[1]: session-7.scope: Deactivated successfully. Sep 13 00:06:36.674224 systemd-logind[1206]: Session 7 logged out. Waiting for processes to exit. Sep 13 00:06:36.674879 systemd-logind[1206]: Removed session 7. Sep 13 00:06:41.674523 systemd[1]: Started sshd@7-10.0.0.20:22-10.0.0.1:37976.service. Sep 13 00:06:41.725901 sshd[3337]: Accepted publickey for core from 10.0.0.1 port 37976 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:41.728002 sshd[3337]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:41.732245 systemd-logind[1206]: New session 8 of user core. Sep 13 00:06:41.733326 systemd[1]: Started session-8.scope. Sep 13 00:06:41.851398 sshd[3337]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:41.855669 systemd[1]: sshd@7-10.0.0.20:22-10.0.0.1:37976.service: Deactivated successfully. Sep 13 00:06:41.856402 systemd[1]: session-8.scope: Deactivated successfully. Sep 13 00:06:41.857224 systemd-logind[1206]: Session 8 logged out. Waiting for processes to exit. Sep 13 00:06:41.858250 systemd-logind[1206]: Removed session 8. Sep 13 00:06:46.856202 systemd[1]: Started sshd@8-10.0.0.20:22-10.0.0.1:37978.service. Sep 13 00:06:46.909123 sshd[3352]: Accepted publickey for core from 10.0.0.1 port 37978 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:46.910757 sshd[3352]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:46.916400 systemd[1]: Started session-9.scope. Sep 13 00:06:46.917666 systemd-logind[1206]: New session 9 of user core. Sep 13 00:06:47.066528 sshd[3352]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:47.071048 systemd[1]: session-9.scope: Deactivated successfully. Sep 13 00:06:47.071842 systemd[1]: sshd@8-10.0.0.20:22-10.0.0.1:37978.service: Deactivated successfully. Sep 13 00:06:47.072485 systemd-logind[1206]: Session 9 logged out. Waiting for processes to exit. Sep 13 00:06:47.075299 systemd-logind[1206]: Removed session 9. Sep 13 00:06:52.076813 systemd[1]: Started sshd@9-10.0.0.20:22-10.0.0.1:41534.service. Sep 13 00:06:52.126787 sshd[3368]: Accepted publickey for core from 10.0.0.1 port 41534 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:52.128130 sshd[3368]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:52.135609 systemd[1]: Started session-10.scope. Sep 13 00:06:52.135690 systemd-logind[1206]: New session 10 of user core. Sep 13 00:06:52.265122 sshd[3368]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:52.269251 systemd[1]: Started sshd@10-10.0.0.20:22-10.0.0.1:41546.service. Sep 13 00:06:52.269890 systemd[1]: sshd@9-10.0.0.20:22-10.0.0.1:41534.service: Deactivated successfully. Sep 13 00:06:52.270878 systemd[1]: session-10.scope: Deactivated successfully. Sep 13 00:06:52.272259 systemd-logind[1206]: Session 10 logged out. Waiting for processes to exit. Sep 13 00:06:52.273358 systemd-logind[1206]: Removed session 10. Sep 13 00:06:52.314315 sshd[3381]: Accepted publickey for core from 10.0.0.1 port 41546 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:52.315820 sshd[3381]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:52.319827 systemd-logind[1206]: New session 11 of user core. Sep 13 00:06:52.320402 systemd[1]: Started session-11.scope. Sep 13 00:06:52.505586 sshd[3381]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:52.512314 systemd[1]: Started sshd@11-10.0.0.20:22-10.0.0.1:41548.service. Sep 13 00:06:52.513412 systemd[1]: sshd@10-10.0.0.20:22-10.0.0.1:41546.service: Deactivated successfully. Sep 13 00:06:52.514311 systemd[1]: session-11.scope: Deactivated successfully. Sep 13 00:06:52.515229 systemd-logind[1206]: Session 11 logged out. Waiting for processes to exit. Sep 13 00:06:52.524533 systemd-logind[1206]: Removed session 11. Sep 13 00:06:52.568537 sshd[3395]: Accepted publickey for core from 10.0.0.1 port 41548 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:52.570279 sshd[3395]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:52.573635 systemd-logind[1206]: New session 12 of user core. Sep 13 00:06:52.574505 systemd[1]: Started session-12.scope. Sep 13 00:06:52.693887 sshd[3395]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:52.697314 systemd[1]: sshd@11-10.0.0.20:22-10.0.0.1:41548.service: Deactivated successfully. Sep 13 00:06:52.698244 systemd[1]: session-12.scope: Deactivated successfully. Sep 13 00:06:52.699419 systemd-logind[1206]: Session 12 logged out. Waiting for processes to exit. Sep 13 00:06:52.700421 systemd-logind[1206]: Removed session 12. Sep 13 00:06:57.700549 systemd[1]: Started sshd@12-10.0.0.20:22-10.0.0.1:41550.service. Sep 13 00:06:57.743840 sshd[3412]: Accepted publickey for core from 10.0.0.1 port 41550 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:06:57.745477 sshd[3412]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:06:57.750844 systemd-logind[1206]: New session 13 of user core. Sep 13 00:06:57.751619 systemd[1]: Started session-13.scope. Sep 13 00:06:57.893022 sshd[3412]: pam_unix(sshd:session): session closed for user core Sep 13 00:06:57.895412 systemd[1]: sshd@12-10.0.0.20:22-10.0.0.1:41550.service: Deactivated successfully. Sep 13 00:06:57.896245 systemd[1]: session-13.scope: Deactivated successfully. Sep 13 00:06:57.896979 systemd-logind[1206]: Session 13 logged out. Waiting for processes to exit. Sep 13 00:06:57.898143 systemd-logind[1206]: Removed session 13. Sep 13 00:07:02.901759 systemd[1]: Started sshd@13-10.0.0.20:22-10.0.0.1:59916.service. Sep 13 00:07:02.945420 sshd[3427]: Accepted publickey for core from 10.0.0.1 port 59916 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:02.947066 sshd[3427]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:02.952147 systemd-logind[1206]: New session 14 of user core. Sep 13 00:07:02.952872 systemd[1]: Started session-14.scope. Sep 13 00:07:03.085348 sshd[3427]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:03.090677 systemd[1]: sshd@13-10.0.0.20:22-10.0.0.1:59916.service: Deactivated successfully. Sep 13 00:07:03.091688 systemd[1]: session-14.scope: Deactivated successfully. Sep 13 00:07:03.093472 systemd-logind[1206]: Session 14 logged out. Waiting for processes to exit. Sep 13 00:07:03.096051 systemd[1]: Started sshd@14-10.0.0.20:22-10.0.0.1:59922.service. Sep 13 00:07:03.097822 systemd-logind[1206]: Removed session 14. Sep 13 00:07:03.140512 sshd[3440]: Accepted publickey for core from 10.0.0.1 port 59922 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:03.141829 sshd[3440]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:03.145382 systemd-logind[1206]: New session 15 of user core. Sep 13 00:07:03.146983 systemd[1]: Started session-15.scope. Sep 13 00:07:03.371929 sshd[3440]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:03.373375 systemd[1]: Started sshd@15-10.0.0.20:22-10.0.0.1:59924.service. Sep 13 00:07:03.379414 systemd[1]: sshd@14-10.0.0.20:22-10.0.0.1:59922.service: Deactivated successfully. Sep 13 00:07:03.380060 systemd[1]: session-15.scope: Deactivated successfully. Sep 13 00:07:03.380961 systemd-logind[1206]: Session 15 logged out. Waiting for processes to exit. Sep 13 00:07:03.382029 systemd-logind[1206]: Removed session 15. Sep 13 00:07:03.418699 sshd[3450]: Accepted publickey for core from 10.0.0.1 port 59924 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:03.419965 sshd[3450]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:03.426450 systemd[1]: Started session-16.scope. Sep 13 00:07:03.427000 systemd-logind[1206]: New session 16 of user core. Sep 13 00:07:04.745782 sshd[3450]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:04.754163 systemd[1]: Started sshd@16-10.0.0.20:22-10.0.0.1:59940.service. Sep 13 00:07:04.755213 systemd[1]: sshd@15-10.0.0.20:22-10.0.0.1:59924.service: Deactivated successfully. Sep 13 00:07:04.757317 systemd[1]: session-16.scope: Deactivated successfully. Sep 13 00:07:04.759112 systemd-logind[1206]: Session 16 logged out. Waiting for processes to exit. Sep 13 00:07:04.760204 systemd-logind[1206]: Removed session 16. Sep 13 00:07:04.808107 sshd[3483]: Accepted publickey for core from 10.0.0.1 port 59940 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:04.809769 sshd[3483]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:04.814224 systemd[1]: Started session-17.scope. Sep 13 00:07:04.814383 systemd-logind[1206]: New session 17 of user core. Sep 13 00:07:05.047975 sshd[3483]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:05.052945 systemd[1]: sshd@16-10.0.0.20:22-10.0.0.1:59940.service: Deactivated successfully. Sep 13 00:07:05.053608 systemd[1]: session-17.scope: Deactivated successfully. Sep 13 00:07:05.055618 systemd-logind[1206]: Session 17 logged out. Waiting for processes to exit. Sep 13 00:07:05.056859 systemd[1]: Started sshd@17-10.0.0.20:22-10.0.0.1:59946.service. Sep 13 00:07:05.060124 systemd-logind[1206]: Removed session 17. Sep 13 00:07:05.100202 sshd[3495]: Accepted publickey for core from 10.0.0.1 port 59946 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:05.101485 sshd[3495]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:05.105557 systemd-logind[1206]: New session 18 of user core. Sep 13 00:07:05.108063 systemd[1]: Started session-18.scope. Sep 13 00:07:05.234595 sshd[3495]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:05.237078 systemd[1]: sshd@17-10.0.0.20:22-10.0.0.1:59946.service: Deactivated successfully. Sep 13 00:07:05.237786 systemd[1]: session-18.scope: Deactivated successfully. Sep 13 00:07:05.238291 systemd-logind[1206]: Session 18 logged out. Waiting for processes to exit. Sep 13 00:07:05.238998 systemd-logind[1206]: Removed session 18. Sep 13 00:07:06.739278 kubelet[1920]: E0913 00:07:06.737124 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:10.239226 systemd[1]: Started sshd@18-10.0.0.20:22-10.0.0.1:53926.service. Sep 13 00:07:10.283331 sshd[3510]: Accepted publickey for core from 10.0.0.1 port 53926 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:10.284661 sshd[3510]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:10.288284 systemd-logind[1206]: New session 19 of user core. Sep 13 00:07:10.289739 systemd[1]: Started session-19.scope. Sep 13 00:07:10.406884 sshd[3510]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:10.409379 systemd-logind[1206]: Session 19 logged out. Waiting for processes to exit. Sep 13 00:07:10.409582 systemd[1]: sshd@18-10.0.0.20:22-10.0.0.1:53926.service: Deactivated successfully. Sep 13 00:07:10.410321 systemd[1]: session-19.scope: Deactivated successfully. Sep 13 00:07:10.411102 systemd-logind[1206]: Removed session 19. Sep 13 00:07:11.736316 kubelet[1920]: E0913 00:07:11.736277 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:15.412344 systemd[1]: Started sshd@19-10.0.0.20:22-10.0.0.1:53928.service. Sep 13 00:07:15.462420 sshd[3524]: Accepted publickey for core from 10.0.0.1 port 53928 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:15.464358 sshd[3524]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:15.469648 systemd-logind[1206]: New session 20 of user core. Sep 13 00:07:15.471004 systemd[1]: Started session-20.scope. Sep 13 00:07:15.617202 sshd[3524]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:15.621408 systemd[1]: sshd@19-10.0.0.20:22-10.0.0.1:53928.service: Deactivated successfully. Sep 13 00:07:15.622157 systemd[1]: session-20.scope: Deactivated successfully. Sep 13 00:07:15.622697 systemd-logind[1206]: Session 20 logged out. Waiting for processes to exit. Sep 13 00:07:15.623632 systemd-logind[1206]: Removed session 20. Sep 13 00:07:16.737330 kubelet[1920]: E0913 00:07:16.737288 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:20.624538 systemd[1]: Started sshd@20-10.0.0.20:22-10.0.0.1:53152.service. Sep 13 00:07:20.666244 sshd[3537]: Accepted publickey for core from 10.0.0.1 port 53152 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:20.667556 sshd[3537]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:20.671803 systemd-logind[1206]: New session 21 of user core. Sep 13 00:07:20.672151 systemd[1]: Started session-21.scope. Sep 13 00:07:20.781022 sshd[3537]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:20.785146 systemd[1]: Started sshd@21-10.0.0.20:22-10.0.0.1:53156.service. Sep 13 00:07:20.785747 systemd[1]: sshd@20-10.0.0.20:22-10.0.0.1:53152.service: Deactivated successfully. Sep 13 00:07:20.786759 systemd[1]: session-21.scope: Deactivated successfully. Sep 13 00:07:20.787341 systemd-logind[1206]: Session 21 logged out. Waiting for processes to exit. Sep 13 00:07:20.788191 systemd-logind[1206]: Removed session 21. Sep 13 00:07:20.827886 sshd[3549]: Accepted publickey for core from 10.0.0.1 port 53156 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:20.829277 sshd[3549]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:20.832599 systemd-logind[1206]: New session 22 of user core. Sep 13 00:07:20.833551 systemd[1]: Started session-22.scope. Sep 13 00:07:22.572482 env[1216]: time="2025-09-13T00:07:22.572431792Z" level=info msg="StopContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" with timeout 30 (s)" Sep 13 00:07:22.572856 env[1216]: time="2025-09-13T00:07:22.572822592Z" level=info msg="Stop container \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" with signal terminated" Sep 13 00:07:22.584619 systemd[1]: run-containerd-runc-k8s.io-c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980-runc.KWvBgL.mount: Deactivated successfully. Sep 13 00:07:22.587743 systemd[1]: cri-containerd-17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653.scope: Deactivated successfully. Sep 13 00:07:22.609124 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653-rootfs.mount: Deactivated successfully. Sep 13 00:07:22.616681 env[1216]: time="2025-09-13T00:07:22.616411721Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:07:22.619755 env[1216]: time="2025-09-13T00:07:22.619703242Z" level=info msg="shim disconnected" id=17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653 Sep 13 00:07:22.619865 env[1216]: time="2025-09-13T00:07:22.619756002Z" level=warning msg="cleaning up after shim disconnected" id=17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653 namespace=k8s.io Sep 13 00:07:22.619865 env[1216]: time="2025-09-13T00:07:22.619767082Z" level=info msg="cleaning up dead shim" Sep 13 00:07:22.622368 env[1216]: time="2025-09-13T00:07:22.622317483Z" level=info msg="StopContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" with timeout 2 (s)" Sep 13 00:07:22.622656 env[1216]: time="2025-09-13T00:07:22.622638123Z" level=info msg="Stop container \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" with signal terminated" Sep 13 00:07:22.628996 systemd-networkd[1044]: lxc_health: Link DOWN Sep 13 00:07:22.629003 systemd-networkd[1044]: lxc_health: Lost carrier Sep 13 00:07:22.630595 env[1216]: time="2025-09-13T00:07:22.630556084Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3598 runtime=io.containerd.runc.v2\n" Sep 13 00:07:22.632885 env[1216]: time="2025-09-13T00:07:22.632843285Z" level=info msg="StopContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" returns successfully" Sep 13 00:07:22.633534 env[1216]: time="2025-09-13T00:07:22.633503805Z" level=info msg="StopPodSandbox for \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\"" Sep 13 00:07:22.633594 env[1216]: time="2025-09-13T00:07:22.633574245Z" level=info msg="Container to stop \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.635351 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801-shm.mount: Deactivated successfully. Sep 13 00:07:22.641705 systemd[1]: cri-containerd-50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801.scope: Deactivated successfully. Sep 13 00:07:22.658134 systemd[1]: cri-containerd-c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980.scope: Deactivated successfully. Sep 13 00:07:22.658441 systemd[1]: cri-containerd-c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980.scope: Consumed 6.478s CPU time. Sep 13 00:07:22.681506 env[1216]: time="2025-09-13T00:07:22.681444375Z" level=info msg="shim disconnected" id=50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801 Sep 13 00:07:22.681506 env[1216]: time="2025-09-13T00:07:22.681498695Z" level=warning msg="cleaning up after shim disconnected" id=50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801 namespace=k8s.io Sep 13 00:07:22.681506 env[1216]: time="2025-09-13T00:07:22.681508015Z" level=info msg="cleaning up dead shim" Sep 13 00:07:22.681764 env[1216]: time="2025-09-13T00:07:22.681602375Z" level=info msg="shim disconnected" id=c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980 Sep 13 00:07:22.681764 env[1216]: time="2025-09-13T00:07:22.681648735Z" level=warning msg="cleaning up after shim disconnected" id=c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980 namespace=k8s.io Sep 13 00:07:22.681764 env[1216]: time="2025-09-13T00:07:22.681658855Z" level=info msg="cleaning up dead shim" Sep 13 00:07:22.688407 env[1216]: time="2025-09-13T00:07:22.688363417Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3651 runtime=io.containerd.runc.v2\n" Sep 13 00:07:22.688717 env[1216]: time="2025-09-13T00:07:22.688688217Z" level=info msg="TearDown network for sandbox \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\" successfully" Sep 13 00:07:22.688717 env[1216]: time="2025-09-13T00:07:22.688715337Z" level=info msg="StopPodSandbox for \"50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801\" returns successfully" Sep 13 00:07:22.690327 env[1216]: time="2025-09-13T00:07:22.690068977Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3652 runtime=io.containerd.runc.v2\n" Sep 13 00:07:22.706638 env[1216]: time="2025-09-13T00:07:22.706576901Z" level=info msg="StopContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" returns successfully" Sep 13 00:07:22.707419 env[1216]: time="2025-09-13T00:07:22.707385021Z" level=info msg="StopPodSandbox for \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\"" Sep 13 00:07:22.707488 env[1216]: time="2025-09-13T00:07:22.707442341Z" level=info msg="Container to stop \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.707488 env[1216]: time="2025-09-13T00:07:22.707457901Z" level=info msg="Container to stop \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.707488 env[1216]: time="2025-09-13T00:07:22.707470581Z" level=info msg="Container to stop \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.707488 env[1216]: time="2025-09-13T00:07:22.707481861Z" level=info msg="Container to stop \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.707594 env[1216]: time="2025-09-13T00:07:22.707492181Z" level=info msg="Container to stop \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:07:22.715132 systemd[1]: cri-containerd-e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d.scope: Deactivated successfully. Sep 13 00:07:22.738165 env[1216]: time="2025-09-13T00:07:22.738116708Z" level=info msg="shim disconnected" id=e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d Sep 13 00:07:22.738165 env[1216]: time="2025-09-13T00:07:22.738169468Z" level=warning msg="cleaning up after shim disconnected" id=e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d namespace=k8s.io Sep 13 00:07:22.738374 env[1216]: time="2025-09-13T00:07:22.738179988Z" level=info msg="cleaning up dead shim" Sep 13 00:07:22.745510 env[1216]: time="2025-09-13T00:07:22.745471109Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3694 runtime=io.containerd.runc.v2\n" Sep 13 00:07:22.745852 env[1216]: time="2025-09-13T00:07:22.745825629Z" level=info msg="TearDown network for sandbox \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" successfully" Sep 13 00:07:22.745905 env[1216]: time="2025-09-13T00:07:22.745852709Z" level=info msg="StopPodSandbox for \"e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d\" returns successfully" Sep 13 00:07:22.756362 kubelet[1920]: I0913 00:07:22.756309 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-net\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756362 kubelet[1920]: I0913 00:07:22.756351 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-lib-modules\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756362 kubelet[1920]: I0913 00:07:22.756370 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cni-path\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756390 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-config-path\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756407 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-hostproc\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756423 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-hubble-tls\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756438 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-etc-cni-netd\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756452 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-xtables-lock\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.756878 kubelet[1920]: I0913 00:07:22.756466 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-kernel\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756481 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-run\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756498 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl9xs\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-kube-api-access-jl9xs\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756513 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-bpf-maps\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756532 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97zsd\" (UniqueName: \"kubernetes.io/projected/3fa19538-2cb2-498e-b692-f575000fce93-kube-api-access-97zsd\") pod \"3fa19538-2cb2-498e-b692-f575000fce93\" (UID: \"3fa19538-2cb2-498e-b692-f575000fce93\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756548 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3fa19538-2cb2-498e-b692-f575000fce93-cilium-config-path\") pod \"3fa19538-2cb2-498e-b692-f575000fce93\" (UID: \"3fa19538-2cb2-498e-b692-f575000fce93\") " Sep 13 00:07:22.757038 kubelet[1920]: I0913 00:07:22.756562 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-cgroup\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.757169 kubelet[1920]: I0913 00:07:22.756582 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0498e092-a3ef-4562-8c1f-832bdfc5c680-clustermesh-secrets\") pod \"0498e092-a3ef-4562-8c1f-832bdfc5c680\" (UID: \"0498e092-a3ef-4562-8c1f-832bdfc5c680\") " Sep 13 00:07:22.760302 kubelet[1920]: I0913 00:07:22.758284 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760302 kubelet[1920]: I0913 00:07:22.758307 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760302 kubelet[1920]: I0913 00:07:22.758284 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760302 kubelet[1920]: I0913 00:07:22.758344 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-hostproc" (OuterVolumeSpecName: "hostproc") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760302 kubelet[1920]: I0913 00:07:22.758347 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760500 kubelet[1920]: I0913 00:07:22.759000 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760500 kubelet[1920]: I0913 00:07:22.760203 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:07:22.760500 kubelet[1920]: I0913 00:07:22.760252 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760670 kubelet[1920]: I0913 00:07:22.760653 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cni-path" (OuterVolumeSpecName: "cni-path") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760775 kubelet[1920]: I0913 00:07:22.760758 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.760852 kubelet[1920]: I0913 00:07:22.760839 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:22.761948 kubelet[1920]: I0913 00:07:22.761910 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fa19538-2cb2-498e-b692-f575000fce93-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3fa19538-2cb2-498e-b692-f575000fce93" (UID: "3fa19538-2cb2-498e-b692-f575000fce93"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:07:22.762119 kubelet[1920]: I0913 00:07:22.762093 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-kube-api-access-jl9xs" (OuterVolumeSpecName: "kube-api-access-jl9xs") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "kube-api-access-jl9xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:07:22.763647 kubelet[1920]: I0913 00:07:22.763614 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:07:22.763751 kubelet[1920]: I0913 00:07:22.763702 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fa19538-2cb2-498e-b692-f575000fce93-kube-api-access-97zsd" (OuterVolumeSpecName: "kube-api-access-97zsd") pod "3fa19538-2cb2-498e-b692-f575000fce93" (UID: "3fa19538-2cb2-498e-b692-f575000fce93"). InnerVolumeSpecName "kube-api-access-97zsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:07:22.764172 kubelet[1920]: I0913 00:07:22.764149 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0498e092-a3ef-4562-8c1f-832bdfc5c680-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "0498e092-a3ef-4562-8c1f-832bdfc5c680" (UID: "0498e092-a3ef-4562-8c1f-832bdfc5c680"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:07:22.856907 kubelet[1920]: I0913 00:07:22.856865 1920 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97zsd\" (UniqueName: \"kubernetes.io/projected/3fa19538-2cb2-498e-b692-f575000fce93-kube-api-access-97zsd\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.856907 kubelet[1920]: I0913 00:07:22.856902 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3fa19538-2cb2-498e-b692-f575000fce93-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.856907 kubelet[1920]: I0913 00:07:22.856912 1920 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856921 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856936 1920 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/0498e092-a3ef-4562-8c1f-832bdfc5c680-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856946 1920 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856978 1920 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856986 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.856994 1920 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.857001 1920 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857125 kubelet[1920]: I0913 00:07:22.857008 1920 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857302 kubelet[1920]: I0913 00:07:22.857015 1920 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857302 kubelet[1920]: I0913 00:07:22.857022 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857302 kubelet[1920]: I0913 00:07:22.857030 1920 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl9xs\" (UniqueName: \"kubernetes.io/projected/0498e092-a3ef-4562-8c1f-832bdfc5c680-kube-api-access-jl9xs\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857302 kubelet[1920]: I0913 00:07:22.857037 1920 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.857302 kubelet[1920]: I0913 00:07:22.857045 1920 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/0498e092-a3ef-4562-8c1f-832bdfc5c680-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:22.955383 kubelet[1920]: I0913 00:07:22.955338 1920 scope.go:117] "RemoveContainer" containerID="17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653" Sep 13 00:07:22.958280 env[1216]: time="2025-09-13T00:07:22.957859115Z" level=info msg="RemoveContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\"" Sep 13 00:07:22.960214 systemd[1]: Removed slice kubepods-besteffort-pod3fa19538_2cb2_498e_b692_f575000fce93.slice. Sep 13 00:07:22.964880 systemd[1]: Removed slice kubepods-burstable-pod0498e092_a3ef_4562_8c1f_832bdfc5c680.slice. Sep 13 00:07:22.964973 systemd[1]: kubepods-burstable-pod0498e092_a3ef_4562_8c1f_832bdfc5c680.slice: Consumed 6.602s CPU time. Sep 13 00:07:22.965282 env[1216]: time="2025-09-13T00:07:22.965238357Z" level=info msg="RemoveContainer for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" returns successfully" Sep 13 00:07:22.965613 kubelet[1920]: I0913 00:07:22.965587 1920 scope.go:117] "RemoveContainer" containerID="17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653" Sep 13 00:07:22.966387 env[1216]: time="2025-09-13T00:07:22.966317037Z" level=error msg="ContainerStatus for \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\": not found" Sep 13 00:07:22.967343 kubelet[1920]: E0913 00:07:22.967306 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\": not found" containerID="17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653" Sep 13 00:07:22.967435 kubelet[1920]: I0913 00:07:22.967347 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653"} err="failed to get container status \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\": rpc error: code = NotFound desc = an error occurred when try to find container \"17cda87f29c0ab8cd7a7e2be58e3ba6cbcf89fc4c4c3c9a2dbf0e636760c7653\": not found" Sep 13 00:07:22.967435 kubelet[1920]: I0913 00:07:22.967424 1920 scope.go:117] "RemoveContainer" containerID="c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980" Sep 13 00:07:22.968521 env[1216]: time="2025-09-13T00:07:22.968492718Z" level=info msg="RemoveContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\"" Sep 13 00:07:22.972475 env[1216]: time="2025-09-13T00:07:22.972426879Z" level=info msg="RemoveContainer for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" returns successfully" Sep 13 00:07:22.972668 kubelet[1920]: I0913 00:07:22.972615 1920 scope.go:117] "RemoveContainer" containerID="1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032" Sep 13 00:07:22.974658 env[1216]: time="2025-09-13T00:07:22.974630679Z" level=info msg="RemoveContainer for \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\"" Sep 13 00:07:22.978297 env[1216]: time="2025-09-13T00:07:22.978256920Z" level=info msg="RemoveContainer for \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\" returns successfully" Sep 13 00:07:22.978569 kubelet[1920]: I0913 00:07:22.978532 1920 scope.go:117] "RemoveContainer" containerID="f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544" Sep 13 00:07:22.980511 env[1216]: time="2025-09-13T00:07:22.980470160Z" level=info msg="RemoveContainer for \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\"" Sep 13 00:07:22.984446 env[1216]: time="2025-09-13T00:07:22.984398561Z" level=info msg="RemoveContainer for \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\" returns successfully" Sep 13 00:07:22.984651 kubelet[1920]: I0913 00:07:22.984598 1920 scope.go:117] "RemoveContainer" containerID="168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05" Sep 13 00:07:22.985751 env[1216]: time="2025-09-13T00:07:22.985704481Z" level=info msg="RemoveContainer for \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\"" Sep 13 00:07:22.989010 env[1216]: time="2025-09-13T00:07:22.988968282Z" level=info msg="RemoveContainer for \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\" returns successfully" Sep 13 00:07:22.989191 kubelet[1920]: I0913 00:07:22.989167 1920 scope.go:117] "RemoveContainer" containerID="b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240" Sep 13 00:07:22.990253 env[1216]: time="2025-09-13T00:07:22.990223682Z" level=info msg="RemoveContainer for \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\"" Sep 13 00:07:22.992660 env[1216]: time="2025-09-13T00:07:22.992617083Z" level=info msg="RemoveContainer for \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\" returns successfully" Sep 13 00:07:22.992856 kubelet[1920]: I0913 00:07:22.992835 1920 scope.go:117] "RemoveContainer" containerID="c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980" Sep 13 00:07:22.993171 env[1216]: time="2025-09-13T00:07:22.993102363Z" level=error msg="ContainerStatus for \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\": not found" Sep 13 00:07:22.993382 kubelet[1920]: E0913 00:07:22.993355 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\": not found" containerID="c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980" Sep 13 00:07:22.993422 kubelet[1920]: I0913 00:07:22.993389 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980"} err="failed to get container status \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\": rpc error: code = NotFound desc = an error occurred when try to find container \"c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980\": not found" Sep 13 00:07:22.993422 kubelet[1920]: I0913 00:07:22.993415 1920 scope.go:117] "RemoveContainer" containerID="1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032" Sep 13 00:07:22.993696 env[1216]: time="2025-09-13T00:07:22.993643643Z" level=error msg="ContainerStatus for \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\": not found" Sep 13 00:07:22.993844 kubelet[1920]: E0913 00:07:22.993819 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\": not found" containerID="1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032" Sep 13 00:07:22.993886 kubelet[1920]: I0913 00:07:22.993847 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032"} err="failed to get container status \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\": rpc error: code = NotFound desc = an error occurred when try to find container \"1f4d9954c5f11610af9c2f9fea8208b1e5b5f43ca322efc0f5a6a374963d1032\": not found" Sep 13 00:07:22.993886 kubelet[1920]: I0913 00:07:22.993863 1920 scope.go:117] "RemoveContainer" containerID="f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544" Sep 13 00:07:22.994104 env[1216]: time="2025-09-13T00:07:22.994056883Z" level=error msg="ContainerStatus for \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\": not found" Sep 13 00:07:22.994229 kubelet[1920]: E0913 00:07:22.994209 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\": not found" containerID="f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544" Sep 13 00:07:22.994265 kubelet[1920]: I0913 00:07:22.994235 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544"} err="failed to get container status \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\": rpc error: code = NotFound desc = an error occurred when try to find container \"f6238e9a042828684089349fa8c83b902497fe663c0ffdcbea6946f76b2f3544\": not found" Sep 13 00:07:22.994265 kubelet[1920]: I0913 00:07:22.994250 1920 scope.go:117] "RemoveContainer" containerID="168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05" Sep 13 00:07:22.994480 env[1216]: time="2025-09-13T00:07:22.994432803Z" level=error msg="ContainerStatus for \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\": not found" Sep 13 00:07:22.994605 kubelet[1920]: E0913 00:07:22.994583 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\": not found" containerID="168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05" Sep 13 00:07:22.994650 kubelet[1920]: I0913 00:07:22.994608 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05"} err="failed to get container status \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\": rpc error: code = NotFound desc = an error occurred when try to find container \"168c1a85870e31b663c228b58cdad66a654e13aef15b73d0e38b2168fd7ebc05\": not found" Sep 13 00:07:22.994650 kubelet[1920]: I0913 00:07:22.994623 1920 scope.go:117] "RemoveContainer" containerID="b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240" Sep 13 00:07:22.994989 env[1216]: time="2025-09-13T00:07:22.994926563Z" level=error msg="ContainerStatus for \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\": not found" Sep 13 00:07:22.995113 kubelet[1920]: E0913 00:07:22.995092 1920 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\": not found" containerID="b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240" Sep 13 00:07:22.995163 kubelet[1920]: I0913 00:07:22.995118 1920 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240"} err="failed to get container status \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\": rpc error: code = NotFound desc = an error occurred when try to find container \"b364dd2870851702f4dcc12716ce81168800e801390b991a9e382befebc5f240\": not found" Sep 13 00:07:23.579637 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c3008fa30d1117a2aeb97f5de7aed03d331ff70c24e04e541d01085a70eff980-rootfs.mount: Deactivated successfully. Sep 13 00:07:23.579759 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-50fe0337bb9927eadc4492053861ce607187d4c2d3d0acb27041d208305ad801-rootfs.mount: Deactivated successfully. Sep 13 00:07:23.579814 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d-rootfs.mount: Deactivated successfully. Sep 13 00:07:23.579869 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e043d200535a92c1705afc5edfcff92619cf945ecae96e814bd5b5df6f0bc75d-shm.mount: Deactivated successfully. Sep 13 00:07:23.579918 systemd[1]: var-lib-kubelet-pods-3fa19538\x2d2cb2\x2d498e\x2db692\x2df575000fce93-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d97zsd.mount: Deactivated successfully. Sep 13 00:07:23.579983 systemd[1]: var-lib-kubelet-pods-0498e092\x2da3ef\x2d4562\x2d8c1f\x2d832bdfc5c680-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djl9xs.mount: Deactivated successfully. Sep 13 00:07:23.580038 systemd[1]: var-lib-kubelet-pods-0498e092\x2da3ef\x2d4562\x2d8c1f\x2d832bdfc5c680-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:07:23.580086 systemd[1]: var-lib-kubelet-pods-0498e092\x2da3ef\x2d4562\x2d8c1f\x2d832bdfc5c680-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:07:23.769215 kubelet[1920]: E0913 00:07:23.769153 1920 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 00:07:24.521885 sshd[3549]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:24.526220 systemd[1]: Started sshd@22-10.0.0.20:22-10.0.0.1:53168.service. Sep 13 00:07:24.526806 systemd[1]: sshd@21-10.0.0.20:22-10.0.0.1:53156.service: Deactivated successfully. Sep 13 00:07:24.527458 systemd[1]: session-22.scope: Deactivated successfully. Sep 13 00:07:24.527611 systemd[1]: session-22.scope: Consumed 1.040s CPU time. Sep 13 00:07:24.528867 systemd-logind[1206]: Session 22 logged out. Waiting for processes to exit. Sep 13 00:07:24.529864 systemd-logind[1206]: Removed session 22. Sep 13 00:07:24.579585 sshd[3713]: Accepted publickey for core from 10.0.0.1 port 53168 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:24.580797 sshd[3713]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:24.584550 systemd-logind[1206]: New session 23 of user core. Sep 13 00:07:24.585048 systemd[1]: Started session-23.scope. Sep 13 00:07:24.738760 kubelet[1920]: I0913 00:07:24.738703 1920 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" path="/var/lib/kubelet/pods/0498e092-a3ef-4562-8c1f-832bdfc5c680/volumes" Sep 13 00:07:24.739290 kubelet[1920]: I0913 00:07:24.739262 1920 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fa19538-2cb2-498e-b692-f575000fce93" path="/var/lib/kubelet/pods/3fa19538-2cb2-498e-b692-f575000fce93/volumes" Sep 13 00:07:25.241076 sshd[3713]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:25.241666 systemd[1]: Started sshd@23-10.0.0.20:22-10.0.0.1:53182.service. Sep 13 00:07:25.244329 systemd[1]: sshd@22-10.0.0.20:22-10.0.0.1:53168.service: Deactivated successfully. Sep 13 00:07:25.245020 systemd[1]: session-23.scope: Deactivated successfully. Sep 13 00:07:25.245714 systemd-logind[1206]: Session 23 logged out. Waiting for processes to exit. Sep 13 00:07:25.246938 systemd-logind[1206]: Removed session 23. Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269617 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="mount-cgroup" Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269651 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="cilium-agent" Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269659 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="apply-sysctl-overwrites" Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269665 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="mount-bpf-fs" Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269671 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3fa19538-2cb2-498e-b692-f575000fce93" containerName="cilium-operator" Sep 13 00:07:25.269661 kubelet[1920]: E0913 00:07:25.269676 1920 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="clean-cilium-state" Sep 13 00:07:25.270125 kubelet[1920]: I0913 00:07:25.269698 1920 memory_manager.go:354] "RemoveStaleState removing state" podUID="0498e092-a3ef-4562-8c1f-832bdfc5c680" containerName="cilium-agent" Sep 13 00:07:25.270125 kubelet[1920]: I0913 00:07:25.269704 1920 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fa19538-2cb2-498e-b692-f575000fce93" containerName="cilium-operator" Sep 13 00:07:25.279497 systemd[1]: Created slice kubepods-burstable-pod09aec6fd_8019_4c6a_a1d6_35d939a1a882.slice. Sep 13 00:07:25.287081 sshd[3725]: Accepted publickey for core from 10.0.0.1 port 53182 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:25.288409 sshd[3725]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:25.293132 systemd-logind[1206]: New session 24 of user core. Sep 13 00:07:25.293936 systemd[1]: Started session-24.scope. Sep 13 00:07:25.434524 sshd[3725]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:25.436483 kubelet[1920]: E0913 00:07:25.436435 1920 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-ipsec-secrets cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-ffp44 lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-d9k8t" podUID="09aec6fd-8019-4c6a-a1d6-35d939a1a882" Sep 13 00:07:25.438827 systemd[1]: Started sshd@24-10.0.0.20:22-10.0.0.1:53184.service. Sep 13 00:07:25.445558 systemd[1]: session-24.scope: Deactivated successfully. Sep 13 00:07:25.446392 systemd-logind[1206]: Session 24 logged out. Waiting for processes to exit. Sep 13 00:07:25.446590 systemd[1]: sshd@23-10.0.0.20:22-10.0.0.1:53182.service: Deactivated successfully. Sep 13 00:07:25.448004 systemd-logind[1206]: Removed session 24. Sep 13 00:07:25.470978 kubelet[1920]: I0913 00:07:25.470927 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-cgroup\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.470978 kubelet[1920]: I0913 00:07:25.470980 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-xtables-lock\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471167 kubelet[1920]: I0913 00:07:25.471103 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-clustermesh-secrets\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471167 kubelet[1920]: I0913 00:07:25.471127 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-config-path\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471221 kubelet[1920]: I0913 00:07:25.471182 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-bpf-maps\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471221 kubelet[1920]: I0913 00:07:25.471206 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hostproc\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471281 kubelet[1920]: I0913 00:07:25.471262 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hubble-tls\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471324 kubelet[1920]: I0913 00:07:25.471297 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffp44\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-kube-api-access-ffp44\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471359 kubelet[1920]: I0913 00:07:25.471330 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-lib-modules\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471359 kubelet[1920]: I0913 00:07:25.471347 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-run\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471405 kubelet[1920]: I0913 00:07:25.471391 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cni-path\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471450 kubelet[1920]: I0913 00:07:25.471410 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-etc-cni-netd\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471489 kubelet[1920]: I0913 00:07:25.471469 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-ipsec-secrets\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471523 kubelet[1920]: I0913 00:07:25.471491 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-net\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.471523 kubelet[1920]: I0913 00:07:25.471507 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-kernel\") pod \"cilium-d9k8t\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " pod="kube-system/cilium-d9k8t" Sep 13 00:07:25.480233 sshd[3739]: Accepted publickey for core from 10.0.0.1 port 53184 ssh2: RSA SHA256:IYYmYtZT7fhBES8dcJq//ghMZv88JUKT/A8TkXgi+lY Sep 13 00:07:25.481594 sshd[3739]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:07:25.485776 systemd-logind[1206]: New session 25 of user core. Sep 13 00:07:25.486389 systemd[1]: Started session-25.scope. Sep 13 00:07:25.974782 kubelet[1920]: I0913 00:07:25.974752 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cni-path\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.974910 kubelet[1920]: I0913 00:07:25.974838 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hubble-tls\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.974910 kubelet[1920]: I0913 00:07:25.974858 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-etc-cni-netd\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.974910 kubelet[1920]: I0913 00:07:25.974872 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-net\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.974910 kubelet[1920]: I0913 00:07:25.974891 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hostproc\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.974917 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-xtables-lock\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.974937 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-ipsec-secrets\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.974955 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-config-path\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.974971 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-clustermesh-secrets\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.974986 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-run\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975039 kubelet[1920]: I0913 00:07:25.975000 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-kernel\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975169 kubelet[1920]: I0913 00:07:25.975013 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-lib-modules\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975169 kubelet[1920]: I0913 00:07:25.975059 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-cgroup\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975169 kubelet[1920]: I0913 00:07:25.975077 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-bpf-maps\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975169 kubelet[1920]: I0913 00:07:25.975093 1920 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffp44\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-kube-api-access-ffp44\") pod \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\" (UID: \"09aec6fd-8019-4c6a-a1d6-35d939a1a882\") " Sep 13 00:07:25.975528 kubelet[1920]: I0913 00:07:25.975494 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.975574 kubelet[1920]: I0913 00:07:25.975535 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cni-path" (OuterVolumeSpecName: "cni-path") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.976437 kubelet[1920]: I0913 00:07:25.976396 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hostproc" (OuterVolumeSpecName: "hostproc") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.976437 kubelet[1920]: I0913 00:07:25.976436 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.976559 kubelet[1920]: I0913 00:07:25.976454 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.979813 kubelet[1920]: I0913 00:07:25.978000 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:07:25.979813 kubelet[1920]: I0913 00:07:25.978051 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.979813 kubelet[1920]: I0913 00:07:25.978073 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.979813 kubelet[1920]: I0913 00:07:25.978079 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:07:25.979813 kubelet[1920]: I0913 00:07:25.978088 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.978990 systemd[1]: var-lib-kubelet-pods-09aec6fd\x2d8019\x2d4c6a\x2da1d6\x2d35d939a1a882-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Sep 13 00:07:25.980183 kubelet[1920]: I0913 00:07:25.978126 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.980183 kubelet[1920]: I0913 00:07:25.978148 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:07:25.980183 kubelet[1920]: I0913 00:07:25.978221 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:07:25.979078 systemd[1]: var-lib-kubelet-pods-09aec6fd\x2d8019\x2d4c6a\x2da1d6\x2d35d939a1a882-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:07:25.980304 kubelet[1920]: I0913 00:07:25.980272 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-kube-api-access-ffp44" (OuterVolumeSpecName: "kube-api-access-ffp44") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "kube-api-access-ffp44". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:07:25.981603 kubelet[1920]: I0913 00:07:25.981571 1920 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "09aec6fd-8019-4c6a-a1d6-35d939a1a882" (UID: "09aec6fd-8019-4c6a-a1d6-35d939a1a882"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.075928 1920 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-lib-modules\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.075962 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.075972 1920 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-bpf-maps\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.075982 1920 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffp44\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-kube-api-access-ffp44\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.075997 1920 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cni-path\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.076005 1920 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.076014 1920 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076001 kubelet[1920]: I0913 00:07:26.076021 1920 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hubble-tls\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076031 1920 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-hostproc\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076038 1920 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-xtables-lock\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076047 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-ipsec-secrets\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076055 1920 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/09aec6fd-8019-4c6a-a1d6-35d939a1a882-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076063 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-run\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076072 1920 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/09aec6fd-8019-4c6a-a1d6-35d939a1a882-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.076314 kubelet[1920]: I0913 00:07:26.076080 1920 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/09aec6fd-8019-4c6a-a1d6-35d939a1a882-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Sep 13 00:07:26.576688 systemd[1]: var-lib-kubelet-pods-09aec6fd\x2d8019\x2d4c6a\x2da1d6\x2d35d939a1a882-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dffp44.mount: Deactivated successfully. Sep 13 00:07:26.576808 systemd[1]: var-lib-kubelet-pods-09aec6fd\x2d8019\x2d4c6a\x2da1d6\x2d35d939a1a882-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:07:26.741837 systemd[1]: Removed slice kubepods-burstable-pod09aec6fd_8019_4c6a_a1d6_35d939a1a882.slice. Sep 13 00:07:27.016605 systemd[1]: Created slice kubepods-burstable-pod7e9bef04_518d_419d_97ec_7f5e94328236.slice. Sep 13 00:07:27.081347 kubelet[1920]: I0913 00:07:27.081293 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-cilium-run\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081377 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-host-proc-sys-kernel\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081403 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgt5c\" (UniqueName: \"kubernetes.io/projected/7e9bef04-518d-419d-97ec-7f5e94328236-kube-api-access-pgt5c\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081423 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-cilium-cgroup\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081440 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7e9bef04-518d-419d-97ec-7f5e94328236-clustermesh-secrets\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081457 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-hostproc\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081703 kubelet[1920]: I0913 00:07:27.081471 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-cni-path\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081488 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-etc-cni-netd\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081509 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-bpf-maps\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081524 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-xtables-lock\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081568 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7e9bef04-518d-419d-97ec-7f5e94328236-cilium-config-path\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081601 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/7e9bef04-518d-419d-97ec-7f5e94328236-cilium-ipsec-secrets\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.081890 kubelet[1920]: I0913 00:07:27.081621 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-host-proc-sys-net\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.082047 kubelet[1920]: I0913 00:07:27.081654 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7e9bef04-518d-419d-97ec-7f5e94328236-hubble-tls\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.082047 kubelet[1920]: I0913 00:07:27.081676 1920 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7e9bef04-518d-419d-97ec-7f5e94328236-lib-modules\") pod \"cilium-lbfnq\" (UID: \"7e9bef04-518d-419d-97ec-7f5e94328236\") " pod="kube-system/cilium-lbfnq" Sep 13 00:07:27.320388 kubelet[1920]: E0913 00:07:27.320262 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:27.321772 env[1216]: time="2025-09-13T00:07:27.321736475Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lbfnq,Uid:7e9bef04-518d-419d-97ec-7f5e94328236,Namespace:kube-system,Attempt:0,}" Sep 13 00:07:27.334251 env[1216]: time="2025-09-13T00:07:27.333886556Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:07:27.334251 env[1216]: time="2025-09-13T00:07:27.333940396Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:07:27.334251 env[1216]: time="2025-09-13T00:07:27.333951076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:07:27.334251 env[1216]: time="2025-09-13T00:07:27.334102277Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac pid=3770 runtime=io.containerd.runc.v2 Sep 13 00:07:27.349021 systemd[1]: Started cri-containerd-b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac.scope. Sep 13 00:07:27.414528 env[1216]: time="2025-09-13T00:07:27.414485345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lbfnq,Uid:7e9bef04-518d-419d-97ec-7f5e94328236,Namespace:kube-system,Attempt:0,} returns sandbox id \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\"" Sep 13 00:07:27.415761 kubelet[1920]: E0913 00:07:27.415360 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:27.417355 env[1216]: time="2025-09-13T00:07:27.417317395Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:07:27.428405 env[1216]: time="2025-09-13T00:07:27.428352352Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2\"" Sep 13 00:07:27.430438 env[1216]: time="2025-09-13T00:07:27.429046354Z" level=info msg="StartContainer for \"9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2\"" Sep 13 00:07:27.444765 systemd[1]: Started cri-containerd-9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2.scope. Sep 13 00:07:27.474510 env[1216]: time="2025-09-13T00:07:27.474458626Z" level=info msg="StartContainer for \"9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2\" returns successfully" Sep 13 00:07:27.483577 systemd[1]: cri-containerd-9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2.scope: Deactivated successfully. Sep 13 00:07:27.516866 env[1216]: time="2025-09-13T00:07:27.516820488Z" level=info msg="shim disconnected" id=9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2 Sep 13 00:07:27.517131 env[1216]: time="2025-09-13T00:07:27.517110728Z" level=warning msg="cleaning up after shim disconnected" id=9ec4ee27a835325d4f4ff499f5834922bde39183f1973cd23a612b8d54fda5e2 namespace=k8s.io Sep 13 00:07:27.517209 env[1216]: time="2025-09-13T00:07:27.517195569Z" level=info msg="cleaning up dead shim" Sep 13 00:07:27.524678 env[1216]: time="2025-09-13T00:07:27.524637514Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:27Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3856 runtime=io.containerd.runc.v2\n" Sep 13 00:07:27.736713 kubelet[1920]: E0913 00:07:27.736664 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:27.972738 kubelet[1920]: E0913 00:07:27.972677 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:27.976327 env[1216]: time="2025-09-13T00:07:27.976287423Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:07:27.990141 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount841957965.mount: Deactivated successfully. Sep 13 00:07:27.992673 env[1216]: time="2025-09-13T00:07:27.992596078Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f\"" Sep 13 00:07:27.993759 env[1216]: time="2025-09-13T00:07:27.993691322Z" level=info msg="StartContainer for \"04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f\"" Sep 13 00:07:28.013459 systemd[1]: Started cri-containerd-04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f.scope. Sep 13 00:07:28.045806 env[1216]: time="2025-09-13T00:07:28.045407120Z" level=info msg="StartContainer for \"04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f\" returns successfully" Sep 13 00:07:28.049781 systemd[1]: cri-containerd-04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f.scope: Deactivated successfully. Sep 13 00:07:28.084209 env[1216]: time="2025-09-13T00:07:28.084160031Z" level=info msg="shim disconnected" id=04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f Sep 13 00:07:28.084448 env[1216]: time="2025-09-13T00:07:28.084429112Z" level=warning msg="cleaning up after shim disconnected" id=04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f namespace=k8s.io Sep 13 00:07:28.084513 env[1216]: time="2025-09-13T00:07:28.084500513Z" level=info msg="cleaning up dead shim" Sep 13 00:07:28.092664 env[1216]: time="2025-09-13T00:07:28.092617144Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3920 runtime=io.containerd.runc.v2\n" Sep 13 00:07:28.576921 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-04f0b12df2e8b221d812f04d0a7b97f2eaa9ae95791c9a421da55f37e57fb20f-rootfs.mount: Deactivated successfully. Sep 13 00:07:28.738286 kubelet[1920]: I0913 00:07:28.738237 1920 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09aec6fd-8019-4c6a-a1d6-35d939a1a882" path="/var/lib/kubelet/pods/09aec6fd-8019-4c6a-a1d6-35d939a1a882/volumes" Sep 13 00:07:28.769797 kubelet[1920]: E0913 00:07:28.769765 1920 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 00:07:28.976623 kubelet[1920]: E0913 00:07:28.976534 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:28.980166 env[1216]: time="2025-09-13T00:07:28.980118895Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:07:29.005477 env[1216]: time="2025-09-13T00:07:29.005428956Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9\"" Sep 13 00:07:29.006310 env[1216]: time="2025-09-13T00:07:29.006170760Z" level=info msg="StartContainer for \"87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9\"" Sep 13 00:07:29.023476 systemd[1]: Started cri-containerd-87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9.scope. Sep 13 00:07:29.056629 env[1216]: time="2025-09-13T00:07:29.056585065Z" level=info msg="StartContainer for \"87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9\" returns successfully" Sep 13 00:07:29.059681 systemd[1]: cri-containerd-87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9.scope: Deactivated successfully. Sep 13 00:07:29.081369 env[1216]: time="2025-09-13T00:07:29.081304415Z" level=info msg="shim disconnected" id=87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9 Sep 13 00:07:29.081602 env[1216]: time="2025-09-13T00:07:29.081581096Z" level=warning msg="cleaning up after shim disconnected" id=87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9 namespace=k8s.io Sep 13 00:07:29.081666 env[1216]: time="2025-09-13T00:07:29.081652016Z" level=info msg="cleaning up dead shim" Sep 13 00:07:29.088419 env[1216]: time="2025-09-13T00:07:29.088377806Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:29Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3978 runtime=io.containerd.runc.v2\n" Sep 13 00:07:29.576928 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-87b6e7277409dea5e4b624234cab3e707aec98ef56189afe0d6c288b5800bee9-rootfs.mount: Deactivated successfully. Sep 13 00:07:29.736416 kubelet[1920]: E0913 00:07:29.736145 1920 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-btkcw" podUID="258c8c50-882e-4189-b148-a427542985f0" Sep 13 00:07:29.979697 kubelet[1920]: E0913 00:07:29.979664 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:29.981907 env[1216]: time="2025-09-13T00:07:29.981858152Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:07:29.993573 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1073968492.mount: Deactivated successfully. Sep 13 00:07:29.998125 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2227480091.mount: Deactivated successfully. Sep 13 00:07:30.002186 env[1216]: time="2025-09-13T00:07:30.002129203Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91\"" Sep 13 00:07:30.002737 env[1216]: time="2025-09-13T00:07:30.002703166Z" level=info msg="StartContainer for \"6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91\"" Sep 13 00:07:30.015452 kubelet[1920]: I0913 00:07:30.014626 1920 setters.go:600] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-13T00:07:30Z","lastTransitionTime":"2025-09-13T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 13 00:07:30.016365 systemd[1]: Started cri-containerd-6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91.scope. Sep 13 00:07:30.050122 env[1216]: time="2025-09-13T00:07:30.050077802Z" level=info msg="StartContainer for \"6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91\" returns successfully" Sep 13 00:07:30.050809 systemd[1]: cri-containerd-6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91.scope: Deactivated successfully. Sep 13 00:07:30.074049 env[1216]: time="2025-09-13T00:07:30.073994842Z" level=info msg="shim disconnected" id=6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91 Sep 13 00:07:30.074049 env[1216]: time="2025-09-13T00:07:30.074040642Z" level=warning msg="cleaning up after shim disconnected" id=6dbf1178ce5595349210593a08178b429c176cc70a4ed81745506df7da43ee91 namespace=k8s.io Sep 13 00:07:30.074049 env[1216]: time="2025-09-13T00:07:30.074051122Z" level=info msg="cleaning up dead shim" Sep 13 00:07:30.081392 env[1216]: time="2025-09-13T00:07:30.081352118Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:07:30Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4034 runtime=io.containerd.runc.v2\n" Sep 13 00:07:30.984180 kubelet[1920]: E0913 00:07:30.984136 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:30.987640 env[1216]: time="2025-09-13T00:07:30.987578963Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:07:30.998915 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount475858334.mount: Deactivated successfully. Sep 13 00:07:31.005596 env[1216]: time="2025-09-13T00:07:31.005550496Z" level=info msg="CreateContainer within sandbox \"b8a5b5b7c1d9738d753c41388cfdaf7d6f83dc62c26b56ae74878f6cc0f110ac\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c\"" Sep 13 00:07:31.013741 env[1216]: time="2025-09-13T00:07:31.013681580Z" level=info msg="StartContainer for \"b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c\"" Sep 13 00:07:31.034139 systemd[1]: Started cri-containerd-b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c.scope. Sep 13 00:07:31.077731 env[1216]: time="2025-09-13T00:07:31.077420372Z" level=info msg="StartContainer for \"b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c\" returns successfully" Sep 13 00:07:31.337762 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm-aes-ce))) Sep 13 00:07:31.736317 kubelet[1920]: E0913 00:07:31.736254 1920 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-btkcw" podUID="258c8c50-882e-4189-b148-a427542985f0" Sep 13 00:07:31.989878 kubelet[1920]: E0913 00:07:31.989770 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:32.006911 kubelet[1920]: I0913 00:07:32.006827 1920 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-lbfnq" podStartSLOduration=6.006811575 podStartE2EDuration="6.006811575s" podCreationTimestamp="2025-09-13 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:07:32.005371846 +0000 UTC m=+93.389389518" watchObservedRunningTime="2025-09-13 00:07:32.006811575 +0000 UTC m=+93.390829247" Sep 13 00:07:33.321951 kubelet[1920]: E0913 00:07:33.321874 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:33.736402 kubelet[1920]: E0913 00:07:33.736354 1920 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-btkcw" podUID="258c8c50-882e-4189-b148-a427542985f0" Sep 13 00:07:34.139900 systemd-networkd[1044]: lxc_health: Link UP Sep 13 00:07:34.153753 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:07:34.154166 systemd-networkd[1044]: lxc_health: Gained carrier Sep 13 00:07:35.322640 kubelet[1920]: E0913 00:07:35.322564 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:35.332898 systemd-networkd[1044]: lxc_health: Gained IPv6LL Sep 13 00:07:35.736295 kubelet[1920]: E0913 00:07:35.736252 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:35.865427 systemd[1]: run-containerd-runc-k8s.io-b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c-runc.uJOhxV.mount: Deactivated successfully. Sep 13 00:07:35.912001 kubelet[1920]: E0913 00:07:35.911963 1920 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:60020->127.0.0.1:40345: write tcp 127.0.0.1:60020->127.0.0.1:40345: write: broken pipe Sep 13 00:07:35.998175 kubelet[1920]: E0913 00:07:35.998068 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:37.005794 kubelet[1920]: E0913 00:07:37.005763 1920 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 13 00:07:37.988969 systemd[1]: run-containerd-runc-k8s.io-b967399db45ac17fdc51202accd5636a7ca65291d3870b5e2287b8cea9b10d8c-runc.bYR3EN.mount: Deactivated successfully. Sep 13 00:07:40.240587 sshd[3739]: pam_unix(sshd:session): session closed for user core Sep 13 00:07:40.243370 systemd[1]: sshd@24-10.0.0.20:22-10.0.0.1:53184.service: Deactivated successfully. Sep 13 00:07:40.244076 systemd[1]: session-25.scope: Deactivated successfully. Sep 13 00:07:40.244566 systemd-logind[1206]: Session 25 logged out. Waiting for processes to exit. Sep 13 00:07:40.245279 systemd-logind[1206]: Removed session 25.