Feb 13 19:05:24.936346 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Feb 13 19:05:24.936366 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Thu Feb 13 17:39:57 -00 2025 Feb 13 19:05:24.936376 kernel: KASLR enabled Feb 13 19:05:24.936381 kernel: efi: EFI v2.7 by EDK II Feb 13 19:05:24.936387 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdbbae018 ACPI 2.0=0xd9b43018 RNG=0xd9b43a18 MEMRESERVE=0xd9b40218 Feb 13 19:05:24.936392 kernel: random: crng init done Feb 13 19:05:24.936399 kernel: secureboot: Secure boot disabled Feb 13 19:05:24.936405 kernel: ACPI: Early table checksum verification disabled Feb 13 19:05:24.936411 kernel: ACPI: RSDP 0x00000000D9B43018 000024 (v02 BOCHS ) Feb 13 19:05:24.936418 kernel: ACPI: XSDT 0x00000000D9B43F18 000064 (v01 BOCHS BXPC 00000001 01000013) Feb 13 19:05:24.936424 kernel: ACPI: FACP 0x00000000D9B43B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936430 kernel: ACPI: DSDT 0x00000000D9B41018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936435 kernel: ACPI: APIC 0x00000000D9B43C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936441 kernel: ACPI: PPTT 0x00000000D9B43098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936448 kernel: ACPI: GTDT 0x00000000D9B43818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936456 kernel: ACPI: MCFG 0x00000000D9B43A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936462 kernel: ACPI: SPCR 0x00000000D9B43918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936468 kernel: ACPI: DBG2 0x00000000D9B43998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936474 kernel: ACPI: IORT 0x00000000D9B43198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:05:24.936480 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Feb 13 19:05:24.936487 kernel: NUMA: Failed to initialise from firmware Feb 13 19:05:24.936493 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Feb 13 19:05:24.936499 kernel: NUMA: NODE_DATA [mem 0xdc958800-0xdc95dfff] Feb 13 19:05:24.936505 kernel: Zone ranges: Feb 13 19:05:24.936511 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Feb 13 19:05:24.936519 kernel: DMA32 empty Feb 13 19:05:24.936525 kernel: Normal empty Feb 13 19:05:24.936531 kernel: Movable zone start for each node Feb 13 19:05:24.936537 kernel: Early memory node ranges Feb 13 19:05:24.936543 kernel: node 0: [mem 0x0000000040000000-0x00000000d967ffff] Feb 13 19:05:24.936549 kernel: node 0: [mem 0x00000000d9680000-0x00000000d968ffff] Feb 13 19:05:24.936579 kernel: node 0: [mem 0x00000000d9690000-0x00000000d976ffff] Feb 13 19:05:24.936586 kernel: node 0: [mem 0x00000000d9770000-0x00000000d9b3ffff] Feb 13 19:05:24.936592 kernel: node 0: [mem 0x00000000d9b40000-0x00000000dce1ffff] Feb 13 19:05:24.936598 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Feb 13 19:05:24.936604 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Feb 13 19:05:24.936611 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Feb 13 19:05:24.936619 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Feb 13 19:05:24.936625 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Feb 13 19:05:24.936631 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Feb 13 19:05:24.936640 kernel: psci: probing for conduit method from ACPI. Feb 13 19:05:24.936647 kernel: psci: PSCIv1.1 detected in firmware. Feb 13 19:05:24.936654 kernel: psci: Using standard PSCI v0.2 function IDs Feb 13 19:05:24.936661 kernel: psci: Trusted OS migration not required Feb 13 19:05:24.936668 kernel: psci: SMC Calling Convention v1.1 Feb 13 19:05:24.936675 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Feb 13 19:05:24.936681 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Feb 13 19:05:24.936688 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Feb 13 19:05:24.936694 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Feb 13 19:05:24.936701 kernel: Detected PIPT I-cache on CPU0 Feb 13 19:05:24.936707 kernel: CPU features: detected: GIC system register CPU interface Feb 13 19:05:24.936714 kernel: CPU features: detected: Hardware dirty bit management Feb 13 19:05:24.936720 kernel: CPU features: detected: Spectre-v4 Feb 13 19:05:24.936728 kernel: CPU features: detected: Spectre-BHB Feb 13 19:05:24.936734 kernel: CPU features: kernel page table isolation forced ON by KASLR Feb 13 19:05:24.936741 kernel: CPU features: detected: Kernel page table isolation (KPTI) Feb 13 19:05:24.936747 kernel: CPU features: detected: ARM erratum 1418040 Feb 13 19:05:24.936754 kernel: CPU features: detected: SSBS not fully self-synchronizing Feb 13 19:05:24.936760 kernel: alternatives: applying boot alternatives Feb 13 19:05:24.936768 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:05:24.936775 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:05:24.936781 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:05:24.936788 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:05:24.936795 kernel: Fallback order for Node 0: 0 Feb 13 19:05:24.936802 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Feb 13 19:05:24.936809 kernel: Policy zone: DMA Feb 13 19:05:24.936815 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:05:24.936822 kernel: software IO TLB: area num 4. Feb 13 19:05:24.936828 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB) Feb 13 19:05:24.936835 kernel: Memory: 2387540K/2572288K available (10304K kernel code, 2186K rwdata, 8092K rodata, 38336K init, 897K bss, 184748K reserved, 0K cma-reserved) Feb 13 19:05:24.936842 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Feb 13 19:05:24.936848 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:05:24.936856 kernel: rcu: RCU event tracing is enabled. Feb 13 19:05:24.936862 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Feb 13 19:05:24.936869 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:05:24.936875 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:05:24.936883 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:05:24.936890 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Feb 13 19:05:24.936896 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Feb 13 19:05:24.936903 kernel: GICv3: 256 SPIs implemented Feb 13 19:05:24.936909 kernel: GICv3: 0 Extended SPIs implemented Feb 13 19:05:24.936916 kernel: Root IRQ handler: gic_handle_irq Feb 13 19:05:24.936922 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Feb 13 19:05:24.936928 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Feb 13 19:05:24.936935 kernel: ITS [mem 0x08080000-0x0809ffff] Feb 13 19:05:24.936942 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400c0000 (indirect, esz 8, psz 64K, shr 1) Feb 13 19:05:24.936949 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400d0000 (flat, esz 8, psz 64K, shr 1) Feb 13 19:05:24.936956 kernel: GICv3: using LPI property table @0x00000000400f0000 Feb 13 19:05:24.936963 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000 Feb 13 19:05:24.936970 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:05:24.936976 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:05:24.936983 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Feb 13 19:05:24.936989 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Feb 13 19:05:24.936996 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Feb 13 19:05:24.937002 kernel: arm-pv: using stolen time PV Feb 13 19:05:24.937009 kernel: Console: colour dummy device 80x25 Feb 13 19:05:24.937016 kernel: ACPI: Core revision 20230628 Feb 13 19:05:24.937023 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Feb 13 19:05:24.937031 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:05:24.937038 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:05:24.937045 kernel: landlock: Up and running. Feb 13 19:05:24.937051 kernel: SELinux: Initializing. Feb 13 19:05:24.937058 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:05:24.937065 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:05:24.937071 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:05:24.937078 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:05:24.937085 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:05:24.937093 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:05:24.937100 kernel: Platform MSI: ITS@0x8080000 domain created Feb 13 19:05:24.937107 kernel: PCI/MSI: ITS@0x8080000 domain created Feb 13 19:05:24.937113 kernel: Remapping and enabling EFI services. Feb 13 19:05:24.937120 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:05:24.937127 kernel: Detected PIPT I-cache on CPU1 Feb 13 19:05:24.937133 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Feb 13 19:05:24.937140 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000 Feb 13 19:05:24.937147 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:05:24.937155 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Feb 13 19:05:24.937162 kernel: Detected PIPT I-cache on CPU2 Feb 13 19:05:24.937173 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Feb 13 19:05:24.937182 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000 Feb 13 19:05:24.937189 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:05:24.937196 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Feb 13 19:05:24.937203 kernel: Detected PIPT I-cache on CPU3 Feb 13 19:05:24.937210 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Feb 13 19:05:24.937217 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000 Feb 13 19:05:24.937225 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:05:24.937232 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Feb 13 19:05:24.937239 kernel: smp: Brought up 1 node, 4 CPUs Feb 13 19:05:24.937246 kernel: SMP: Total of 4 processors activated. Feb 13 19:05:24.937253 kernel: CPU features: detected: 32-bit EL0 Support Feb 13 19:05:24.937260 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Feb 13 19:05:24.937267 kernel: CPU features: detected: Common not Private translations Feb 13 19:05:24.937274 kernel: CPU features: detected: CRC32 instructions Feb 13 19:05:24.937283 kernel: CPU features: detected: Enhanced Virtualization Traps Feb 13 19:05:24.937290 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Feb 13 19:05:24.937297 kernel: CPU features: detected: LSE atomic instructions Feb 13 19:05:24.937304 kernel: CPU features: detected: Privileged Access Never Feb 13 19:05:24.937311 kernel: CPU features: detected: RAS Extension Support Feb 13 19:05:24.937318 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Feb 13 19:05:24.937324 kernel: CPU: All CPU(s) started at EL1 Feb 13 19:05:24.937332 kernel: alternatives: applying system-wide alternatives Feb 13 19:05:24.937339 kernel: devtmpfs: initialized Feb 13 19:05:24.937346 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:05:24.937354 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Feb 13 19:05:24.937361 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:05:24.937368 kernel: SMBIOS 3.0.0 present. Feb 13 19:05:24.937375 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Feb 13 19:05:24.937382 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:05:24.937389 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Feb 13 19:05:24.937397 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Feb 13 19:05:24.937404 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Feb 13 19:05:24.937411 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:05:24.937419 kernel: audit: type=2000 audit(0.025:1): state=initialized audit_enabled=0 res=1 Feb 13 19:05:24.937426 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:05:24.937434 kernel: cpuidle: using governor menu Feb 13 19:05:24.937441 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Feb 13 19:05:24.937448 kernel: ASID allocator initialised with 32768 entries Feb 13 19:05:24.937455 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:05:24.937462 kernel: Serial: AMBA PL011 UART driver Feb 13 19:05:24.937469 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Feb 13 19:05:24.937476 kernel: Modules: 0 pages in range for non-PLT usage Feb 13 19:05:24.937484 kernel: Modules: 509280 pages in range for PLT usage Feb 13 19:05:24.937491 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:05:24.937498 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:05:24.937505 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Feb 13 19:05:24.937512 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Feb 13 19:05:24.937519 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:05:24.937526 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:05:24.937533 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Feb 13 19:05:24.937540 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Feb 13 19:05:24.937549 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:05:24.937568 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:05:24.937576 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:05:24.937583 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:05:24.937590 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:05:24.937597 kernel: ACPI: Interpreter enabled Feb 13 19:05:24.937604 kernel: ACPI: Using GIC for interrupt routing Feb 13 19:05:24.937611 kernel: ACPI: MCFG table detected, 1 entries Feb 13 19:05:24.937618 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Feb 13 19:05:24.937627 kernel: printk: console [ttyAMA0] enabled Feb 13 19:05:24.937635 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 19:05:24.937771 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:05:24.937848 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Feb 13 19:05:24.937916 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Feb 13 19:05:24.937982 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Feb 13 19:05:24.938048 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Feb 13 19:05:24.938059 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Feb 13 19:05:24.938067 kernel: PCI host bridge to bus 0000:00 Feb 13 19:05:24.938139 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Feb 13 19:05:24.938200 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Feb 13 19:05:24.938261 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Feb 13 19:05:24.938321 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 19:05:24.938402 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Feb 13 19:05:24.938486 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Feb 13 19:05:24.938633 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Feb 13 19:05:24.938713 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Feb 13 19:05:24.938782 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Feb 13 19:05:24.938850 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Feb 13 19:05:24.938920 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Feb 13 19:05:24.938999 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Feb 13 19:05:24.939068 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Feb 13 19:05:24.939127 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Feb 13 19:05:24.939186 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Feb 13 19:05:24.939195 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Feb 13 19:05:24.939202 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Feb 13 19:05:24.939209 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Feb 13 19:05:24.939216 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Feb 13 19:05:24.939226 kernel: iommu: Default domain type: Translated Feb 13 19:05:24.939233 kernel: iommu: DMA domain TLB invalidation policy: strict mode Feb 13 19:05:24.939240 kernel: efivars: Registered efivars operations Feb 13 19:05:24.939248 kernel: vgaarb: loaded Feb 13 19:05:24.939255 kernel: clocksource: Switched to clocksource arch_sys_counter Feb 13 19:05:24.939262 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:05:24.939269 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:05:24.939276 kernel: pnp: PnP ACPI init Feb 13 19:05:24.939354 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Feb 13 19:05:24.939366 kernel: pnp: PnP ACPI: found 1 devices Feb 13 19:05:24.939374 kernel: NET: Registered PF_INET protocol family Feb 13 19:05:24.939381 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:05:24.939388 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:05:24.939395 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:05:24.939403 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:05:24.939410 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:05:24.939417 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:05:24.939425 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:05:24.939433 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:05:24.939440 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:05:24.939447 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:05:24.939454 kernel: kvm [1]: HYP mode not available Feb 13 19:05:24.939462 kernel: Initialise system trusted keyrings Feb 13 19:05:24.939469 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:05:24.939476 kernel: Key type asymmetric registered Feb 13 19:05:24.939483 kernel: Asymmetric key parser 'x509' registered Feb 13 19:05:24.939490 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Feb 13 19:05:24.939498 kernel: io scheduler mq-deadline registered Feb 13 19:05:24.939505 kernel: io scheduler kyber registered Feb 13 19:05:24.939512 kernel: io scheduler bfq registered Feb 13 19:05:24.939520 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Feb 13 19:05:24.939527 kernel: ACPI: button: Power Button [PWRB] Feb 13 19:05:24.939534 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Feb 13 19:05:24.939623 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Feb 13 19:05:24.939634 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:05:24.939641 kernel: thunder_xcv, ver 1.0 Feb 13 19:05:24.939651 kernel: thunder_bgx, ver 1.0 Feb 13 19:05:24.939658 kernel: nicpf, ver 1.0 Feb 13 19:05:24.939665 kernel: nicvf, ver 1.0 Feb 13 19:05:24.939740 kernel: rtc-efi rtc-efi.0: registered as rtc0 Feb 13 19:05:24.939806 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-02-13T19:05:24 UTC (1739473524) Feb 13 19:05:24.939815 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 19:05:24.939822 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Feb 13 19:05:24.939829 kernel: watchdog: Delayed init of the lockup detector failed: -19 Feb 13 19:05:24.939838 kernel: watchdog: Hard watchdog permanently disabled Feb 13 19:05:24.939845 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:05:24.939853 kernel: Segment Routing with IPv6 Feb 13 19:05:24.939860 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:05:24.939867 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:05:24.939874 kernel: Key type dns_resolver registered Feb 13 19:05:24.939881 kernel: registered taskstats version 1 Feb 13 19:05:24.939888 kernel: Loading compiled-in X.509 certificates Feb 13 19:05:24.939896 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 58bec1a0c6b8a133d1af4ea745973da0351f7027' Feb 13 19:05:24.939910 kernel: Key type .fscrypt registered Feb 13 19:05:24.939917 kernel: Key type fscrypt-provisioning registered Feb 13 19:05:24.939925 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:05:24.939932 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:05:24.939941 kernel: ima: No architecture policies found Feb 13 19:05:24.939950 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Feb 13 19:05:24.939960 kernel: clk: Disabling unused clocks Feb 13 19:05:24.939967 kernel: Freeing unused kernel memory: 38336K Feb 13 19:05:24.939974 kernel: Run /init as init process Feb 13 19:05:24.939983 kernel: with arguments: Feb 13 19:05:24.939990 kernel: /init Feb 13 19:05:24.939997 kernel: with environment: Feb 13 19:05:24.940004 kernel: HOME=/ Feb 13 19:05:24.940013 kernel: TERM=linux Feb 13 19:05:24.940020 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:05:24.940028 systemd[1]: Successfully made /usr/ read-only. Feb 13 19:05:24.940039 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:05:24.940060 systemd[1]: Detected virtualization kvm. Feb 13 19:05:24.940071 systemd[1]: Detected architecture arm64. Feb 13 19:05:24.940083 systemd[1]: Running in initrd. Feb 13 19:05:24.940090 systemd[1]: No hostname configured, using default hostname. Feb 13 19:05:24.940098 systemd[1]: Hostname set to . Feb 13 19:05:24.940105 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:05:24.940113 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:05:24.940121 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:05:24.940130 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:05:24.940138 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:05:24.940146 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:05:24.940153 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:05:24.940162 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:05:24.940170 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:05:24.940179 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:05:24.940187 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:05:24.940195 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:05:24.940202 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:05:24.940210 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:05:24.940218 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:05:24.940225 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:05:24.940233 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:05:24.940240 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:05:24.940250 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:05:24.940257 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Feb 13 19:05:24.940265 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:05:24.940272 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:05:24.940280 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:05:24.940288 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:05:24.940295 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:05:24.940303 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:05:24.940312 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:05:24.940319 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:05:24.940327 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:05:24.940335 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:05:24.940342 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:05:24.940350 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:05:24.940358 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:05:24.940367 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:05:24.940375 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:05:24.940401 systemd-journald[237]: Collecting audit messages is disabled. Feb 13 19:05:24.940421 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:05:24.940429 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:05:24.940437 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:05:24.940445 systemd-journald[237]: Journal started Feb 13 19:05:24.940463 systemd-journald[237]: Runtime Journal (/run/log/journal/88dace6cfa6445e8a6cc25991f14bbc2) is 5.9M, max 47.3M, 41.4M free. Feb 13 19:05:24.925206 systemd-modules-load[239]: Inserted module 'overlay' Feb 13 19:05:24.944615 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:05:24.944638 kernel: Bridge firewalling registered Feb 13 19:05:24.945965 systemd-modules-load[239]: Inserted module 'br_netfilter' Feb 13 19:05:24.946942 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:05:24.960740 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:05:24.962534 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:05:24.966967 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:05:24.970090 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:05:24.971845 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:05:24.979418 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:05:24.980989 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:05:24.993774 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:05:24.995085 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:05:24.998008 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:05:25.011233 dracut-cmdline[279]: dracut-dracut-053 Feb 13 19:05:25.014036 dracut-cmdline[279]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:05:25.033647 systemd-resolved[276]: Positive Trust Anchors: Feb 13 19:05:25.033667 systemd-resolved[276]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:05:25.033699 systemd-resolved[276]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:05:25.038379 systemd-resolved[276]: Defaulting to hostname 'linux'. Feb 13 19:05:25.039435 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:05:25.043962 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:05:25.095594 kernel: SCSI subsystem initialized Feb 13 19:05:25.100581 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:05:25.111591 kernel: iscsi: registered transport (tcp) Feb 13 19:05:25.125585 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:05:25.125615 kernel: QLogic iSCSI HBA Driver Feb 13 19:05:25.173322 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:05:25.180776 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:05:25.201596 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:05:25.201664 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:05:25.201676 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:05:25.249620 kernel: raid6: neonx8 gen() 15647 MB/s Feb 13 19:05:25.266594 kernel: raid6: neonx4 gen() 15728 MB/s Feb 13 19:05:25.283600 kernel: raid6: neonx2 gen() 13145 MB/s Feb 13 19:05:25.300594 kernel: raid6: neonx1 gen() 10469 MB/s Feb 13 19:05:25.317589 kernel: raid6: int64x8 gen() 6780 MB/s Feb 13 19:05:25.334586 kernel: raid6: int64x4 gen() 7344 MB/s Feb 13 19:05:25.351586 kernel: raid6: int64x2 gen() 6106 MB/s Feb 13 19:05:25.368717 kernel: raid6: int64x1 gen() 5049 MB/s Feb 13 19:05:25.368729 kernel: raid6: using algorithm neonx4 gen() 15728 MB/s Feb 13 19:05:25.386752 kernel: raid6: .... xor() 12446 MB/s, rmw enabled Feb 13 19:05:25.386765 kernel: raid6: using neon recovery algorithm Feb 13 19:05:25.392977 kernel: xor: measuring software checksum speed Feb 13 19:05:25.392994 kernel: 8regs : 21624 MB/sec Feb 13 19:05:25.393011 kernel: 32regs : 21670 MB/sec Feb 13 19:05:25.393663 kernel: arm64_neon : 27738 MB/sec Feb 13 19:05:25.393675 kernel: xor: using function: arm64_neon (27738 MB/sec) Feb 13 19:05:25.452154 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:05:25.467629 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:05:25.481784 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:05:25.495271 systemd-udevd[461]: Using default interface naming scheme 'v255'. Feb 13 19:05:25.499138 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:05:25.509953 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:05:25.522383 dracut-pre-trigger[472]: rd.md=0: removing MD RAID activation Feb 13 19:05:25.551625 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:05:25.564762 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:05:25.607601 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:05:25.618807 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:05:25.630661 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:05:25.632371 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:05:25.634147 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:05:25.638090 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:05:25.646728 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:05:25.661605 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:05:25.676008 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Feb 13 19:05:25.687180 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Feb 13 19:05:25.687296 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:05:25.687313 kernel: GPT:9289727 != 19775487 Feb 13 19:05:25.687323 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:05:25.687334 kernel: GPT:9289727 != 19775487 Feb 13 19:05:25.687409 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:05:25.687424 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:05:25.688782 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:05:25.689024 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:05:25.691650 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:05:25.694064 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:05:25.694331 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:05:25.699707 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:05:25.710747 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:05:25.712401 kernel: BTRFS: device fsid 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (508) Feb 13 19:05:25.718575 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (519) Feb 13 19:05:25.732986 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:05:25.742364 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Feb 13 19:05:25.750452 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Feb 13 19:05:25.757119 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Feb 13 19:05:25.758545 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Feb 13 19:05:25.768366 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:05:25.785758 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:05:25.787742 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:05:25.794895 disk-uuid[553]: Primary Header is updated. Feb 13 19:05:25.794895 disk-uuid[553]: Secondary Entries is updated. Feb 13 19:05:25.794895 disk-uuid[553]: Secondary Header is updated. Feb 13 19:05:25.800125 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:05:25.814128 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:05:26.821519 disk-uuid[554]: The operation has completed successfully. Feb 13 19:05:26.822662 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:05:26.851593 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:05:26.851691 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:05:26.894753 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:05:26.897712 sh[573]: Success Feb 13 19:05:26.915590 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Feb 13 19:05:26.963582 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:05:26.965546 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:05:26.968088 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:05:26.982020 kernel: BTRFS info (device dm-0): first mount of filesystem 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 Feb 13 19:05:26.982078 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:05:26.984598 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:05:26.984646 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:05:26.984658 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:05:26.989733 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:05:26.991126 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:05:27.001787 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:05:27.003419 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:05:27.014868 kernel: BTRFS info (device vda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:05:27.014914 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:05:27.014925 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:05:27.017658 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:05:27.024886 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:05:27.027308 kernel: BTRFS info (device vda6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:05:27.032871 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:05:27.040867 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:05:27.115082 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:05:27.126731 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:05:27.156410 ignition[669]: Ignition 2.20.0 Feb 13 19:05:27.156421 ignition[669]: Stage: fetch-offline Feb 13 19:05:27.156455 ignition[669]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:27.156463 ignition[669]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:27.158958 systemd-networkd[768]: lo: Link UP Feb 13 19:05:27.156644 ignition[669]: parsed url from cmdline: "" Feb 13 19:05:27.158962 systemd-networkd[768]: lo: Gained carrier Feb 13 19:05:27.156648 ignition[669]: no config URL provided Feb 13 19:05:27.160056 systemd-networkd[768]: Enumeration completed Feb 13 19:05:27.156654 ignition[669]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:05:27.160606 systemd-networkd[768]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:05:27.156663 ignition[669]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:05:27.160609 systemd-networkd[768]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:05:27.156686 ignition[669]: op(1): [started] loading QEMU firmware config module Feb 13 19:05:27.160933 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:05:27.156690 ignition[669]: op(1): executing: "modprobe" "qemu_fw_cfg" Feb 13 19:05:27.161712 systemd-networkd[768]: eth0: Link UP Feb 13 19:05:27.173574 ignition[669]: op(1): [finished] loading QEMU firmware config module Feb 13 19:05:27.161715 systemd-networkd[768]: eth0: Gained carrier Feb 13 19:05:27.161722 systemd-networkd[768]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:05:27.162342 systemd[1]: Reached target network.target - Network. Feb 13 19:05:27.184609 systemd-networkd[768]: eth0: DHCPv4 address 10.0.0.60/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:05:27.199681 ignition[669]: parsing config with SHA512: 09f3bd3b1dc22e77d5c5d24a1eb533139bf6e775bcc0c8f6f4938da39b1649b525b5f16b5aaab79790c266d7bfb8e2b7b02db30e1c71b95d71bd38ae83ca61dd Feb 13 19:05:27.204195 unknown[669]: fetched base config from "system" Feb 13 19:05:27.204204 unknown[669]: fetched user config from "qemu" Feb 13 19:05:27.204599 ignition[669]: fetch-offline: fetch-offline passed Feb 13 19:05:27.206449 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:05:27.204679 ignition[669]: Ignition finished successfully Feb 13 19:05:27.207749 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Feb 13 19:05:27.218740 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:05:27.231406 ignition[777]: Ignition 2.20.0 Feb 13 19:05:27.231422 ignition[777]: Stage: kargs Feb 13 19:05:27.231623 ignition[777]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:27.234767 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:05:27.231634 ignition[777]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:27.232597 ignition[777]: kargs: kargs passed Feb 13 19:05:27.232641 ignition[777]: Ignition finished successfully Feb 13 19:05:27.252786 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:05:27.262321 ignition[787]: Ignition 2.20.0 Feb 13 19:05:27.262333 ignition[787]: Stage: disks Feb 13 19:05:27.262503 ignition[787]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:27.262512 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:27.265179 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:05:27.263396 ignition[787]: disks: disks passed Feb 13 19:05:27.267129 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:05:27.263441 ignition[787]: Ignition finished successfully Feb 13 19:05:27.268920 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:05:27.270753 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:05:27.272752 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:05:27.274554 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:05:27.282721 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:05:27.292513 systemd-fsck[798]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 19:05:27.297027 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:05:27.992676 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:05:28.038584 kernel: EXT4-fs (vda9): mounted filesystem 24882d04-b1a5-4a27-95f1-925956e69b18 r/w with ordered data mode. Quota mode: none. Feb 13 19:05:28.038632 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:05:28.040036 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:05:28.052655 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:05:28.054532 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:05:28.055844 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 19:05:28.055925 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:05:28.064190 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (806) Feb 13 19:05:28.055954 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:05:28.069163 kernel: BTRFS info (device vda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:05:28.069184 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:05:28.069195 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:05:28.060288 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:05:28.062794 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:05:28.072612 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:05:28.074063 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:05:28.107891 initrd-setup-root[831]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:05:28.112230 initrd-setup-root[838]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:05:28.116741 initrd-setup-root[845]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:05:28.120615 initrd-setup-root[852]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:05:28.198444 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:05:28.216742 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:05:28.219197 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:05:28.224575 kernel: BTRFS info (device vda6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:05:28.239852 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:05:28.241774 ignition[921]: INFO : Ignition 2.20.0 Feb 13 19:05:28.241774 ignition[921]: INFO : Stage: mount Feb 13 19:05:28.241774 ignition[921]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:28.241774 ignition[921]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:28.247232 ignition[921]: INFO : mount: mount passed Feb 13 19:05:28.247232 ignition[921]: INFO : Ignition finished successfully Feb 13 19:05:28.243772 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:05:28.256688 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:05:28.415894 systemd-networkd[768]: eth0: Gained IPv6LL Feb 13 19:05:28.980134 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:05:28.990760 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:05:28.996590 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (933) Feb 13 19:05:28.996629 kernel: BTRFS info (device vda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:05:28.998681 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:05:28.998706 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:05:29.001576 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:05:29.002712 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:05:29.019740 ignition[950]: INFO : Ignition 2.20.0 Feb 13 19:05:29.019740 ignition[950]: INFO : Stage: files Feb 13 19:05:29.021480 ignition[950]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:29.021480 ignition[950]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:29.021480 ignition[950]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:05:29.025101 ignition[950]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:05:29.025101 ignition[950]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:05:29.028282 ignition[950]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:05:29.029654 ignition[950]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:05:29.029654 ignition[950]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:05:29.028862 unknown[950]: wrote ssh authorized keys file for user: core Feb 13 19:05:29.033586 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Feb 13 19:05:29.033586 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Feb 13 19:05:29.103238 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:05:30.130815 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Feb 13 19:05:30.133264 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Feb 13 19:05:30.368459 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:05:30.621470 ignition[950]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Feb 13 19:05:30.621470 ignition[950]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Feb 13 19:05:30.625359 ignition[950]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Feb 13 19:05:30.649819 ignition[950]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:05:30.653629 ignition[950]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:05:30.655209 ignition[950]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Feb 13 19:05:30.655209 ignition[950]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:05:30.655209 ignition[950]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:05:30.655209 ignition[950]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:05:30.655209 ignition[950]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:05:30.655209 ignition[950]: INFO : files: files passed Feb 13 19:05:30.655209 ignition[950]: INFO : Ignition finished successfully Feb 13 19:05:30.657268 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:05:30.669796 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:05:30.671902 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:05:30.673341 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:05:30.673437 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:05:30.679972 initrd-setup-root-after-ignition[979]: grep: /sysroot/oem/oem-release: No such file or directory Feb 13 19:05:30.686746 initrd-setup-root-after-ignition[981]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:05:30.686746 initrd-setup-root-after-ignition[981]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:05:30.690068 initrd-setup-root-after-ignition[985]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:05:30.689334 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:05:30.691806 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:05:30.701799 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:05:30.723501 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:05:30.723763 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:05:30.726006 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:05:30.727972 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:05:30.729981 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:05:30.730889 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:05:30.746951 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:05:30.761780 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:05:30.770092 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:05:30.771459 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:05:30.773615 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:05:30.775503 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:05:30.775656 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:05:30.778340 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:05:30.780557 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:05:30.782335 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:05:30.784181 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:05:30.786296 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:05:30.788438 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:05:30.790450 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:05:30.792610 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:05:30.794837 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:05:30.796591 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:05:30.798225 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:05:30.798368 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:05:30.800788 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:05:30.802869 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:05:30.804885 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:05:30.805635 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:05:30.807030 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:05:30.807169 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:05:30.810107 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:05:30.810253 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:05:30.812308 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:05:30.814144 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:05:30.817158 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:05:30.818648 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:05:30.820577 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:05:30.822886 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:05:30.822983 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:05:30.824698 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:05:30.824777 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:05:30.826608 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:05:30.826730 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:05:30.828628 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:05:30.828738 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:05:30.844861 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:05:30.845896 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:05:30.846042 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:05:30.849158 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:05:30.851105 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:05:30.851253 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:05:30.858073 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:05:30.858200 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:05:30.861741 ignition[1005]: INFO : Ignition 2.20.0 Feb 13 19:05:30.861741 ignition[1005]: INFO : Stage: umount Feb 13 19:05:30.861741 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:05:30.861741 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:05:30.865917 ignition[1005]: INFO : umount: umount passed Feb 13 19:05:30.865917 ignition[1005]: INFO : Ignition finished successfully Feb 13 19:05:30.864502 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:05:30.865870 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:05:30.869083 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:05:30.874243 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:05:30.874345 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:05:30.876736 systemd[1]: Stopped target network.target - Network. Feb 13 19:05:30.877766 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:05:30.877857 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:05:30.879726 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:05:30.879784 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:05:30.881934 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:05:30.882001 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:05:30.883719 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:05:30.883769 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:05:30.888000 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:05:30.889853 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:05:30.892052 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:05:30.892156 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:05:30.894396 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:05:30.894485 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:05:30.899474 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:05:30.899626 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:05:30.904825 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Feb 13 19:05:30.905075 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:05:30.905230 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:05:30.908324 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Feb 13 19:05:30.909080 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:05:30.909138 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:05:30.918676 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:05:30.919977 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:05:30.920052 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:05:30.926044 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:05:30.926110 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:05:30.929411 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:05:30.929468 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:05:30.930875 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:05:30.930926 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:05:30.934133 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:05:30.936198 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 13 19:05:30.936263 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:05:30.946714 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:05:30.946821 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:05:30.950318 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:05:30.950466 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:05:30.952854 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:05:30.952891 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:05:30.954090 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:05:30.954123 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:05:30.956318 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:05:30.956373 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:05:30.959330 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:05:30.959382 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:05:30.962343 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:05:30.962393 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:05:30.982769 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:05:30.983924 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:05:30.983996 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:05:30.987347 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:05:30.987398 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:05:30.991536 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Feb 13 19:05:30.991631 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:05:30.991920 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:05:30.992005 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:05:30.995146 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:05:30.997233 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:05:31.007188 systemd[1]: Switching root. Feb 13 19:05:31.034928 systemd-journald[237]: Journal stopped Feb 13 19:05:31.858908 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). Feb 13 19:05:31.858965 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:05:31.858978 kernel: SELinux: policy capability open_perms=1 Feb 13 19:05:31.858988 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:05:31.858998 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:05:31.859008 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:05:31.859017 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:05:31.859031 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:05:31.859040 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:05:31.859051 kernel: audit: type=1403 audit(1739473531.174:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:05:31.859061 systemd[1]: Successfully loaded SELinux policy in 33.024ms. Feb 13 19:05:31.859077 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 12.036ms. Feb 13 19:05:31.859089 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:05:31.859099 systemd[1]: Detected virtualization kvm. Feb 13 19:05:31.859109 systemd[1]: Detected architecture arm64. Feb 13 19:05:31.859120 systemd[1]: Detected first boot. Feb 13 19:05:31.859131 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:05:31.859183 zram_generator::config[1052]: No configuration found. Feb 13 19:05:31.859203 kernel: NET: Registered PF_VSOCK protocol family Feb 13 19:05:31.859213 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:05:31.859224 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Feb 13 19:05:31.859235 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:05:31.859245 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:05:31.859255 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:05:31.859266 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:05:31.859277 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:05:31.859290 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:05:31.859301 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:05:31.859312 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:05:31.859323 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:05:31.859333 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:05:31.859344 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:05:31.859355 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:05:31.859366 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:05:31.859376 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:05:31.859388 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:05:31.859399 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:05:31.859409 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:05:31.859419 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Feb 13 19:05:31.859431 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:05:31.859441 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:05:31.859474 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:05:31.859489 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:05:31.859500 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:05:31.859510 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:05:31.859521 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:05:31.859531 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:05:31.859548 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:05:31.859616 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:05:31.859632 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:05:31.859642 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Feb 13 19:05:31.859655 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:05:31.859666 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:05:31.859676 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:05:31.859686 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:05:31.859696 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:05:31.859707 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:05:31.859718 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:05:31.859728 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:05:31.859739 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:05:31.859752 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:05:31.859768 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:05:31.859779 systemd[1]: Reached target machines.target - Containers. Feb 13 19:05:31.859789 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:05:31.859800 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:05:31.859811 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:05:31.859821 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:05:31.859832 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:05:31.859848 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:05:31.859861 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:05:31.859871 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:05:31.859881 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:05:31.859892 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:05:31.859902 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:05:31.859911 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:05:31.859925 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:05:31.859935 kernel: fuse: init (API version 7.39) Feb 13 19:05:31.859947 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:05:31.859958 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:05:31.859968 kernel: loop: module loaded Feb 13 19:05:31.859977 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:05:31.859989 kernel: ACPI: bus type drm_connector registered Feb 13 19:05:31.859999 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:05:31.860009 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:05:31.860019 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:05:31.860029 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Feb 13 19:05:31.860040 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:05:31.860051 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:05:31.860061 systemd[1]: Stopped verity-setup.service. Feb 13 19:05:31.860098 systemd-journald[1124]: Collecting audit messages is disabled. Feb 13 19:05:31.860126 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:05:31.860138 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:05:31.860149 systemd-journald[1124]: Journal started Feb 13 19:05:31.860170 systemd-journald[1124]: Runtime Journal (/run/log/journal/88dace6cfa6445e8a6cc25991f14bbc2) is 5.9M, max 47.3M, 41.4M free. Feb 13 19:05:31.617855 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:05:31.628610 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Feb 13 19:05:31.629009 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:05:31.864414 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:05:31.865146 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:05:31.866378 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:05:31.867703 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:05:31.869043 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:05:31.871602 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:05:31.873126 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:05:31.874778 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:05:31.874949 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:05:31.876580 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:05:31.876752 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:05:31.878466 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:05:31.878681 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:05:31.880102 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:05:31.880263 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:05:31.883023 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:05:31.883192 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:05:31.884689 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:05:31.884851 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:05:31.886441 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:05:31.888042 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:05:31.889781 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:05:31.891582 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Feb 13 19:05:31.905880 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:05:31.923703 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:05:31.926144 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:05:31.927383 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:05:31.927427 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:05:31.929535 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Feb 13 19:05:31.932056 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:05:31.934489 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:05:31.935785 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:05:31.937002 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:05:31.939374 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:05:31.940743 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:05:31.941743 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:05:31.943014 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:05:31.946778 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:05:31.951130 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:05:31.954399 systemd-journald[1124]: Time spent on flushing to /var/log/journal/88dace6cfa6445e8a6cc25991f14bbc2 is 15.394ms for 868 entries. Feb 13 19:05:31.954399 systemd-journald[1124]: System Journal (/var/log/journal/88dace6cfa6445e8a6cc25991f14bbc2) is 8M, max 195.6M, 187.6M free. Feb 13 19:05:31.976218 systemd-journald[1124]: Received client request to flush runtime journal. Feb 13 19:05:31.976260 kernel: loop0: detected capacity change from 0 to 113512 Feb 13 19:05:31.957132 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:05:31.962520 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:05:31.968513 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:05:31.979276 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:05:31.986198 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:05:31.982058 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:05:31.986823 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:05:31.990798 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:05:31.993549 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:05:31.999250 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:05:32.009779 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Feb 13 19:05:32.012879 kernel: loop1: detected capacity change from 0 to 123192 Feb 13 19:05:32.013119 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:05:32.014820 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:05:32.021812 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:05:32.035637 udevadm[1185]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 13 19:05:32.047480 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Feb 13 19:05:32.049736 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Feb 13 19:05:32.051721 kernel: loop2: detected capacity change from 0 to 194096 Feb 13 19:05:32.049755 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Feb 13 19:05:32.054065 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:05:32.095640 kernel: loop3: detected capacity change from 0 to 113512 Feb 13 19:05:32.102598 kernel: loop4: detected capacity change from 0 to 123192 Feb 13 19:05:32.114585 kernel: loop5: detected capacity change from 0 to 194096 Feb 13 19:05:32.122464 (sd-merge)[1195]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Feb 13 19:05:32.122930 (sd-merge)[1195]: Merged extensions into '/usr'. Feb 13 19:05:32.129387 systemd[1]: Reload requested from client PID 1169 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:05:32.129403 systemd[1]: Reloading... Feb 13 19:05:32.177755 zram_generator::config[1220]: No configuration found. Feb 13 19:05:32.276911 ldconfig[1164]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:05:32.292144 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:05:32.341808 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:05:32.341963 systemd[1]: Reloading finished in 212 ms. Feb 13 19:05:32.358380 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:05:32.360115 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:05:32.378233 systemd[1]: Starting ensure-sysext.service... Feb 13 19:05:32.380299 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:05:32.392524 systemd[1]: Reload requested from client PID 1257 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:05:32.392547 systemd[1]: Reloading... Feb 13 19:05:32.399294 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:05:32.399502 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:05:32.400713 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:05:32.401030 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Feb 13 19:05:32.401157 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Feb 13 19:05:32.409947 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:05:32.410030 systemd-tmpfiles[1258]: Skipping /boot Feb 13 19:05:32.420155 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:05:32.420170 systemd-tmpfiles[1258]: Skipping /boot Feb 13 19:05:32.443593 zram_generator::config[1285]: No configuration found. Feb 13 19:05:32.540098 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:05:32.594403 systemd[1]: Reloading finished in 201 ms. Feb 13 19:05:32.607017 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:05:32.624078 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:05:32.634396 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:05:32.637307 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:05:32.639817 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:05:32.642870 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:05:32.648941 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:05:32.651375 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:05:32.658188 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:05:32.659922 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:05:32.665995 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:05:32.668607 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:05:32.669870 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:05:32.670027 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:05:32.671656 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:05:32.673412 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:05:32.673655 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:05:32.675683 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:05:32.675887 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:05:32.677724 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:05:32.677961 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:05:32.692930 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:05:32.698692 systemd-udevd[1328]: Using default interface naming scheme 'v255'. Feb 13 19:05:32.708597 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:05:32.711070 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:05:32.715360 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:05:32.717498 augenrules[1358]: No rules Feb 13 19:05:32.718014 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:05:32.718161 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:05:32.719922 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:05:32.722815 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:05:32.726088 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:05:32.729139 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:05:32.729379 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:05:32.733232 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:05:32.735930 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:05:32.739404 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:05:32.739685 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:05:32.743128 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:05:32.743329 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:05:32.745213 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:05:32.745376 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:05:32.752975 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:05:32.768632 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1377) Feb 13 19:05:32.768982 systemd[1]: Finished ensure-sysext.service. Feb 13 19:05:32.785703 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:05:32.787278 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:05:32.789238 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:05:32.791731 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:05:32.796532 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:05:32.802777 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:05:32.803900 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:05:32.803972 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:05:32.807460 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:05:32.813774 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Feb 13 19:05:32.817638 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:05:32.818197 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:05:32.819840 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:05:32.820020 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:05:32.820076 augenrules[1392]: /sbin/augenrules: No change Feb 13 19:05:32.821666 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:05:32.821841 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:05:32.823245 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:05:32.823417 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:05:32.825043 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:05:32.825215 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:05:32.834833 augenrules[1429]: No rules Feb 13 19:05:32.835475 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:05:32.836409 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:05:32.854117 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Feb 13 19:05:32.860800 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:05:32.872037 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:05:32.873763 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:05:32.873845 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:05:32.891840 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:05:32.895600 systemd-resolved[1326]: Positive Trust Anchors: Feb 13 19:05:32.895619 systemd-resolved[1326]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:05:32.895652 systemd-resolved[1326]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:05:32.903118 systemd-resolved[1326]: Defaulting to hostname 'linux'. Feb 13 19:05:32.906886 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:05:32.908649 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:05:32.914192 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Feb 13 19:05:32.915963 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:05:32.918968 systemd-networkd[1409]: lo: Link UP Feb 13 19:05:32.918976 systemd-networkd[1409]: lo: Gained carrier Feb 13 19:05:32.919885 systemd-networkd[1409]: Enumeration completed Feb 13 19:05:32.919986 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:05:32.921580 systemd[1]: Reached target network.target - Network. Feb 13 19:05:32.922153 systemd-networkd[1409]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:05:32.922162 systemd-networkd[1409]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:05:32.922822 systemd-networkd[1409]: eth0: Link UP Feb 13 19:05:32.922828 systemd-networkd[1409]: eth0: Gained carrier Feb 13 19:05:32.922843 systemd-networkd[1409]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:05:32.932895 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Feb 13 19:05:32.933243 systemd-networkd[1409]: eth0: DHCPv4 address 10.0.0.60/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:05:32.933876 systemd-timesyncd[1412]: Network configuration changed, trying to establish connection. Feb 13 19:05:32.936774 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:05:32.936833 systemd-timesyncd[1412]: Contacted time server 10.0.0.1:123 (10.0.0.1). Feb 13 19:05:32.936880 systemd-timesyncd[1412]: Initial clock synchronization to Thu 2025-02-13 19:05:32.583642 UTC. Feb 13 19:05:32.952321 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:05:32.954011 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Feb 13 19:05:32.966672 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:05:32.978997 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:05:33.012602 lvm[1450]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:05:33.017237 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:05:33.044223 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:05:33.045825 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:05:33.047023 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:05:33.048203 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:05:33.049495 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:05:33.050948 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:05:33.052274 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:05:33.053607 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:05:33.054827 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:05:33.054864 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:05:33.055840 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:05:33.057459 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:05:33.060182 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:05:33.063769 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Feb 13 19:05:33.065272 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Feb 13 19:05:33.066589 systemd[1]: Reached target ssh-access.target - SSH Access Available. Feb 13 19:05:33.070046 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:05:33.071988 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Feb 13 19:05:33.074444 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:05:33.076297 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:05:33.077586 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:05:33.078551 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:05:33.079570 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:05:33.079618 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:05:33.080693 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:05:33.082496 lvm[1458]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:05:33.083807 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:05:33.086721 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:05:33.090262 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:05:33.091453 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:05:33.092713 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:05:33.097817 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:05:33.100823 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:05:33.105806 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:05:33.111121 jq[1461]: false Feb 13 19:05:33.114791 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:05:33.120892 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:05:33.121073 extend-filesystems[1462]: Found loop3 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found loop4 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found loop5 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda1 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda2 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda3 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found usr Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda4 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda6 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda7 Feb 13 19:05:33.121073 extend-filesystems[1462]: Found vda9 Feb 13 19:05:33.121073 extend-filesystems[1462]: Checking size of /dev/vda9 Feb 13 19:05:33.121439 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:05:33.147737 dbus-daemon[1460]: [system] SELinux support is enabled Feb 13 19:05:33.158371 extend-filesystems[1462]: Resized partition /dev/vda9 Feb 13 19:05:33.124789 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:05:33.159459 jq[1476]: true Feb 13 19:05:33.126938 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:05:33.132691 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:05:33.139732 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:05:33.139916 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:05:33.141477 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:05:33.141693 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:05:33.154885 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:05:33.159507 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:05:33.161836 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:05:33.177746 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1369) Feb 13 19:05:33.180578 extend-filesystems[1485]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:05:33.190626 (ntainerd)[1488]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:05:33.202358 jq[1486]: true Feb 13 19:05:33.213143 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Feb 13 19:05:33.213214 tar[1484]: linux-arm64/helm Feb 13 19:05:33.217202 update_engine[1473]: I20250213 19:05:33.211491 1473 main.cc:92] Flatcar Update Engine starting Feb 13 19:05:33.210613 systemd-logind[1469]: Watching system buttons on /dev/input/event0 (Power Button) Feb 13 19:05:33.213587 systemd-logind[1469]: New seat seat0. Feb 13 19:05:33.216024 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:05:33.222196 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:05:33.223198 update_engine[1473]: I20250213 19:05:33.223124 1473 update_check_scheduler.cc:74] Next update check in 7m39s Feb 13 19:05:33.225348 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:05:33.225578 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:05:33.227514 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:05:33.227635 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:05:33.236860 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:05:33.367772 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Feb 13 19:05:33.392843 locksmithd[1507]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:05:33.544398 containerd[1488]: time="2025-02-13T19:05:33.543787722Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 19:05:33.545316 extend-filesystems[1485]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Feb 13 19:05:33.545316 extend-filesystems[1485]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 19:05:33.545316 extend-filesystems[1485]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Feb 13 19:05:33.552810 extend-filesystems[1462]: Resized filesystem in /dev/vda9 Feb 13 19:05:33.549240 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:05:33.562398 bash[1514]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:05:33.549461 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:05:33.556730 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:05:33.564781 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Feb 13 19:05:33.581737 containerd[1488]: time="2025-02-13T19:05:33.581686069Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.583111 containerd[1488]: time="2025-02-13T19:05:33.583062140Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:05:33.583863 sshd_keygen[1483]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583187282Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583218864Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583372186Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583389544Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583458902Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583471787Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583704982Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583721231Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583735072Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583744516Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584015 containerd[1488]: time="2025-02-13T19:05:33.583838536Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584195 containerd[1488]: time="2025-02-13T19:05:33.584028639Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584195 containerd[1488]: time="2025-02-13T19:05:33.584144376Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:05:33.584195 containerd[1488]: time="2025-02-13T19:05:33.584157070Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:05:33.584248 containerd[1488]: time="2025-02-13T19:05:33.584231207Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:05:33.584347 containerd[1488]: time="2025-02-13T19:05:33.584272998Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:05:33.587858 containerd[1488]: time="2025-02-13T19:05:33.587821150Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588087532Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588115788Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588131923Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588146834Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588311435Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588553959Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588684645Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588714316Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588728462Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588742648Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588762491Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588775262Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588787497Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590581 containerd[1488]: time="2025-02-13T19:05:33.588800535Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588812158Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588825082Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588835749Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588855211Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588868287Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588879643Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588893866Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588907631Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588920325Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588932216Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588948771Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588961045Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588974542Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.590880 containerd[1488]: time="2025-02-13T19:05:33.588986012Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.588997865Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589009680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589023444Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589049138Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589063629Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589073914Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589233162Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589250329Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589259812Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589270326Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589279541Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589291088Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589300494Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:05:33.591096 containerd[1488]: time="2025-02-13T19:05:33.589310320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:05:33.591316 containerd[1488]: time="2025-02-13T19:05:33.589693394Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:05:33.591316 containerd[1488]: time="2025-02-13T19:05:33.589744094Z" level=info msg="Connect containerd service" Feb 13 19:05:33.591316 containerd[1488]: time="2025-02-13T19:05:33.589783285Z" level=info msg="using legacy CRI server" Feb 13 19:05:33.591316 containerd[1488]: time="2025-02-13T19:05:33.589790129Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:05:33.591316 containerd[1488]: time="2025-02-13T19:05:33.590006538Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:05:33.591874 containerd[1488]: time="2025-02-13T19:05:33.591847123Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:05:33.592499 containerd[1488]: time="2025-02-13T19:05:33.592448557Z" level=info msg="Start subscribing containerd event" Feb 13 19:05:33.592642 containerd[1488]: time="2025-02-13T19:05:33.592525906Z" level=info msg="Start recovering state" Feb 13 19:05:33.592642 containerd[1488]: time="2025-02-13T19:05:33.592628643Z" level=info msg="Start event monitor" Feb 13 19:05:33.592688 containerd[1488]: time="2025-02-13T19:05:33.592640304Z" level=info msg="Start snapshots syncer" Feb 13 19:05:33.592688 containerd[1488]: time="2025-02-13T19:05:33.592653304Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:05:33.592688 containerd[1488]: time="2025-02-13T19:05:33.592660339Z" level=info msg="Start streaming server" Feb 13 19:05:33.593159 containerd[1488]: time="2025-02-13T19:05:33.593129289Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:05:33.593470 containerd[1488]: time="2025-02-13T19:05:33.593452832Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:05:33.593801 containerd[1488]: time="2025-02-13T19:05:33.593786087Z" level=info msg="containerd successfully booted in 0.071394s" Feb 13 19:05:33.593919 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:05:33.604402 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:05:33.612923 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:05:33.618324 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:05:33.618577 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:05:33.622920 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:05:33.638490 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:05:33.649281 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:05:33.652168 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Feb 13 19:05:33.653729 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:05:33.757873 tar[1484]: linux-arm64/LICENSE Feb 13 19:05:33.757969 tar[1484]: linux-arm64/README.md Feb 13 19:05:33.772208 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:05:34.559689 systemd-networkd[1409]: eth0: Gained IPv6LL Feb 13 19:05:34.563627 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:05:34.565612 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:05:34.582035 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Feb 13 19:05:34.584715 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:05:34.586974 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:05:34.601620 systemd[1]: coreos-metadata.service: Deactivated successfully. Feb 13 19:05:34.601850 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Feb 13 19:05:34.604052 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:05:34.607463 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:05:35.068190 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:35.069880 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:05:35.071062 systemd[1]: Startup finished in 574ms (kernel) + 6.468s (initrd) + 3.931s (userspace) = 10.974s. Feb 13 19:05:35.071956 (kubelet)[1573]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:05:35.572829 kubelet[1573]: E0213 19:05:35.572753 1573 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:05:35.574737 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:05:35.574903 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:05:35.575224 systemd[1]: kubelet.service: Consumed 849ms CPU time, 241.4M memory peak. Feb 13 19:05:37.070953 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:05:37.072009 systemd[1]: Started sshd@0-10.0.0.60:22-10.0.0.1:59740.service - OpenSSH per-connection server daemon (10.0.0.1:59740). Feb 13 19:05:37.131822 sshd[1587]: Accepted publickey for core from 10.0.0.1 port 59740 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:05:37.133519 sshd-session[1587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:05:37.139595 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:05:37.148896 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:05:37.154025 systemd-logind[1469]: New session 1 of user core. Feb 13 19:05:37.157294 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:05:37.159788 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:05:37.165336 (systemd)[1591]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:05:37.167252 systemd-logind[1469]: New session c1 of user core. Feb 13 19:05:37.265659 systemd[1591]: Queued start job for default target default.target. Feb 13 19:05:37.281512 systemd[1591]: Created slice app.slice - User Application Slice. Feb 13 19:05:37.281548 systemd[1591]: Reached target paths.target - Paths. Feb 13 19:05:37.281601 systemd[1591]: Reached target timers.target - Timers. Feb 13 19:05:37.282823 systemd[1591]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:05:37.291781 systemd[1591]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:05:37.291838 systemd[1591]: Reached target sockets.target - Sockets. Feb 13 19:05:37.291885 systemd[1591]: Reached target basic.target - Basic System. Feb 13 19:05:37.291915 systemd[1591]: Reached target default.target - Main User Target. Feb 13 19:05:37.291939 systemd[1591]: Startup finished in 119ms. Feb 13 19:05:37.292121 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:05:37.293336 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:05:37.370998 systemd[1]: Started sshd@1-10.0.0.60:22-10.0.0.1:59750.service - OpenSSH per-connection server daemon (10.0.0.1:59750). Feb 13 19:05:37.415678 sshd[1602]: Accepted publickey for core from 10.0.0.1 port 59750 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:05:37.416790 sshd-session[1602]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:05:37.421224 systemd-logind[1469]: New session 2 of user core. Feb 13 19:05:37.432719 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:05:37.513619 sshd[1604]: Connection closed by 10.0.0.1 port 59750 Feb 13 19:05:37.513943 sshd-session[1602]: pam_unix(sshd:session): session closed for user core Feb 13 19:05:37.526469 systemd[1]: sshd@1-10.0.0.60:22-10.0.0.1:59750.service: Deactivated successfully. Feb 13 19:05:37.527765 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:05:37.529262 systemd-logind[1469]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:05:37.546930 systemd[1]: Started sshd@2-10.0.0.60:22-10.0.0.1:59758.service - OpenSSH per-connection server daemon (10.0.0.1:59758). Feb 13 19:05:37.548901 systemd-logind[1469]: Removed session 2. Feb 13 19:05:37.587356 sshd[1609]: Accepted publickey for core from 10.0.0.1 port 59758 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:05:37.588427 sshd-session[1609]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:05:37.592407 systemd-logind[1469]: New session 3 of user core. Feb 13 19:05:37.599751 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:05:37.645825 sshd[1612]: Connection closed by 10.0.0.1 port 59758 Feb 13 19:05:37.646182 sshd-session[1609]: pam_unix(sshd:session): session closed for user core Feb 13 19:05:37.662178 systemd[1]: sshd@2-10.0.0.60:22-10.0.0.1:59758.service: Deactivated successfully. Feb 13 19:05:37.663553 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:05:37.664903 systemd-logind[1469]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:05:37.679830 systemd[1]: Started sshd@3-10.0.0.60:22-10.0.0.1:59770.service - OpenSSH per-connection server daemon (10.0.0.1:59770). Feb 13 19:05:37.680870 systemd-logind[1469]: Removed session 3. Feb 13 19:05:37.731407 sshd[1617]: Accepted publickey for core from 10.0.0.1 port 59770 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:05:37.732523 sshd-session[1617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:05:37.736440 systemd-logind[1469]: New session 4 of user core. Feb 13 19:05:37.751757 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:05:37.804656 sshd[1620]: Connection closed by 10.0.0.1 port 59770 Feb 13 19:05:37.805086 sshd-session[1617]: pam_unix(sshd:session): session closed for user core Feb 13 19:05:37.814426 systemd[1]: sshd@3-10.0.0.60:22-10.0.0.1:59770.service: Deactivated successfully. Feb 13 19:05:37.815796 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:05:37.817135 systemd-logind[1469]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:05:37.818340 systemd[1]: Started sshd@4-10.0.0.60:22-10.0.0.1:59776.service - OpenSSH per-connection server daemon (10.0.0.1:59776). Feb 13 19:05:37.819040 systemd-logind[1469]: Removed session 4. Feb 13 19:05:37.860592 sshd[1625]: Accepted publickey for core from 10.0.0.1 port 59776 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:05:37.861643 sshd-session[1625]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:05:37.865573 systemd-logind[1469]: New session 5 of user core. Feb 13 19:05:37.873753 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:05:37.937640 sudo[1629]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:05:37.937931 sudo[1629]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:05:38.318799 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:05:38.318943 (dockerd)[1649]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:05:38.594607 dockerd[1649]: time="2025-02-13T19:05:38.593954873Z" level=info msg="Starting up" Feb 13 19:05:38.797614 dockerd[1649]: time="2025-02-13T19:05:38.797569349Z" level=info msg="Loading containers: start." Feb 13 19:05:38.935580 kernel: Initializing XFRM netlink socket Feb 13 19:05:38.996010 systemd-networkd[1409]: docker0: Link UP Feb 13 19:05:39.026162 dockerd[1649]: time="2025-02-13T19:05:39.025722766Z" level=info msg="Loading containers: done." Feb 13 19:05:39.042201 dockerd[1649]: time="2025-02-13T19:05:39.042121660Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:05:39.042341 dockerd[1649]: time="2025-02-13T19:05:39.042218974Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Feb 13 19:05:39.042941 dockerd[1649]: time="2025-02-13T19:05:39.042377766Z" level=info msg="Daemon has completed initialization" Feb 13 19:05:39.076030 dockerd[1649]: time="2025-02-13T19:05:39.075978373Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:05:39.076075 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:05:39.731961 containerd[1488]: time="2025-02-13T19:05:39.731913527Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.10\"" Feb 13 19:05:40.432693 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount968529367.mount: Deactivated successfully. Feb 13 19:05:41.284407 containerd[1488]: time="2025-02-13T19:05:41.284309860Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:41.285333 containerd[1488]: time="2025-02-13T19:05:41.285263930Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.10: active requests=0, bytes read=29865209" Feb 13 19:05:41.285842 containerd[1488]: time="2025-02-13T19:05:41.285813902Z" level=info msg="ImageCreate event name:\"sha256:deaeae5e8513d8c5921aee5b515f0fc2ac63b71dfe965318f71eb49468e74a4f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:41.288945 containerd[1488]: time="2025-02-13T19:05:41.288910190Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:63b2b4b4e9b5dcb5b1b6cec9d5f5f538791a40cd8cb273ef530e6d6535aa0b43\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:41.290116 containerd[1488]: time="2025-02-13T19:05:41.290084548Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.10\" with image id \"sha256:deaeae5e8513d8c5921aee5b515f0fc2ac63b71dfe965318f71eb49468e74a4f\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:63b2b4b4e9b5dcb5b1b6cec9d5f5f538791a40cd8cb273ef530e6d6535aa0b43\", size \"29862007\" in 1.558128245s" Feb 13 19:05:41.290159 containerd[1488]: time="2025-02-13T19:05:41.290119766Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.10\" returns image reference \"sha256:deaeae5e8513d8c5921aee5b515f0fc2ac63b71dfe965318f71eb49468e74a4f\"" Feb 13 19:05:41.308513 containerd[1488]: time="2025-02-13T19:05:41.308472630Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.10\"" Feb 13 19:05:42.653613 containerd[1488]: time="2025-02-13T19:05:42.653413243Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:42.654025 containerd[1488]: time="2025-02-13T19:05:42.653826486Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.10: active requests=0, bytes read=26898596" Feb 13 19:05:42.654880 containerd[1488]: time="2025-02-13T19:05:42.654852766Z" level=info msg="ImageCreate event name:\"sha256:e31753dd49b05da8fcb7deb26f2a5942a6747a0e6d4492f3dc8544123b97a3a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:42.657807 containerd[1488]: time="2025-02-13T19:05:42.657765402Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:99b3336343ea48be24f1e64774825e9f8d5170bd2ed482ff336548eb824f5f58\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:42.660038 containerd[1488]: time="2025-02-13T19:05:42.660009507Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.10\" with image id \"sha256:e31753dd49b05da8fcb7deb26f2a5942a6747a0e6d4492f3dc8544123b97a3a2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:99b3336343ea48be24f1e64774825e9f8d5170bd2ed482ff336548eb824f5f58\", size \"28302323\" in 1.351494277s" Feb 13 19:05:42.660038 containerd[1488]: time="2025-02-13T19:05:42.660044595Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.10\" returns image reference \"sha256:e31753dd49b05da8fcb7deb26f2a5942a6747a0e6d4492f3dc8544123b97a3a2\"" Feb 13 19:05:42.678685 containerd[1488]: time="2025-02-13T19:05:42.678653691Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.10\"" Feb 13 19:05:43.540690 containerd[1488]: time="2025-02-13T19:05:43.540642387Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:43.541410 containerd[1488]: time="2025-02-13T19:05:43.541362093Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.10: active requests=0, bytes read=16164936" Feb 13 19:05:43.542964 containerd[1488]: time="2025-02-13T19:05:43.542907857Z" level=info msg="ImageCreate event name:\"sha256:ea60c047fad7c01bf50f1f0259a4aeea2cc4401850d5a95802cc1d07d9021eb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:43.545870 containerd[1488]: time="2025-02-13T19:05:43.545838198Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:cf7eb256192f1f51093fe278c209a9368f0675eb61ed01b148af47d2f21c002d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:43.547064 containerd[1488]: time="2025-02-13T19:05:43.547004696Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.10\" with image id \"sha256:ea60c047fad7c01bf50f1f0259a4aeea2cc4401850d5a95802cc1d07d9021eb4\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:cf7eb256192f1f51093fe278c209a9368f0675eb61ed01b148af47d2f21c002d\", size \"17568681\" in 868.184391ms" Feb 13 19:05:43.547064 containerd[1488]: time="2025-02-13T19:05:43.547038381Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.10\" returns image reference \"sha256:ea60c047fad7c01bf50f1f0259a4aeea2cc4401850d5a95802cc1d07d9021eb4\"" Feb 13 19:05:43.565382 containerd[1488]: time="2025-02-13T19:05:43.565333490Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.10\"" Feb 13 19:05:44.478175 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1283518866.mount: Deactivated successfully. Feb 13 19:05:44.674384 containerd[1488]: time="2025-02-13T19:05:44.674336252Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:44.678035 containerd[1488]: time="2025-02-13T19:05:44.676691481Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.10: active requests=0, bytes read=25663372" Feb 13 19:05:44.678693 containerd[1488]: time="2025-02-13T19:05:44.678640955Z" level=info msg="ImageCreate event name:\"sha256:fa8af75a6512774cc93242474a9841ace82a7d0646001149fc65d92a8bb0c00a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:44.681017 containerd[1488]: time="2025-02-13T19:05:44.680978763Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:d112e804e548fce28d9f1e3282c9ce54e374451e6a2c41b1ca9d7fca5d1fcc48\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:44.681704 containerd[1488]: time="2025-02-13T19:05:44.681624298Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.10\" with image id \"sha256:fa8af75a6512774cc93242474a9841ace82a7d0646001149fc65d92a8bb0c00a\", repo tag \"registry.k8s.io/kube-proxy:v1.30.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:d112e804e548fce28d9f1e3282c9ce54e374451e6a2c41b1ca9d7fca5d1fcc48\", size \"25662389\" in 1.116248883s" Feb 13 19:05:44.681704 containerd[1488]: time="2025-02-13T19:05:44.681660170Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.10\" returns image reference \"sha256:fa8af75a6512774cc93242474a9841ace82a7d0646001149fc65d92a8bb0c00a\"" Feb 13 19:05:44.699354 containerd[1488]: time="2025-02-13T19:05:44.699321203Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Feb 13 19:05:45.347409 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1297804177.mount: Deactivated successfully. Feb 13 19:05:45.825784 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:05:45.836229 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:05:45.926080 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:45.930334 (kubelet)[2001]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:05:46.042165 kubelet[2001]: E0213 19:05:46.042111 2001 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:05:46.045433 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:05:46.045638 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:05:46.046117 systemd[1]: kubelet.service: Consumed 136ms CPU time, 97.1M memory peak. Feb 13 19:05:46.048489 containerd[1488]: time="2025-02-13T19:05:46.048450511Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.049407 containerd[1488]: time="2025-02-13T19:05:46.049301877Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383" Feb 13 19:05:46.050753 containerd[1488]: time="2025-02-13T19:05:46.050703808Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.056543 containerd[1488]: time="2025-02-13T19:05:46.056489022Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.057973 containerd[1488]: time="2025-02-13T19:05:46.057885873Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.358369726s" Feb 13 19:05:46.057973 containerd[1488]: time="2025-02-13T19:05:46.057929808Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Feb 13 19:05:46.076710 containerd[1488]: time="2025-02-13T19:05:46.076577322Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Feb 13 19:05:46.560765 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3592433642.mount: Deactivated successfully. Feb 13 19:05:46.564158 containerd[1488]: time="2025-02-13T19:05:46.564113270Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.564558 containerd[1488]: time="2025-02-13T19:05:46.564491863Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823" Feb 13 19:05:46.570111 containerd[1488]: time="2025-02-13T19:05:46.570066725Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.577053 containerd[1488]: time="2025-02-13T19:05:46.577008354Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:46.578015 containerd[1488]: time="2025-02-13T19:05:46.577880398Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 501.245131ms" Feb 13 19:05:46.578015 containerd[1488]: time="2025-02-13T19:05:46.577914412Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Feb 13 19:05:46.596455 containerd[1488]: time="2025-02-13T19:05:46.596413489Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Feb 13 19:05:47.432449 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2565448864.mount: Deactivated successfully. Feb 13 19:05:48.995316 containerd[1488]: time="2025-02-13T19:05:48.995231232Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:48.995934 containerd[1488]: time="2025-02-13T19:05:48.995875850Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191474" Feb 13 19:05:48.996779 containerd[1488]: time="2025-02-13T19:05:48.996697926Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:49.000146 containerd[1488]: time="2025-02-13T19:05:49.000104599Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:05:49.001639 containerd[1488]: time="2025-02-13T19:05:49.001602379Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 2.405145015s" Feb 13 19:05:49.001702 containerd[1488]: time="2025-02-13T19:05:49.001639903Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" Feb 13 19:05:54.751525 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:54.751684 systemd[1]: kubelet.service: Consumed 136ms CPU time, 97.1M memory peak. Feb 13 19:05:54.762815 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:05:54.780209 systemd[1]: Reload requested from client PID 2153 ('systemctl') (unit session-5.scope)... Feb 13 19:05:54.780223 systemd[1]: Reloading... Feb 13 19:05:54.866130 zram_generator::config[2198]: No configuration found. Feb 13 19:05:54.993218 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:05:55.066939 systemd[1]: Reloading finished in 286 ms. Feb 13 19:05:55.106858 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:55.109308 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:05:55.110638 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:05:55.110846 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:55.110889 systemd[1]: kubelet.service: Consumed 77ms CPU time, 82.3M memory peak. Feb 13 19:05:55.112380 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:05:55.205028 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:05:55.208525 (kubelet)[2244]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:05:55.257537 kubelet[2244]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:05:55.257537 kubelet[2244]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 13 19:05:55.257537 kubelet[2244]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:05:55.257936 kubelet[2244]: I0213 19:05:55.257723 2244 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:05:56.727981 kubelet[2244]: I0213 19:05:56.727929 2244 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Feb 13 19:05:56.727981 kubelet[2244]: I0213 19:05:56.727965 2244 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:05:56.728350 kubelet[2244]: I0213 19:05:56.728175 2244 server.go:927] "Client rotation is on, will bootstrap in background" Feb 13 19:05:56.748146 kubelet[2244]: E0213 19:05:56.748097 2244 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.60:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.748457 kubelet[2244]: I0213 19:05:56.748215 2244 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:05:56.755228 kubelet[2244]: I0213 19:05:56.755198 2244 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:05:56.756417 kubelet[2244]: I0213 19:05:56.756365 2244 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:05:56.756624 kubelet[2244]: I0213 19:05:56.756417 2244 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Feb 13 19:05:56.756707 kubelet[2244]: I0213 19:05:56.756699 2244 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:05:56.756733 kubelet[2244]: I0213 19:05:56.756709 2244 container_manager_linux.go:301] "Creating device plugin manager" Feb 13 19:05:56.756990 kubelet[2244]: I0213 19:05:56.756964 2244 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:05:56.759684 kubelet[2244]: I0213 19:05:56.759662 2244 kubelet.go:400] "Attempting to sync node with API server" Feb 13 19:05:56.759732 kubelet[2244]: I0213 19:05:56.759686 2244 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:05:56.760276 kubelet[2244]: I0213 19:05:56.759894 2244 kubelet.go:312] "Adding apiserver pod source" Feb 13 19:05:56.760276 kubelet[2244]: I0213 19:05:56.759974 2244 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:05:56.761659 kubelet[2244]: W0213 19:05:56.761600 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.761792 kubelet[2244]: E0213 19:05:56.761778 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.761846 kubelet[2244]: W0213 19:05:56.761762 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.761927 kubelet[2244]: E0213 19:05:56.761918 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.765255 kubelet[2244]: I0213 19:05:56.765171 2244 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:05:56.765685 kubelet[2244]: I0213 19:05:56.765663 2244 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:05:56.765811 kubelet[2244]: W0213 19:05:56.765778 2244 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:05:56.766781 kubelet[2244]: I0213 19:05:56.766746 2244 server.go:1264] "Started kubelet" Feb 13 19:05:56.767507 kubelet[2244]: I0213 19:05:56.766953 2244 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:05:56.767507 kubelet[2244]: I0213 19:05:56.767318 2244 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:05:56.767696 kubelet[2244]: I0213 19:05:56.767680 2244 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:05:56.769086 kubelet[2244]: I0213 19:05:56.768069 2244 server.go:455] "Adding debug handlers to kubelet server" Feb 13 19:05:56.769086 kubelet[2244]: E0213 19:05:56.768741 2244 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.60:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.60:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1823d9fbd743f516 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-02-13 19:05:56.766717206 +0000 UTC m=+1.555010766,LastTimestamp:2025-02-13 19:05:56.766717206 +0000 UTC m=+1.555010766,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Feb 13 19:05:56.770325 kubelet[2244]: I0213 19:05:56.770300 2244 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:05:56.770660 kubelet[2244]: E0213 19:05:56.770626 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:05:56.771029 kubelet[2244]: I0213 19:05:56.771000 2244 volume_manager.go:291] "Starting Kubelet Volume Manager" Feb 13 19:05:56.771139 kubelet[2244]: I0213 19:05:56.771123 2244 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:05:56.771261 kubelet[2244]: I0213 19:05:56.771248 2244 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:05:56.771625 kubelet[2244]: W0213 19:05:56.771581 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.771676 kubelet[2244]: E0213 19:05:56.771633 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.772166 kubelet[2244]: E0213 19:05:56.772139 2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="200ms" Feb 13 19:05:56.773325 kubelet[2244]: E0213 19:05:56.773293 2244 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:05:56.773425 kubelet[2244]: I0213 19:05:56.773401 2244 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:05:56.773791 kubelet[2244]: I0213 19:05:56.773533 2244 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:05:56.774648 kubelet[2244]: I0213 19:05:56.774624 2244 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:05:56.785039 kubelet[2244]: I0213 19:05:56.784974 2244 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:05:56.786195 kubelet[2244]: I0213 19:05:56.786145 2244 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:05:56.786360 kubelet[2244]: I0213 19:05:56.786343 2244 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 13 19:05:56.786402 kubelet[2244]: I0213 19:05:56.786378 2244 kubelet.go:2337] "Starting kubelet main sync loop" Feb 13 19:05:56.786458 kubelet[2244]: E0213 19:05:56.786425 2244 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:05:56.790864 kubelet[2244]: W0213 19:05:56.790814 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.790864 kubelet[2244]: E0213 19:05:56.790867 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:56.791385 kubelet[2244]: I0213 19:05:56.791366 2244 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 13 19:05:56.791385 kubelet[2244]: I0213 19:05:56.791381 2244 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 13 19:05:56.791492 kubelet[2244]: I0213 19:05:56.791400 2244 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:05:56.854815 kubelet[2244]: I0213 19:05:56.854783 2244 policy_none.go:49] "None policy: Start" Feb 13 19:05:56.855529 kubelet[2244]: I0213 19:05:56.855507 2244 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 13 19:05:56.855529 kubelet[2244]: I0213 19:05:56.855539 2244 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:05:56.861596 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:05:56.871809 kubelet[2244]: I0213 19:05:56.871772 2244 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:05:56.872245 kubelet[2244]: E0213 19:05:56.872210 2244 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Feb 13 19:05:56.872894 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:05:56.875544 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:05:56.886832 kubelet[2244]: E0213 19:05:56.886790 2244 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:05:56.890630 kubelet[2244]: I0213 19:05:56.890598 2244 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:05:56.890885 kubelet[2244]: I0213 19:05:56.890839 2244 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:05:56.890980 kubelet[2244]: I0213 19:05:56.890961 2244 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:05:56.892864 kubelet[2244]: E0213 19:05:56.892823 2244 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Feb 13 19:05:56.973129 kubelet[2244]: E0213 19:05:56.973048 2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="400ms" Feb 13 19:05:57.074459 kubelet[2244]: I0213 19:05:57.074329 2244 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:05:57.074684 kubelet[2244]: E0213 19:05:57.074656 2244 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Feb 13 19:05:57.087176 kubelet[2244]: I0213 19:05:57.087132 2244 topology_manager.go:215] "Topology Admit Handler" podUID="ff32d1fa611dab9efd204acbcf41f8ae" podNamespace="kube-system" podName="kube-apiserver-localhost" Feb 13 19:05:57.088596 kubelet[2244]: I0213 19:05:57.088397 2244 topology_manager.go:215] "Topology Admit Handler" podUID="dd3721fb1a67092819e35b40473f4063" podNamespace="kube-system" podName="kube-controller-manager-localhost" Feb 13 19:05:57.089319 kubelet[2244]: I0213 19:05:57.089282 2244 topology_manager.go:215] "Topology Admit Handler" podUID="8d610d6c43052dbc8df47eb68906a982" podNamespace="kube-system" podName="kube-scheduler-localhost" Feb 13 19:05:57.095300 systemd[1]: Created slice kubepods-burstable-podff32d1fa611dab9efd204acbcf41f8ae.slice - libcontainer container kubepods-burstable-podff32d1fa611dab9efd204acbcf41f8ae.slice. Feb 13 19:05:57.108886 systemd[1]: Created slice kubepods-burstable-poddd3721fb1a67092819e35b40473f4063.slice - libcontainer container kubepods-burstable-poddd3721fb1a67092819e35b40473f4063.slice. Feb 13 19:05:57.126259 systemd[1]: Created slice kubepods-burstable-pod8d610d6c43052dbc8df47eb68906a982.slice - libcontainer container kubepods-burstable-pod8d610d6c43052dbc8df47eb68906a982.slice. Feb 13 19:05:57.172908 kubelet[2244]: I0213 19:05:57.172873 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:05:57.173212 kubelet[2244]: I0213 19:05:57.173072 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:05:57.173212 kubelet[2244]: I0213 19:05:57.173105 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:05:57.173212 kubelet[2244]: I0213 19:05:57.173121 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:05:57.173212 kubelet[2244]: I0213 19:05:57.173140 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:05:57.173212 kubelet[2244]: I0213 19:05:57.173157 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:05:57.173384 kubelet[2244]: I0213 19:05:57.173175 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:05:57.173384 kubelet[2244]: I0213 19:05:57.173223 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:05:57.173384 kubelet[2244]: I0213 19:05:57.173273 2244 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8d610d6c43052dbc8df47eb68906a982-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"8d610d6c43052dbc8df47eb68906a982\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:05:57.374032 kubelet[2244]: E0213 19:05:57.373893 2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="800ms" Feb 13 19:05:57.406363 kubelet[2244]: E0213 19:05:57.406326 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:57.407149 containerd[1488]: time="2025-02-13T19:05:57.407114898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:ff32d1fa611dab9efd204acbcf41f8ae,Namespace:kube-system,Attempt:0,}" Feb 13 19:05:57.424319 kubelet[2244]: E0213 19:05:57.424236 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:57.424717 containerd[1488]: time="2025-02-13T19:05:57.424682790Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:dd3721fb1a67092819e35b40473f4063,Namespace:kube-system,Attempt:0,}" Feb 13 19:05:57.429046 kubelet[2244]: E0213 19:05:57.428966 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:57.429407 containerd[1488]: time="2025-02-13T19:05:57.429374415Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:8d610d6c43052dbc8df47eb68906a982,Namespace:kube-system,Attempt:0,}" Feb 13 19:05:57.476744 kubelet[2244]: I0213 19:05:57.476712 2244 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:05:57.477079 kubelet[2244]: E0213 19:05:57.477042 2244 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Feb 13 19:05:57.835930 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4256824694.mount: Deactivated successfully. Feb 13 19:05:57.839246 containerd[1488]: time="2025-02-13T19:05:57.839117660Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:05:57.840670 containerd[1488]: time="2025-02-13T19:05:57.840586226Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175" Feb 13 19:05:57.843238 containerd[1488]: time="2025-02-13T19:05:57.843171669Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:05:57.844761 containerd[1488]: time="2025-02-13T19:05:57.844732030Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:05:57.845716 containerd[1488]: time="2025-02-13T19:05:57.845668271Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:05:57.846585 containerd[1488]: time="2025-02-13T19:05:57.846147012Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:05:57.846989 containerd[1488]: time="2025-02-13T19:05:57.846917790Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:05:57.850536 containerd[1488]: time="2025-02-13T19:05:57.850502401Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:05:57.851662 containerd[1488]: time="2025-02-13T19:05:57.851417719Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 426.651479ms" Feb 13 19:05:57.852469 containerd[1488]: time="2025-02-13T19:05:57.852435015Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 422.990445ms" Feb 13 19:05:57.853043 containerd[1488]: time="2025-02-13T19:05:57.852802476Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 445.609837ms" Feb 13 19:05:57.889144 kubelet[2244]: W0213 19:05:57.888972 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:57.889144 kubelet[2244]: E0213 19:05:57.889121 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:57.968787 kubelet[2244]: W0213 19:05:57.968700 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:57.968787 kubelet[2244]: E0213 19:05:57.968765 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:58.037351 containerd[1488]: time="2025-02-13T19:05:58.037220966Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:05:58.037351 containerd[1488]: time="2025-02-13T19:05:58.037299323Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:05:58.037351 containerd[1488]: time="2025-02-13T19:05:58.037315578Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.037520 containerd[1488]: time="2025-02-13T19:05:58.037410948Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.038620 containerd[1488]: time="2025-02-13T19:05:58.038445245Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:05:58.038620 containerd[1488]: time="2025-02-13T19:05:58.038509944Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:05:58.038620 containerd[1488]: time="2025-02-13T19:05:58.038527915Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.039579 containerd[1488]: time="2025-02-13T19:05:58.039513449Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.040154 containerd[1488]: time="2025-02-13T19:05:58.040083235Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:05:58.040239 containerd[1488]: time="2025-02-13T19:05:58.040129363Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:05:58.040239 containerd[1488]: time="2025-02-13T19:05:58.040141064Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.040663 containerd[1488]: time="2025-02-13T19:05:58.040219981Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:05:58.063785 systemd[1]: Started cri-containerd-3b5d9025ab6a3d6c3ecfa6e89ffbcbb1a2a3d187e2558ec6ee6925d5ffde594e.scope - libcontainer container 3b5d9025ab6a3d6c3ecfa6e89ffbcbb1a2a3d187e2558ec6ee6925d5ffde594e. Feb 13 19:05:58.065235 systemd[1]: Started cri-containerd-c92bdfccb27e7906a18d8fa94b81d6c436e1c51a593e38e025c061561cf7eee7.scope - libcontainer container c92bdfccb27e7906a18d8fa94b81d6c436e1c51a593e38e025c061561cf7eee7. Feb 13 19:05:58.066889 systemd[1]: Started cri-containerd-f55bdf8689d264b38a2079dd0d715c73fd06c66c72b857434be99fde841e9ebf.scope - libcontainer container f55bdf8689d264b38a2079dd0d715c73fd06c66c72b857434be99fde841e9ebf. Feb 13 19:05:58.067262 kubelet[2244]: W0213 19:05:58.067213 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:58.067324 kubelet[2244]: E0213 19:05:58.067274 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:58.101069 containerd[1488]: time="2025-02-13T19:05:58.100833560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:8d610d6c43052dbc8df47eb68906a982,Namespace:kube-system,Attempt:0,} returns sandbox id \"3b5d9025ab6a3d6c3ecfa6e89ffbcbb1a2a3d187e2558ec6ee6925d5ffde594e\"" Feb 13 19:05:58.102662 containerd[1488]: time="2025-02-13T19:05:58.101641453Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:ff32d1fa611dab9efd204acbcf41f8ae,Namespace:kube-system,Attempt:0,} returns sandbox id \"c92bdfccb27e7906a18d8fa94b81d6c436e1c51a593e38e025c061561cf7eee7\"" Feb 13 19:05:58.103746 kubelet[2244]: E0213 19:05:58.103673 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:58.104005 kubelet[2244]: E0213 19:05:58.103920 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:58.106916 containerd[1488]: time="2025-02-13T19:05:58.106621359Z" level=info msg="CreateContainer within sandbox \"3b5d9025ab6a3d6c3ecfa6e89ffbcbb1a2a3d187e2558ec6ee6925d5ffde594e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:05:58.107598 containerd[1488]: time="2025-02-13T19:05:58.107572667Z" level=info msg="CreateContainer within sandbox \"c92bdfccb27e7906a18d8fa94b81d6c436e1c51a593e38e025c061561cf7eee7\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:05:58.108908 containerd[1488]: time="2025-02-13T19:05:58.108863322Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:dd3721fb1a67092819e35b40473f4063,Namespace:kube-system,Attempt:0,} returns sandbox id \"f55bdf8689d264b38a2079dd0d715c73fd06c66c72b857434be99fde841e9ebf\"" Feb 13 19:05:58.109967 kubelet[2244]: E0213 19:05:58.109651 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:58.111543 containerd[1488]: time="2025-02-13T19:05:58.111508532Z" level=info msg="CreateContainer within sandbox \"f55bdf8689d264b38a2079dd0d715c73fd06c66c72b857434be99fde841e9ebf\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:05:58.123355 containerd[1488]: time="2025-02-13T19:05:58.123307060Z" level=info msg="CreateContainer within sandbox \"3b5d9025ab6a3d6c3ecfa6e89ffbcbb1a2a3d187e2558ec6ee6925d5ffde594e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"8619169766527ff282abe63da3b04e7c23c59e4b028e249640acdc7af1a8c503\"" Feb 13 19:05:58.123986 containerd[1488]: time="2025-02-13T19:05:58.123954964Z" level=info msg="StartContainer for \"8619169766527ff282abe63da3b04e7c23c59e4b028e249640acdc7af1a8c503\"" Feb 13 19:05:58.130451 containerd[1488]: time="2025-02-13T19:05:58.130408878Z" level=info msg="CreateContainer within sandbox \"c92bdfccb27e7906a18d8fa94b81d6c436e1c51a593e38e025c061561cf7eee7\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"507cf3e958cba7c858bbeef5aaebebbad4f60e33111f777a0d79ca5fd058ef70\"" Feb 13 19:05:58.130955 containerd[1488]: time="2025-02-13T19:05:58.130915043Z" level=info msg="StartContainer for \"507cf3e958cba7c858bbeef5aaebebbad4f60e33111f777a0d79ca5fd058ef70\"" Feb 13 19:05:58.134478 containerd[1488]: time="2025-02-13T19:05:58.134402132Z" level=info msg="CreateContainer within sandbox \"f55bdf8689d264b38a2079dd0d715c73fd06c66c72b857434be99fde841e9ebf\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"4bbf38aaff258434603b512352a459ed7847bb29ba84744668b9e1e5f0a4c683\"" Feb 13 19:05:58.136014 containerd[1488]: time="2025-02-13T19:05:58.134846835Z" level=info msg="StartContainer for \"4bbf38aaff258434603b512352a459ed7847bb29ba84744668b9e1e5f0a4c683\"" Feb 13 19:05:58.149758 systemd[1]: Started cri-containerd-8619169766527ff282abe63da3b04e7c23c59e4b028e249640acdc7af1a8c503.scope - libcontainer container 8619169766527ff282abe63da3b04e7c23c59e4b028e249640acdc7af1a8c503. Feb 13 19:05:58.153947 systemd[1]: Started cri-containerd-507cf3e958cba7c858bbeef5aaebebbad4f60e33111f777a0d79ca5fd058ef70.scope - libcontainer container 507cf3e958cba7c858bbeef5aaebebbad4f60e33111f777a0d79ca5fd058ef70. Feb 13 19:05:58.169731 systemd[1]: Started cri-containerd-4bbf38aaff258434603b512352a459ed7847bb29ba84744668b9e1e5f0a4c683.scope - libcontainer container 4bbf38aaff258434603b512352a459ed7847bb29ba84744668b9e1e5f0a4c683. Feb 13 19:05:58.175120 kubelet[2244]: E0213 19:05:58.175076 2244 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="1.6s" Feb 13 19:05:58.201949 containerd[1488]: time="2025-02-13T19:05:58.201869918Z" level=info msg="StartContainer for \"507cf3e958cba7c858bbeef5aaebebbad4f60e33111f777a0d79ca5fd058ef70\" returns successfully" Feb 13 19:05:58.202196 containerd[1488]: time="2025-02-13T19:05:58.201923034Z" level=info msg="StartContainer for \"8619169766527ff282abe63da3b04e7c23c59e4b028e249640acdc7af1a8c503\" returns successfully" Feb 13 19:05:58.232903 containerd[1488]: time="2025-02-13T19:05:58.232838090Z" level=info msg="StartContainer for \"4bbf38aaff258434603b512352a459ed7847bb29ba84744668b9e1e5f0a4c683\" returns successfully" Feb 13 19:05:58.282639 kubelet[2244]: I0213 19:05:58.281067 2244 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:05:58.282639 kubelet[2244]: E0213 19:05:58.281394 2244 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Feb 13 19:05:58.353259 kubelet[2244]: W0213 19:05:58.352268 2244 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:58.353259 kubelet[2244]: E0213 19:05:58.352329 2244 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Feb 13 19:05:58.797759 kubelet[2244]: E0213 19:05:58.797654 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:58.801590 kubelet[2244]: E0213 19:05:58.799439 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:58.801590 kubelet[2244]: E0213 19:05:58.800974 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:59.807097 kubelet[2244]: E0213 19:05:59.807055 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:05:59.822289 kubelet[2244]: E0213 19:05:59.822255 2244 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Feb 13 19:05:59.882731 kubelet[2244]: I0213 19:05:59.882695 2244 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:06:00.003123 kubelet[2244]: I0213 19:06:00.003080 2244 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Feb 13 19:06:00.015066 kubelet[2244]: E0213 19:06:00.015028 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.116244 kubelet[2244]: E0213 19:06:00.116106 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.216732 kubelet[2244]: E0213 19:06:00.216688 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.317648 kubelet[2244]: E0213 19:06:00.317602 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.418291 kubelet[2244]: E0213 19:06:00.418176 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.518811 kubelet[2244]: E0213 19:06:00.518744 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.619300 kubelet[2244]: E0213 19:06:00.619254 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.719965 kubelet[2244]: E0213 19:06:00.719862 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:00.820551 kubelet[2244]: E0213 19:06:00.820510 2244 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:06:01.528611 kubelet[2244]: E0213 19:06:01.528557 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:01.763709 kubelet[2244]: I0213 19:06:01.763664 2244 apiserver.go:52] "Watching apiserver" Feb 13 19:06:01.771431 kubelet[2244]: I0213 19:06:01.771393 2244 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:06:01.808330 kubelet[2244]: E0213 19:06:01.808089 2244 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:02.185353 systemd[1]: Reload requested from client PID 2520 ('systemctl') (unit session-5.scope)... Feb 13 19:06:02.185368 systemd[1]: Reloading... Feb 13 19:06:02.273591 zram_generator::config[2567]: No configuration found. Feb 13 19:06:02.374128 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:06:02.464509 systemd[1]: Reloading finished in 278 ms. Feb 13 19:06:02.487893 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:06:02.503390 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:06:02.503718 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:06:02.503773 systemd[1]: kubelet.service: Consumed 1.886s CPU time, 112.7M memory peak. Feb 13 19:06:02.514800 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:06:02.621488 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:06:02.625497 (kubelet)[2606]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:06:02.666023 kubelet[2606]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:06:02.666023 kubelet[2606]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 13 19:06:02.666023 kubelet[2606]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:06:02.666433 kubelet[2606]: I0213 19:06:02.666062 2606 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:06:02.671711 kubelet[2606]: I0213 19:06:02.671673 2606 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Feb 13 19:06:02.671711 kubelet[2606]: I0213 19:06:02.671704 2606 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:06:02.672485 kubelet[2606]: I0213 19:06:02.671881 2606 server.go:927] "Client rotation is on, will bootstrap in background" Feb 13 19:06:02.675638 kubelet[2606]: I0213 19:06:02.675606 2606 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:06:02.677274 kubelet[2606]: I0213 19:06:02.677237 2606 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:06:02.685225 kubelet[2606]: I0213 19:06:02.685191 2606 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:06:02.685422 kubelet[2606]: I0213 19:06:02.685389 2606 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:06:02.685634 kubelet[2606]: I0213 19:06:02.685415 2606 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Feb 13 19:06:02.685712 kubelet[2606]: I0213 19:06:02.685635 2606 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:06:02.685712 kubelet[2606]: I0213 19:06:02.685645 2606 container_manager_linux.go:301] "Creating device plugin manager" Feb 13 19:06:02.685712 kubelet[2606]: I0213 19:06:02.685683 2606 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:06:02.685798 kubelet[2606]: I0213 19:06:02.685787 2606 kubelet.go:400] "Attempting to sync node with API server" Feb 13 19:06:02.685825 kubelet[2606]: I0213 19:06:02.685799 2606 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:06:02.685851 kubelet[2606]: I0213 19:06:02.685826 2606 kubelet.go:312] "Adding apiserver pod source" Feb 13 19:06:02.685851 kubelet[2606]: I0213 19:06:02.685841 2606 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:06:02.686541 kubelet[2606]: I0213 19:06:02.686307 2606 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:06:02.686541 kubelet[2606]: I0213 19:06:02.686498 2606 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:06:02.687985 kubelet[2606]: I0213 19:06:02.687964 2606 server.go:1264] "Started kubelet" Feb 13 19:06:02.688454 kubelet[2606]: I0213 19:06:02.688312 2606 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:06:02.688762 kubelet[2606]: I0213 19:06:02.688743 2606 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:06:02.689577 kubelet[2606]: I0213 19:06:02.689212 2606 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:06:02.690229 kubelet[2606]: I0213 19:06:02.690211 2606 volume_manager.go:291] "Starting Kubelet Volume Manager" Feb 13 19:06:02.690646 kubelet[2606]: I0213 19:06:02.690286 2606 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:06:02.694393 kubelet[2606]: I0213 19:06:02.694354 2606 server.go:455] "Adding debug handlers to kubelet server" Feb 13 19:06:02.695115 kubelet[2606]: I0213 19:06:02.695090 2606 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:06:02.695255 kubelet[2606]: I0213 19:06:02.695237 2606 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:06:02.706840 kubelet[2606]: E0213 19:06:02.706503 2606 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:06:02.706840 kubelet[2606]: I0213 19:06:02.706785 2606 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:06:02.709880 kubelet[2606]: I0213 19:06:02.708713 2606 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:06:02.710378 kubelet[2606]: I0213 19:06:02.710333 2606 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:06:02.710428 kubelet[2606]: I0213 19:06:02.710381 2606 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 13 19:06:02.710428 kubelet[2606]: I0213 19:06:02.710399 2606 kubelet.go:2337] "Starting kubelet main sync loop" Feb 13 19:06:02.710470 kubelet[2606]: E0213 19:06:02.710446 2606 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:06:02.712908 kubelet[2606]: I0213 19:06:02.712768 2606 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:06:02.712908 kubelet[2606]: I0213 19:06:02.712793 2606 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:06:02.743874 kubelet[2606]: I0213 19:06:02.743783 2606 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 13 19:06:02.744419 kubelet[2606]: I0213 19:06:02.744267 2606 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 13 19:06:02.744419 kubelet[2606]: I0213 19:06:02.744302 2606 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:06:02.745141 kubelet[2606]: I0213 19:06:02.744467 2606 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:06:02.745141 kubelet[2606]: I0213 19:06:02.744478 2606 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:06:02.745141 kubelet[2606]: I0213 19:06:02.744495 2606 policy_none.go:49] "None policy: Start" Feb 13 19:06:02.745141 kubelet[2606]: I0213 19:06:02.745094 2606 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 13 19:06:02.745141 kubelet[2606]: I0213 19:06:02.745117 2606 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:06:02.745256 kubelet[2606]: I0213 19:06:02.745239 2606 state_mem.go:75] "Updated machine memory state" Feb 13 19:06:02.749369 kubelet[2606]: I0213 19:06:02.749339 2606 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:06:02.749596 kubelet[2606]: I0213 19:06:02.749542 2606 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:06:02.749678 kubelet[2606]: I0213 19:06:02.749660 2606 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:06:02.806896 kubelet[2606]: I0213 19:06:02.806863 2606 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Feb 13 19:06:02.810796 kubelet[2606]: I0213 19:06:02.810741 2606 topology_manager.go:215] "Topology Admit Handler" podUID="ff32d1fa611dab9efd204acbcf41f8ae" podNamespace="kube-system" podName="kube-apiserver-localhost" Feb 13 19:06:02.810870 kubelet[2606]: I0213 19:06:02.810861 2606 topology_manager.go:215] "Topology Admit Handler" podUID="dd3721fb1a67092819e35b40473f4063" podNamespace="kube-system" podName="kube-controller-manager-localhost" Feb 13 19:06:02.811023 kubelet[2606]: I0213 19:06:02.810899 2606 topology_manager.go:215] "Topology Admit Handler" podUID="8d610d6c43052dbc8df47eb68906a982" podNamespace="kube-system" podName="kube-scheduler-localhost" Feb 13 19:06:02.835807 kubelet[2606]: E0213 19:06:02.835772 2606 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Feb 13 19:06:02.838852 kubelet[2606]: I0213 19:06:02.838698 2606 kubelet_node_status.go:112] "Node was previously registered" node="localhost" Feb 13 19:06:02.838852 kubelet[2606]: I0213 19:06:02.838778 2606 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Feb 13 19:06:02.896274 kubelet[2606]: I0213 19:06:02.896227 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:06:02.896274 kubelet[2606]: I0213 19:06:02.896262 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:02.896274 kubelet[2606]: I0213 19:06:02.896285 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:06:02.896610 kubelet[2606]: I0213 19:06:02.896300 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ff32d1fa611dab9efd204acbcf41f8ae-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"ff32d1fa611dab9efd204acbcf41f8ae\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:06:02.896610 kubelet[2606]: I0213 19:06:02.896320 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:02.896610 kubelet[2606]: I0213 19:06:02.896339 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8d610d6c43052dbc8df47eb68906a982-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"8d610d6c43052dbc8df47eb68906a982\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:06:02.896610 kubelet[2606]: I0213 19:06:02.896353 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:02.896610 kubelet[2606]: I0213 19:06:02.896372 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:02.896722 kubelet[2606]: I0213 19:06:02.896392 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dd3721fb1a67092819e35b40473f4063-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd3721fb1a67092819e35b40473f4063\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:03.131355 kubelet[2606]: E0213 19:06:03.131033 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.131355 kubelet[2606]: E0213 19:06:03.131163 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.136905 kubelet[2606]: E0213 19:06:03.136843 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.686440 kubelet[2606]: I0213 19:06:03.686389 2606 apiserver.go:52] "Watching apiserver" Feb 13 19:06:03.696056 kubelet[2606]: I0213 19:06:03.695987 2606 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:06:03.725034 kubelet[2606]: E0213 19:06:03.722886 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.729791 kubelet[2606]: E0213 19:06:03.729763 2606 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Feb 13 19:06:03.730026 kubelet[2606]: E0213 19:06:03.729926 2606 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Feb 13 19:06:03.730487 kubelet[2606]: E0213 19:06:03.730471 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.731126 kubelet[2606]: E0213 19:06:03.731031 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:03.753155 kubelet[2606]: I0213 19:06:03.753083 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.753067378 podStartE2EDuration="1.753067378s" podCreationTimestamp="2025-02-13 19:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:03.740256839 +0000 UTC m=+1.111713818" watchObservedRunningTime="2025-02-13 19:06:03.753067378 +0000 UTC m=+1.124524277" Feb 13 19:06:03.760806 kubelet[2606]: I0213 19:06:03.760740 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.7607251 podStartE2EDuration="2.7607251s" podCreationTimestamp="2025-02-13 19:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:03.753380206 +0000 UTC m=+1.124837145" watchObservedRunningTime="2025-02-13 19:06:03.7607251 +0000 UTC m=+1.132182039" Feb 13 19:06:03.768060 kubelet[2606]: I0213 19:06:03.767911 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.767897213 podStartE2EDuration="1.767897213s" podCreationTimestamp="2025-02-13 19:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:03.760680975 +0000 UTC m=+1.132137954" watchObservedRunningTime="2025-02-13 19:06:03.767897213 +0000 UTC m=+1.139354152" Feb 13 19:06:03.988403 sudo[1629]: pam_unix(sudo:session): session closed for user root Feb 13 19:06:03.989641 sshd[1628]: Connection closed by 10.0.0.1 port 59776 Feb 13 19:06:03.990066 sshd-session[1625]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:03.992998 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:06:03.993203 systemd[1]: session-5.scope: Consumed 7.325s CPU time, 254.1M memory peak. Feb 13 19:06:03.995194 systemd[1]: sshd@4-10.0.0.60:22-10.0.0.1:59776.service: Deactivated successfully. Feb 13 19:06:03.997847 systemd-logind[1469]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:06:03.998835 systemd-logind[1469]: Removed session 5. Feb 13 19:06:04.724144 kubelet[2606]: E0213 19:06:04.724103 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:04.724144 kubelet[2606]: E0213 19:06:04.724131 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:06.297991 kubelet[2606]: E0213 19:06:06.297950 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:08.409438 kubelet[2606]: E0213 19:06:08.409314 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:08.562748 kubelet[2606]: E0213 19:06:08.562715 2606 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:06:16.837334 kubelet[2606]: I0213 19:06:16.837285 2606 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:06:16.837884 containerd[1488]: time="2025-02-13T19:06:16.837727715Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:06:16.838087 kubelet[2606]: I0213 19:06:16.837896 2606 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:06:17.579824 kubelet[2606]: I0213 19:06:17.579439 2606 topology_manager.go:215] "Topology Admit Handler" podUID="12162f3d-0ddc-48c1-8f4d-b710e8a864e9" podNamespace="kube-system" podName="kube-proxy-6mldx" Feb 13 19:06:17.583062 kubelet[2606]: I0213 19:06:17.583012 2606 topology_manager.go:215] "Topology Admit Handler" podUID="97fdc299-96ed-44a6-933b-4cf8196e7a62" podNamespace="kube-flannel" podName="kube-flannel-ds-56xl5" Feb 13 19:06:17.592166 systemd[1]: Created slice kubepods-besteffort-pod12162f3d_0ddc_48c1_8f4d_b710e8a864e9.slice - libcontainer container kubepods-besteffort-pod12162f3d_0ddc_48c1_8f4d_b710e8a864e9.slice. Feb 13 19:06:17.595773 kubelet[2606]: I0213 19:06:17.595338 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/12162f3d-0ddc-48c1-8f4d-b710e8a864e9-lib-modules\") pod \"kube-proxy-6mldx\" (UID: \"12162f3d-0ddc-48c1-8f4d-b710e8a864e9\") " pod="kube-system/kube-proxy-6mldx" Feb 13 19:06:17.595773 kubelet[2606]: I0213 19:06:17.595377 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/97fdc299-96ed-44a6-933b-4cf8196e7a62-cni\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.595773 kubelet[2606]: I0213 19:06:17.595397 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/97fdc299-96ed-44a6-933b-4cf8196e7a62-run\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.595773 kubelet[2606]: I0213 19:06:17.595416 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/12162f3d-0ddc-48c1-8f4d-b710e8a864e9-kube-proxy\") pod \"kube-proxy-6mldx\" (UID: \"12162f3d-0ddc-48c1-8f4d-b710e8a864e9\") " pod="kube-system/kube-proxy-6mldx" Feb 13 19:06:17.595773 kubelet[2606]: I0213 19:06:17.595442 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/97fdc299-96ed-44a6-933b-4cf8196e7a62-flannel-cfg\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.595965 kubelet[2606]: I0213 19:06:17.595459 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/97fdc299-96ed-44a6-933b-4cf8196e7a62-xtables-lock\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.595965 kubelet[2606]: I0213 19:06:17.595474 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crmjd\" (UniqueName: \"kubernetes.io/projected/97fdc299-96ed-44a6-933b-4cf8196e7a62-kube-api-access-crmjd\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.595965 kubelet[2606]: I0213 19:06:17.595494 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/12162f3d-0ddc-48c1-8f4d-b710e8a864e9-xtables-lock\") pod \"kube-proxy-6mldx\" (UID: \"12162f3d-0ddc-48c1-8f4d-b710e8a864e9\") " pod="kube-system/kube-proxy-6mldx" Feb 13 19:06:17.595965 kubelet[2606]: I0213 19:06:17.595520 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9hr6\" (UniqueName: \"kubernetes.io/projected/12162f3d-0ddc-48c1-8f4d-b710e8a864e9-kube-api-access-m9hr6\") pod \"kube-proxy-6mldx\" (UID: \"12162f3d-0ddc-48c1-8f4d-b710e8a864e9\") " pod="kube-system/kube-proxy-6mldx" Feb 13 19:06:17.595965 kubelet[2606]: I0213 19:06:17.595544 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/97fdc299-96ed-44a6-933b-4cf8196e7a62-cni-plugin\") pod \"kube-flannel-ds-56xl5\" (UID: \"97fdc299-96ed-44a6-933b-4cf8196e7a62\") " pod="kube-flannel/kube-flannel-ds-56xl5" Feb 13 19:06:17.609353 systemd[1]: Created slice kubepods-burstable-pod97fdc299_96ed_44a6_933b_4cf8196e7a62.slice - libcontainer container kubepods-burstable-pod97fdc299_96ed_44a6_933b_4cf8196e7a62.slice. Feb 13 19:06:17.907043 containerd[1488]: time="2025-02-13T19:06:17.907000196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6mldx,Uid:12162f3d-0ddc-48c1-8f4d-b710e8a864e9,Namespace:kube-system,Attempt:0,}" Feb 13 19:06:17.913267 containerd[1488]: time="2025-02-13T19:06:17.913229438Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-56xl5,Uid:97fdc299-96ed-44a6-933b-4cf8196e7a62,Namespace:kube-flannel,Attempt:0,}" Feb 13 19:06:17.929425 containerd[1488]: time="2025-02-13T19:06:17.929202825Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:06:17.929425 containerd[1488]: time="2025-02-13T19:06:17.929265235Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:06:17.929425 containerd[1488]: time="2025-02-13T19:06:17.929277877Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:17.929425 containerd[1488]: time="2025-02-13T19:06:17.929350008Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:17.942003 containerd[1488]: time="2025-02-13T19:06:17.941887744Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:06:17.943034 containerd[1488]: time="2025-02-13T19:06:17.942758239Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:06:17.943034 containerd[1488]: time="2025-02-13T19:06:17.942806846Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:17.943034 containerd[1488]: time="2025-02-13T19:06:17.942899341Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:17.956764 systemd[1]: Started cri-containerd-677beeb488bbd511ee18748c43103cf51f51a1d99cd4ba5a7df6bd1914338299.scope - libcontainer container 677beeb488bbd511ee18748c43103cf51f51a1d99cd4ba5a7df6bd1914338299. Feb 13 19:06:17.960493 systemd[1]: Started cri-containerd-2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb.scope - libcontainer container 2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb. Feb 13 19:06:17.977725 containerd[1488]: time="2025-02-13T19:06:17.977683072Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6mldx,Uid:12162f3d-0ddc-48c1-8f4d-b710e8a864e9,Namespace:kube-system,Attempt:0,} returns sandbox id \"677beeb488bbd511ee18748c43103cf51f51a1d99cd4ba5a7df6bd1914338299\"" Feb 13 19:06:17.981514 containerd[1488]: time="2025-02-13T19:06:17.981480739Z" level=info msg="CreateContainer within sandbox \"677beeb488bbd511ee18748c43103cf51f51a1d99cd4ba5a7df6bd1914338299\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:06:17.995809 containerd[1488]: time="2025-02-13T19:06:17.995710777Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-56xl5,Uid:97fdc299-96ed-44a6-933b-4cf8196e7a62,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\"" Feb 13 19:06:17.997618 containerd[1488]: time="2025-02-13T19:06:17.997590867Z" level=info msg="CreateContainer within sandbox \"677beeb488bbd511ee18748c43103cf51f51a1d99cd4ba5a7df6bd1914338299\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"635c5811d2cae59c6ccedf183b32b30d49470dcc568a7b8cdda72e702f332cba\"" Feb 13 19:06:18.000100 containerd[1488]: time="2025-02-13T19:06:18.000070450Z" level=info msg="StartContainer for \"635c5811d2cae59c6ccedf183b32b30d49470dcc568a7b8cdda72e702f332cba\"" Feb 13 19:06:18.003505 containerd[1488]: time="2025-02-13T19:06:18.003430507Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Feb 13 19:06:18.025712 systemd[1]: Started cri-containerd-635c5811d2cae59c6ccedf183b32b30d49470dcc568a7b8cdda72e702f332cba.scope - libcontainer container 635c5811d2cae59c6ccedf183b32b30d49470dcc568a7b8cdda72e702f332cba. Feb 13 19:06:18.050125 containerd[1488]: time="2025-02-13T19:06:18.050086753Z" level=info msg="StartContainer for \"635c5811d2cae59c6ccedf183b32b30d49470dcc568a7b8cdda72e702f332cba\" returns successfully" Feb 13 19:06:18.097515 update_engine[1473]: I20250213 19:06:18.097429 1473 update_attempter.cc:509] Updating boot flags... Feb 13 19:06:18.131376 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2801) Feb 13 19:06:18.180608 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2805) Feb 13 19:06:18.223578 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2805) Feb 13 19:06:18.755686 kubelet[2606]: I0213 19:06:18.755622 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-6mldx" podStartSLOduration=1.755592922 podStartE2EDuration="1.755592922s" podCreationTimestamp="2025-02-13 19:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:18.755452741 +0000 UTC m=+16.126909680" watchObservedRunningTime="2025-02-13 19:06:18.755592922 +0000 UTC m=+16.127049901" Feb 13 19:06:19.324185 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3370281704.mount: Deactivated successfully. Feb 13 19:06:19.355836 containerd[1488]: time="2025-02-13T19:06:19.355781074Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:19.356292 containerd[1488]: time="2025-02-13T19:06:19.356237457Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3673532" Feb 13 19:06:19.357461 containerd[1488]: time="2025-02-13T19:06:19.357412021Z" level=info msg="ImageCreate event name:\"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:19.359760 containerd[1488]: time="2025-02-13T19:06:19.359701541Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:19.360518 containerd[1488]: time="2025-02-13T19:06:19.360474048Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3662650\" in 1.357012177s" Feb 13 19:06:19.360518 containerd[1488]: time="2025-02-13T19:06:19.360507293Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\"" Feb 13 19:06:19.362843 containerd[1488]: time="2025-02-13T19:06:19.362731683Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Feb 13 19:06:19.373432 containerd[1488]: time="2025-02-13T19:06:19.373371368Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709\"" Feb 13 19:06:19.373860 containerd[1488]: time="2025-02-13T19:06:19.373833072Z" level=info msg="StartContainer for \"d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709\"" Feb 13 19:06:19.402773 systemd[1]: Started cri-containerd-d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709.scope - libcontainer container d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709. Feb 13 19:06:19.424893 containerd[1488]: time="2025-02-13T19:06:19.424847230Z" level=info msg="StartContainer for \"d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709\" returns successfully" Feb 13 19:06:19.426476 systemd[1]: cri-containerd-d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709.scope: Deactivated successfully. Feb 13 19:06:19.471973 containerd[1488]: time="2025-02-13T19:06:19.471912717Z" level=info msg="shim disconnected" id=d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709 namespace=k8s.io Feb 13 19:06:19.471973 containerd[1488]: time="2025-02-13T19:06:19.471967085Z" level=warning msg="cleaning up after shim disconnected" id=d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709 namespace=k8s.io Feb 13 19:06:19.471973 containerd[1488]: time="2025-02-13T19:06:19.471976966Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:06:19.707197 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d70c660a39b25a4f40fd2cee3e4a2664f13bc6c3cd3133083a2d42498269c709-rootfs.mount: Deactivated successfully. Feb 13 19:06:19.750952 containerd[1488]: time="2025-02-13T19:06:19.750917607Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Feb 13 19:06:21.134244 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount494365126.mount: Deactivated successfully. Feb 13 19:06:22.420395 containerd[1488]: time="2025-02-13T19:06:22.420342250Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:22.420898 containerd[1488]: time="2025-02-13T19:06:22.420824148Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26874261" Feb 13 19:06:22.423239 containerd[1488]: time="2025-02-13T19:06:22.423182752Z" level=info msg="ImageCreate event name:\"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:22.426094 containerd[1488]: time="2025-02-13T19:06:22.426047537Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:06:22.427537 containerd[1488]: time="2025-02-13T19:06:22.427411862Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26863435\" in 2.67645865s" Feb 13 19:06:22.427537 containerd[1488]: time="2025-02-13T19:06:22.427447746Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\"" Feb 13 19:06:22.437125 containerd[1488]: time="2025-02-13T19:06:22.437086547Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:06:22.453761 containerd[1488]: time="2025-02-13T19:06:22.453720351Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e\"" Feb 13 19:06:22.454472 containerd[1488]: time="2025-02-13T19:06:22.454442518Z" level=info msg="StartContainer for \"4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e\"" Feb 13 19:06:22.483717 systemd[1]: Started cri-containerd-4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e.scope - libcontainer container 4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e. Feb 13 19:06:22.502755 containerd[1488]: time="2025-02-13T19:06:22.502651287Z" level=info msg="StartContainer for \"4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e\" returns successfully" Feb 13 19:06:22.508651 systemd[1]: cri-containerd-4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e.scope: Deactivated successfully. Feb 13 19:06:22.564581 containerd[1488]: time="2025-02-13T19:06:22.564471415Z" level=info msg="shim disconnected" id=4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e namespace=k8s.io Feb 13 19:06:22.564581 containerd[1488]: time="2025-02-13T19:06:22.564540463Z" level=warning msg="cleaning up after shim disconnected" id=4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e namespace=k8s.io Feb 13 19:06:22.564581 containerd[1488]: time="2025-02-13T19:06:22.564551825Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:06:22.592826 kubelet[2606]: I0213 19:06:22.592791 2606 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Feb 13 19:06:22.615550 kubelet[2606]: I0213 19:06:22.615512 2606 topology_manager.go:215] "Topology Admit Handler" podUID="0cc7b41b-4248-4e19-96b2-3f4bffb88b90" podNamespace="kube-system" podName="coredns-7db6d8ff4d-hdmcl" Feb 13 19:06:22.616606 kubelet[2606]: I0213 19:06:22.616148 2606 topology_manager.go:215] "Topology Admit Handler" podUID="ad8eda04-e7ce-469d-b354-eba78e8e6145" podNamespace="kube-system" podName="coredns-7db6d8ff4d-5kvnd" Feb 13 19:06:22.623857 systemd[1]: Created slice kubepods-burstable-pod0cc7b41b_4248_4e19_96b2_3f4bffb88b90.slice - libcontainer container kubepods-burstable-pod0cc7b41b_4248_4e19_96b2_3f4bffb88b90.slice. Feb 13 19:06:22.630331 systemd[1]: Created slice kubepods-burstable-podad8eda04_e7ce_469d_b354_eba78e8e6145.slice - libcontainer container kubepods-burstable-podad8eda04_e7ce_469d_b354_eba78e8e6145.slice. Feb 13 19:06:22.631707 kubelet[2606]: I0213 19:06:22.631672 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsqpc\" (UniqueName: \"kubernetes.io/projected/0cc7b41b-4248-4e19-96b2-3f4bffb88b90-kube-api-access-lsqpc\") pod \"coredns-7db6d8ff4d-hdmcl\" (UID: \"0cc7b41b-4248-4e19-96b2-3f4bffb88b90\") " pod="kube-system/coredns-7db6d8ff4d-hdmcl" Feb 13 19:06:22.631784 kubelet[2606]: I0213 19:06:22.631717 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mldcx\" (UniqueName: \"kubernetes.io/projected/ad8eda04-e7ce-469d-b354-eba78e8e6145-kube-api-access-mldcx\") pod \"coredns-7db6d8ff4d-5kvnd\" (UID: \"ad8eda04-e7ce-469d-b354-eba78e8e6145\") " pod="kube-system/coredns-7db6d8ff4d-5kvnd" Feb 13 19:06:22.631784 kubelet[2606]: I0213 19:06:22.631738 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc7b41b-4248-4e19-96b2-3f4bffb88b90-config-volume\") pod \"coredns-7db6d8ff4d-hdmcl\" (UID: \"0cc7b41b-4248-4e19-96b2-3f4bffb88b90\") " pod="kube-system/coredns-7db6d8ff4d-hdmcl" Feb 13 19:06:22.631784 kubelet[2606]: I0213 19:06:22.631760 2606 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8eda04-e7ce-469d-b354-eba78e8e6145-config-volume\") pod \"coredns-7db6d8ff4d-5kvnd\" (UID: \"ad8eda04-e7ce-469d-b354-eba78e8e6145\") " pod="kube-system/coredns-7db6d8ff4d-5kvnd" Feb 13 19:06:22.757316 containerd[1488]: time="2025-02-13T19:06:22.757042776Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Feb 13 19:06:22.783320 containerd[1488]: time="2025-02-13T19:06:22.783214569Z" level=info msg="CreateContainer within sandbox \"2069c00db242cdf780d21b8a4fc4363fc3291e4340870ce834a4fbceaec8cdbb\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"84ddd869b60c6ef4dc578fe772876d34734fc840b988978656629aca5d723d86\"" Feb 13 19:06:22.783958 containerd[1488]: time="2025-02-13T19:06:22.783895411Z" level=info msg="StartContainer for \"84ddd869b60c6ef4dc578fe772876d34734fc840b988978656629aca5d723d86\"" Feb 13 19:06:22.806743 systemd[1]: Started cri-containerd-84ddd869b60c6ef4dc578fe772876d34734fc840b988978656629aca5d723d86.scope - libcontainer container 84ddd869b60c6ef4dc578fe772876d34734fc840b988978656629aca5d723d86. Feb 13 19:06:22.829980 containerd[1488]: time="2025-02-13T19:06:22.829921117Z" level=info msg="StartContainer for \"84ddd869b60c6ef4dc578fe772876d34734fc840b988978656629aca5d723d86\" returns successfully" Feb 13 19:06:22.929226 containerd[1488]: time="2025-02-13T19:06:22.929151792Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hdmcl,Uid:0cc7b41b-4248-4e19-96b2-3f4bffb88b90,Namespace:kube-system,Attempt:0,}" Feb 13 19:06:22.933932 containerd[1488]: time="2025-02-13T19:06:22.933747106Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5kvnd,Uid:ad8eda04-e7ce-469d-b354-eba78e8e6145,Namespace:kube-system,Attempt:0,}" Feb 13 19:06:23.009407 containerd[1488]: time="2025-02-13T19:06:23.009288524Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hdmcl,Uid:0cc7b41b-4248-4e19-96b2-3f4bffb88b90,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ca0a57f430eb4aaa7aaad347c3dec2606f2b1a47766bb1b75ebf2f72554ac05d\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:06:23.010087 kubelet[2606]: E0213 19:06:23.009775 2606 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca0a57f430eb4aaa7aaad347c3dec2606f2b1a47766bb1b75ebf2f72554ac05d\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:06:23.010087 kubelet[2606]: E0213 19:06:23.009829 2606 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca0a57f430eb4aaa7aaad347c3dec2606f2b1a47766bb1b75ebf2f72554ac05d\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-hdmcl" Feb 13 19:06:23.010087 kubelet[2606]: E0213 19:06:23.009847 2606 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca0a57f430eb4aaa7aaad347c3dec2606f2b1a47766bb1b75ebf2f72554ac05d\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-hdmcl" Feb 13 19:06:23.010087 kubelet[2606]: E0213 19:06:23.009887 2606 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-hdmcl_kube-system(0cc7b41b-4248-4e19-96b2-3f4bffb88b90)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-hdmcl_kube-system(0cc7b41b-4248-4e19-96b2-3f4bffb88b90)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ca0a57f430eb4aaa7aaad347c3dec2606f2b1a47766bb1b75ebf2f72554ac05d\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-7db6d8ff4d-hdmcl" podUID="0cc7b41b-4248-4e19-96b2-3f4bffb88b90" Feb 13 19:06:23.010467 containerd[1488]: time="2025-02-13T19:06:23.010419854Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5kvnd,Uid:ad8eda04-e7ce-469d-b354-eba78e8e6145,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"cd8c3d2960538adf0d6e0277964b67b10a1edb4b2a2a278bd57181deecaa9089\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:06:23.010604 kubelet[2606]: E0213 19:06:23.010582 2606 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd8c3d2960538adf0d6e0277964b67b10a1edb4b2a2a278bd57181deecaa9089\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:06:23.010663 kubelet[2606]: E0213 19:06:23.010617 2606 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd8c3d2960538adf0d6e0277964b67b10a1edb4b2a2a278bd57181deecaa9089\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-5kvnd" Feb 13 19:06:23.010663 kubelet[2606]: E0213 19:06:23.010632 2606 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd8c3d2960538adf0d6e0277964b67b10a1edb4b2a2a278bd57181deecaa9089\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-5kvnd" Feb 13 19:06:23.010711 kubelet[2606]: E0213 19:06:23.010669 2606 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5kvnd_kube-system(ad8eda04-e7ce-469d-b354-eba78e8e6145)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5kvnd_kube-system(ad8eda04-e7ce-469d-b354-eba78e8e6145)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cd8c3d2960538adf0d6e0277964b67b10a1edb4b2a2a278bd57181deecaa9089\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-7db6d8ff4d-5kvnd" podUID="ad8eda04-e7ce-469d-b354-eba78e8e6145" Feb 13 19:06:23.461507 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4deada253c51f35f1d4ea345bae97bbd803b03adc4ebab4509f8d9f83a74681e-rootfs.mount: Deactivated successfully. Feb 13 19:06:23.780869 kubelet[2606]: I0213 19:06:23.780727 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-56xl5" podStartSLOduration=2.347199067 podStartE2EDuration="6.780711008s" podCreationTimestamp="2025-02-13 19:06:17 +0000 UTC" firstStartedPulling="2025-02-13 19:06:17.996934846 +0000 UTC m=+15.368391745" lastFinishedPulling="2025-02-13 19:06:22.430446747 +0000 UTC m=+19.801903686" observedRunningTime="2025-02-13 19:06:23.780602755 +0000 UTC m=+21.152059814" watchObservedRunningTime="2025-02-13 19:06:23.780711008 +0000 UTC m=+21.152167907" Feb 13 19:06:23.904799 systemd-networkd[1409]: flannel.1: Link UP Feb 13 19:06:23.904805 systemd-networkd[1409]: flannel.1: Gained carrier Feb 13 19:06:25.247701 systemd-networkd[1409]: flannel.1: Gained IPv6LL Feb 13 19:06:27.541195 systemd[1]: Started sshd@5-10.0.0.60:22-10.0.0.1:33062.service - OpenSSH per-connection server daemon (10.0.0.1:33062). Feb 13 19:06:27.588463 sshd[3258]: Accepted publickey for core from 10.0.0.1 port 33062 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:27.589978 sshd-session[3258]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:27.594552 systemd-logind[1469]: New session 6 of user core. Feb 13 19:06:27.603786 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:06:27.731326 sshd[3260]: Connection closed by 10.0.0.1 port 33062 Feb 13 19:06:27.731803 sshd-session[3258]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:27.738158 systemd[1]: sshd@5-10.0.0.60:22-10.0.0.1:33062.service: Deactivated successfully. Feb 13 19:06:27.740218 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:06:27.741863 systemd-logind[1469]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:06:27.743093 systemd-logind[1469]: Removed session 6. Feb 13 19:06:32.743458 systemd[1]: Started sshd@6-10.0.0.60:22-10.0.0.1:48934.service - OpenSSH per-connection server daemon (10.0.0.1:48934). Feb 13 19:06:32.813591 sshd[3296]: Accepted publickey for core from 10.0.0.1 port 48934 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:32.815036 sshd-session[3296]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:32.819493 systemd-logind[1469]: New session 7 of user core. Feb 13 19:06:32.831787 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:06:32.973273 sshd[3298]: Connection closed by 10.0.0.1 port 48934 Feb 13 19:06:32.973993 sshd-session[3296]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:32.977928 systemd[1]: sshd@6-10.0.0.60:22-10.0.0.1:48934.service: Deactivated successfully. Feb 13 19:06:32.979967 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:06:32.981272 systemd-logind[1469]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:06:32.982284 systemd-logind[1469]: Removed session 7. Feb 13 19:06:33.711556 containerd[1488]: time="2025-02-13T19:06:33.711483875Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hdmcl,Uid:0cc7b41b-4248-4e19-96b2-3f4bffb88b90,Namespace:kube-system,Attempt:0,}" Feb 13 19:06:33.755855 systemd-networkd[1409]: cni0: Link UP Feb 13 19:06:33.772944 kernel: cni0: port 1(veth449cf36f) entered blocking state Feb 13 19:06:33.773058 kernel: cni0: port 1(veth449cf36f) entered disabled state Feb 13 19:06:33.773074 kernel: veth449cf36f: entered allmulticast mode Feb 13 19:06:33.781675 kernel: veth449cf36f: entered promiscuous mode Feb 13 19:06:33.781764 kernel: cni0: port 1(veth449cf36f) entered blocking state Feb 13 19:06:33.783597 kernel: cni0: port 1(veth449cf36f) entered forwarding state Feb 13 19:06:33.788204 systemd-networkd[1409]: veth449cf36f: Link UP Feb 13 19:06:33.788588 kernel: cni0: port 1(veth449cf36f) entered disabled state Feb 13 19:06:33.800857 kernel: cni0: port 1(veth449cf36f) entered blocking state Feb 13 19:06:33.800959 kernel: cni0: port 1(veth449cf36f) entered forwarding state Feb 13 19:06:33.800957 systemd-networkd[1409]: veth449cf36f: Gained carrier Feb 13 19:06:33.801253 systemd-networkd[1409]: cni0: Gained carrier Feb 13 19:06:33.806035 containerd[1488]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x40000a68e8), "name":"cbr0", "type":"bridge"} Feb 13 19:06:33.806035 containerd[1488]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:06:33.823166 containerd[1488]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-02-13T19:06:33.823048140Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:06:33.823166 containerd[1488]: time="2025-02-13T19:06:33.823122346Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:06:33.823166 containerd[1488]: time="2025-02-13T19:06:33.823139867Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:33.823398 containerd[1488]: time="2025-02-13T19:06:33.823232154Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:33.841764 systemd[1]: Started cri-containerd-3d7d9829c7076e5de74ac544a5f09945603d435a35fd328582667c32f1c7ca3d.scope - libcontainer container 3d7d9829c7076e5de74ac544a5f09945603d435a35fd328582667c32f1c7ca3d. Feb 13 19:06:33.852826 systemd-resolved[1326]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:06:33.870491 containerd[1488]: time="2025-02-13T19:06:33.870445822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-hdmcl,Uid:0cc7b41b-4248-4e19-96b2-3f4bffb88b90,Namespace:kube-system,Attempt:0,} returns sandbox id \"3d7d9829c7076e5de74ac544a5f09945603d435a35fd328582667c32f1c7ca3d\"" Feb 13 19:06:33.874537 containerd[1488]: time="2025-02-13T19:06:33.874501167Z" level=info msg="CreateContainer within sandbox \"3d7d9829c7076e5de74ac544a5f09945603d435a35fd328582667c32f1c7ca3d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:06:33.889440 containerd[1488]: time="2025-02-13T19:06:33.889390526Z" level=info msg="CreateContainer within sandbox \"3d7d9829c7076e5de74ac544a5f09945603d435a35fd328582667c32f1c7ca3d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ebf35980746b41ed80cecd4ebb0f8fa2c6b77a419eb62dfe7781ecabe7059971\"" Feb 13 19:06:33.890160 containerd[1488]: time="2025-02-13T19:06:33.890124981Z" level=info msg="StartContainer for \"ebf35980746b41ed80cecd4ebb0f8fa2c6b77a419eb62dfe7781ecabe7059971\"" Feb 13 19:06:33.918793 systemd[1]: Started cri-containerd-ebf35980746b41ed80cecd4ebb0f8fa2c6b77a419eb62dfe7781ecabe7059971.scope - libcontainer container ebf35980746b41ed80cecd4ebb0f8fa2c6b77a419eb62dfe7781ecabe7059971. Feb 13 19:06:33.943799 containerd[1488]: time="2025-02-13T19:06:33.943758132Z" level=info msg="StartContainer for \"ebf35980746b41ed80cecd4ebb0f8fa2c6b77a419eb62dfe7781ecabe7059971\" returns successfully" Feb 13 19:06:34.729234 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1925373860.mount: Deactivated successfully. Feb 13 19:06:34.894738 kubelet[2606]: I0213 19:06:34.894593 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-hdmcl" podStartSLOduration=17.894547342 podStartE2EDuration="17.894547342s" podCreationTimestamp="2025-02-13 19:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:34.805744631 +0000 UTC m=+32.177201570" watchObservedRunningTime="2025-02-13 19:06:34.894547342 +0000 UTC m=+32.266004281" Feb 13 19:06:34.912687 systemd-networkd[1409]: cni0: Gained IPv6LL Feb 13 19:06:35.615740 systemd-networkd[1409]: veth449cf36f: Gained IPv6LL Feb 13 19:06:36.712049 containerd[1488]: time="2025-02-13T19:06:36.712005766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5kvnd,Uid:ad8eda04-e7ce-469d-b354-eba78e8e6145,Namespace:kube-system,Attempt:0,}" Feb 13 19:06:36.748899 systemd-networkd[1409]: veth848d7c0c: Link UP Feb 13 19:06:36.750676 kernel: cni0: port 2(veth848d7c0c) entered blocking state Feb 13 19:06:36.750746 kernel: cni0: port 2(veth848d7c0c) entered disabled state Feb 13 19:06:36.751672 kernel: veth848d7c0c: entered allmulticast mode Feb 13 19:06:36.752659 kernel: veth848d7c0c: entered promiscuous mode Feb 13 19:06:36.758247 kernel: cni0: port 2(veth848d7c0c) entered blocking state Feb 13 19:06:36.758655 kernel: cni0: port 2(veth848d7c0c) entered forwarding state Feb 13 19:06:36.758393 systemd-networkd[1409]: veth848d7c0c: Gained carrier Feb 13 19:06:36.762522 containerd[1488]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x4000012938), "name":"cbr0", "type":"bridge"} Feb 13 19:06:36.762522 containerd[1488]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:06:36.783982 containerd[1488]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-02-13T19:06:36.783855332Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:06:36.784122 containerd[1488]: time="2025-02-13T19:06:36.784012702Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:06:36.784122 containerd[1488]: time="2025-02-13T19:06:36.784049025Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:36.784380 containerd[1488]: time="2025-02-13T19:06:36.784152712Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:06:36.804960 systemd[1]: Started cri-containerd-4b34cbd19942c4b769cb661c1a00c2134f8c37c81fc89ca6d12675437f35b128.scope - libcontainer container 4b34cbd19942c4b769cb661c1a00c2134f8c37c81fc89ca6d12675437f35b128. Feb 13 19:06:36.819427 systemd-resolved[1326]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:06:36.839022 containerd[1488]: time="2025-02-13T19:06:36.838979049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5kvnd,Uid:ad8eda04-e7ce-469d-b354-eba78e8e6145,Namespace:kube-system,Attempt:0,} returns sandbox id \"4b34cbd19942c4b769cb661c1a00c2134f8c37c81fc89ca6d12675437f35b128\"" Feb 13 19:06:36.841747 containerd[1488]: time="2025-02-13T19:06:36.841705153Z" level=info msg="CreateContainer within sandbox \"4b34cbd19942c4b769cb661c1a00c2134f8c37c81fc89ca6d12675437f35b128\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:06:36.852717 containerd[1488]: time="2025-02-13T19:06:36.852674733Z" level=info msg="CreateContainer within sandbox \"4b34cbd19942c4b769cb661c1a00c2134f8c37c81fc89ca6d12675437f35b128\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"efa342beacb476fb4b2f54b1940c60587bed8333a0c9f45b29926179f5ef1522\"" Feb 13 19:06:36.853517 containerd[1488]: time="2025-02-13T19:06:36.853354378Z" level=info msg="StartContainer for \"efa342beacb476fb4b2f54b1940c60587bed8333a0c9f45b29926179f5ef1522\"" Feb 13 19:06:36.880804 systemd[1]: Started cri-containerd-efa342beacb476fb4b2f54b1940c60587bed8333a0c9f45b29926179f5ef1522.scope - libcontainer container efa342beacb476fb4b2f54b1940c60587bed8333a0c9f45b29926179f5ef1522. Feb 13 19:06:36.907210 containerd[1488]: time="2025-02-13T19:06:36.907145926Z" level=info msg="StartContainer for \"efa342beacb476fb4b2f54b1940c60587bed8333a0c9f45b29926179f5ef1522\" returns successfully" Feb 13 19:06:37.763354 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1998400313.mount: Deactivated successfully. Feb 13 19:06:37.816720 kubelet[2606]: I0213 19:06:37.816661 2606 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-5kvnd" podStartSLOduration=20.8166367 podStartE2EDuration="20.8166367s" podCreationTimestamp="2025-02-13 19:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:06:37.800716503 +0000 UTC m=+35.172173402" watchObservedRunningTime="2025-02-13 19:06:37.8166367 +0000 UTC m=+35.188093639" Feb 13 19:06:37.855699 systemd-networkd[1409]: veth848d7c0c: Gained IPv6LL Feb 13 19:06:37.986078 systemd[1]: Started sshd@7-10.0.0.60:22-10.0.0.1:48946.service - OpenSSH per-connection server daemon (10.0.0.1:48946). Feb 13 19:06:38.038826 sshd[3568]: Accepted publickey for core from 10.0.0.1 port 48946 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:38.040207 sshd-session[3568]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:38.044644 systemd-logind[1469]: New session 8 of user core. Feb 13 19:06:38.056759 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:06:38.173495 sshd[3570]: Connection closed by 10.0.0.1 port 48946 Feb 13 19:06:38.174350 sshd-session[3568]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:38.182972 systemd[1]: sshd@7-10.0.0.60:22-10.0.0.1:48946.service: Deactivated successfully. Feb 13 19:06:38.184810 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:06:38.185599 systemd-logind[1469]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:06:38.197897 systemd[1]: Started sshd@8-10.0.0.60:22-10.0.0.1:48958.service - OpenSSH per-connection server daemon (10.0.0.1:48958). Feb 13 19:06:38.199062 systemd-logind[1469]: Removed session 8. Feb 13 19:06:38.241974 sshd[3583]: Accepted publickey for core from 10.0.0.1 port 48958 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:38.241487 sshd-session[3583]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:38.247017 systemd-logind[1469]: New session 9 of user core. Feb 13 19:06:38.252974 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:06:38.395575 sshd[3586]: Connection closed by 10.0.0.1 port 48958 Feb 13 19:06:38.396148 sshd-session[3583]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:38.419872 systemd[1]: sshd@8-10.0.0.60:22-10.0.0.1:48958.service: Deactivated successfully. Feb 13 19:06:38.424473 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:06:38.426214 systemd-logind[1469]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:06:38.437914 systemd[1]: Started sshd@9-10.0.0.60:22-10.0.0.1:48970.service - OpenSSH per-connection server daemon (10.0.0.1:48970). Feb 13 19:06:38.439122 systemd-logind[1469]: Removed session 9. Feb 13 19:06:38.479451 sshd[3596]: Accepted publickey for core from 10.0.0.1 port 48970 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:38.480899 sshd-session[3596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:38.486273 systemd-logind[1469]: New session 10 of user core. Feb 13 19:06:38.496751 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:06:38.613318 sshd[3599]: Connection closed by 10.0.0.1 port 48970 Feb 13 19:06:38.613679 sshd-session[3596]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:38.617035 systemd[1]: sshd@9-10.0.0.60:22-10.0.0.1:48970.service: Deactivated successfully. Feb 13 19:06:38.619693 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:06:38.620423 systemd-logind[1469]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:06:38.621505 systemd-logind[1469]: Removed session 10. Feb 13 19:06:43.624970 systemd[1]: Started sshd@10-10.0.0.60:22-10.0.0.1:43592.service - OpenSSH per-connection server daemon (10.0.0.1:43592). Feb 13 19:06:43.671815 sshd[3633]: Accepted publickey for core from 10.0.0.1 port 43592 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:43.670926 sshd-session[3633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:43.675669 systemd-logind[1469]: New session 11 of user core. Feb 13 19:06:43.687797 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:06:43.800785 sshd[3635]: Connection closed by 10.0.0.1 port 43592 Feb 13 19:06:43.801243 sshd-session[3633]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:43.811819 systemd[1]: sshd@10-10.0.0.60:22-10.0.0.1:43592.service: Deactivated successfully. Feb 13 19:06:43.813361 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:06:43.814994 systemd-logind[1469]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:06:43.828891 systemd[1]: Started sshd@11-10.0.0.60:22-10.0.0.1:43596.service - OpenSSH per-connection server daemon (10.0.0.1:43596). Feb 13 19:06:43.829790 systemd-logind[1469]: Removed session 11. Feb 13 19:06:43.871021 sshd[3648]: Accepted publickey for core from 10.0.0.1 port 43596 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:43.872375 sshd-session[3648]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:43.876796 systemd-logind[1469]: New session 12 of user core. Feb 13 19:06:43.893764 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:06:44.107278 sshd[3651]: Connection closed by 10.0.0.1 port 43596 Feb 13 19:06:44.107966 sshd-session[3648]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:44.123018 systemd[1]: sshd@11-10.0.0.60:22-10.0.0.1:43596.service: Deactivated successfully. Feb 13 19:06:44.124678 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:06:44.125317 systemd-logind[1469]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:06:44.130872 systemd[1]: Started sshd@12-10.0.0.60:22-10.0.0.1:43602.service - OpenSSH per-connection server daemon (10.0.0.1:43602). Feb 13 19:06:44.133959 systemd-logind[1469]: Removed session 12. Feb 13 19:06:44.177156 sshd[3682]: Accepted publickey for core from 10.0.0.1 port 43602 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:44.178469 sshd-session[3682]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:44.182474 systemd-logind[1469]: New session 13 of user core. Feb 13 19:06:44.193733 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:06:45.441426 sshd[3685]: Connection closed by 10.0.0.1 port 43602 Feb 13 19:06:45.442150 sshd-session[3682]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:45.468983 systemd[1]: Started sshd@13-10.0.0.60:22-10.0.0.1:43604.service - OpenSSH per-connection server daemon (10.0.0.1:43604). Feb 13 19:06:45.469552 systemd[1]: sshd@12-10.0.0.60:22-10.0.0.1:43602.service: Deactivated successfully. Feb 13 19:06:45.471951 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:06:45.473514 systemd-logind[1469]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:06:45.475002 systemd-logind[1469]: Removed session 13. Feb 13 19:06:45.513228 sshd[3702]: Accepted publickey for core from 10.0.0.1 port 43604 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:45.514416 sshd-session[3702]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:45.519166 systemd-logind[1469]: New session 14 of user core. Feb 13 19:06:45.528701 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:06:45.748093 sshd[3707]: Connection closed by 10.0.0.1 port 43604 Feb 13 19:06:45.749010 sshd-session[3702]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:45.768930 systemd[1]: Started sshd@14-10.0.0.60:22-10.0.0.1:43616.service - OpenSSH per-connection server daemon (10.0.0.1:43616). Feb 13 19:06:45.769497 systemd[1]: sshd@13-10.0.0.60:22-10.0.0.1:43604.service: Deactivated successfully. Feb 13 19:06:45.771302 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:06:45.775480 systemd-logind[1469]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:06:45.777274 systemd-logind[1469]: Removed session 14. Feb 13 19:06:45.811248 sshd[3716]: Accepted publickey for core from 10.0.0.1 port 43616 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:45.812470 sshd-session[3716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:45.819015 systemd-logind[1469]: New session 15 of user core. Feb 13 19:06:45.828729 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:06:45.936960 sshd[3721]: Connection closed by 10.0.0.1 port 43616 Feb 13 19:06:45.937270 sshd-session[3716]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:45.940341 systemd[1]: sshd@14-10.0.0.60:22-10.0.0.1:43616.service: Deactivated successfully. Feb 13 19:06:45.941959 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:06:45.942621 systemd-logind[1469]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:06:45.943638 systemd-logind[1469]: Removed session 15. Feb 13 19:06:50.948246 systemd[1]: Started sshd@15-10.0.0.60:22-10.0.0.1:43628.service - OpenSSH per-connection server daemon (10.0.0.1:43628). Feb 13 19:06:50.991457 sshd[3760]: Accepted publickey for core from 10.0.0.1 port 43628 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:50.992692 sshd-session[3760]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:50.996622 systemd-logind[1469]: New session 16 of user core. Feb 13 19:06:51.007783 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:06:51.116597 sshd[3762]: Connection closed by 10.0.0.1 port 43628 Feb 13 19:06:51.117760 sshd-session[3760]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:51.120223 systemd[1]: sshd@15-10.0.0.60:22-10.0.0.1:43628.service: Deactivated successfully. Feb 13 19:06:51.121761 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:06:51.126803 systemd-logind[1469]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:06:51.128191 systemd-logind[1469]: Removed session 16. Feb 13 19:06:56.129061 systemd[1]: Started sshd@16-10.0.0.60:22-10.0.0.1:50930.service - OpenSSH per-connection server daemon (10.0.0.1:50930). Feb 13 19:06:56.174698 sshd[3797]: Accepted publickey for core from 10.0.0.1 port 50930 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:06:56.175972 sshd-session[3797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:06:56.180887 systemd-logind[1469]: New session 17 of user core. Feb 13 19:06:56.190764 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:06:56.314943 sshd[3799]: Connection closed by 10.0.0.1 port 50930 Feb 13 19:06:56.315288 sshd-session[3797]: pam_unix(sshd:session): session closed for user core Feb 13 19:06:56.319630 systemd[1]: sshd@16-10.0.0.60:22-10.0.0.1:50930.service: Deactivated successfully. Feb 13 19:06:56.321321 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:06:56.324482 systemd-logind[1469]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:06:56.325507 systemd-logind[1469]: Removed session 17. Feb 13 19:07:01.330587 systemd[1]: Started sshd@17-10.0.0.60:22-10.0.0.1:50934.service - OpenSSH per-connection server daemon (10.0.0.1:50934). Feb 13 19:07:01.380031 sshd[3834]: Accepted publickey for core from 10.0.0.1 port 50934 ssh2: RSA SHA256:QyQQN4NlJHXH6/vW7NxDLOKgT/2dxBjCkGLAHoHnd3w Feb 13 19:07:01.380617 sshd-session[3834]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:07:01.385405 systemd-logind[1469]: New session 18 of user core. Feb 13 19:07:01.395791 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:07:01.534338 sshd[3836]: Connection closed by 10.0.0.1 port 50934 Feb 13 19:07:01.534881 sshd-session[3834]: pam_unix(sshd:session): session closed for user core Feb 13 19:07:01.541174 systemd[1]: sshd@17-10.0.0.60:22-10.0.0.1:50934.service: Deactivated successfully. Feb 13 19:07:01.542900 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:07:01.543631 systemd-logind[1469]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:07:01.544677 systemd-logind[1469]: Removed session 18.