Mar 7 00:49:59.885489 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Mar 7 00:49:59.885513 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Mar 6 22:59:59 -00 2026 Mar 7 00:49:59.885524 kernel: KASLR enabled Mar 7 00:49:59.885530 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Mar 7 00:49:59.885535 kernel: efi: SMBIOS 3.0=0x139ed0000 MEMATTR=0x1390c1018 ACPI 2.0=0x136760018 RNG=0x13676e918 MEMRESERVE=0x136b43d18 Mar 7 00:49:59.885541 kernel: random: crng init done Mar 7 00:49:59.885548 kernel: ACPI: Early table checksum verification disabled Mar 7 00:49:59.885554 kernel: ACPI: RSDP 0x0000000136760018 000024 (v02 BOCHS ) Mar 7 00:49:59.885560 kernel: ACPI: XSDT 0x000000013676FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Mar 7 00:49:59.885579 kernel: ACPI: FACP 0x000000013676FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885586 kernel: ACPI: DSDT 0x0000000136767518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885592 kernel: ACPI: APIC 0x000000013676FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885598 kernel: ACPI: PPTT 0x000000013676FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885604 kernel: ACPI: GTDT 0x000000013676D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885612 kernel: ACPI: MCFG 0x000000013676FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885620 kernel: ACPI: SPCR 0x000000013676E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885627 kernel: ACPI: DBG2 0x000000013676E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885633 kernel: ACPI: IORT 0x000000013676E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 7 00:49:59.885640 kernel: ACPI: BGRT 0x000000013676E798 000038 (v01 INTEL EDK2 00000002 01000013) Mar 7 00:49:59.885646 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Mar 7 00:49:59.885652 kernel: NUMA: Failed to initialise from firmware Mar 7 00:49:59.885659 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Mar 7 00:49:59.885665 kernel: NUMA: NODE_DATA [mem 0x13966f800-0x139674fff] Mar 7 00:49:59.885671 kernel: Zone ranges: Mar 7 00:49:59.885678 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Mar 7 00:49:59.885686 kernel: DMA32 empty Mar 7 00:49:59.885692 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Mar 7 00:49:59.885699 kernel: Movable zone start for each node Mar 7 00:49:59.885705 kernel: Early memory node ranges Mar 7 00:49:59.885711 kernel: node 0: [mem 0x0000000040000000-0x000000013676ffff] Mar 7 00:49:59.885718 kernel: node 0: [mem 0x0000000136770000-0x0000000136b3ffff] Mar 7 00:49:59.885724 kernel: node 0: [mem 0x0000000136b40000-0x0000000139e1ffff] Mar 7 00:49:59.885731 kernel: node 0: [mem 0x0000000139e20000-0x0000000139eaffff] Mar 7 00:49:59.885737 kernel: node 0: [mem 0x0000000139eb0000-0x0000000139ebffff] Mar 7 00:49:59.885743 kernel: node 0: [mem 0x0000000139ec0000-0x0000000139fdffff] Mar 7 00:49:59.885750 kernel: node 0: [mem 0x0000000139fe0000-0x0000000139ffffff] Mar 7 00:49:59.885756 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Mar 7 00:49:59.885764 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Mar 7 00:49:59.885771 kernel: psci: probing for conduit method from ACPI. Mar 7 00:49:59.885777 kernel: psci: PSCIv1.1 detected in firmware. Mar 7 00:49:59.885786 kernel: psci: Using standard PSCI v0.2 function IDs Mar 7 00:49:59.885793 kernel: psci: Trusted OS migration not required Mar 7 00:49:59.885800 kernel: psci: SMC Calling Convention v1.1 Mar 7 00:49:59.885809 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Mar 7 00:49:59.885815 kernel: percpu: Embedded 30 pages/cpu s85736 r8192 d28952 u122880 Mar 7 00:49:59.885822 kernel: pcpu-alloc: s85736 r8192 d28952 u122880 alloc=30*4096 Mar 7 00:49:59.885829 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 7 00:49:59.885836 kernel: Detected PIPT I-cache on CPU0 Mar 7 00:49:59.885843 kernel: CPU features: detected: GIC system register CPU interface Mar 7 00:49:59.885849 kernel: CPU features: detected: Hardware dirty bit management Mar 7 00:49:59.885856 kernel: CPU features: detected: Spectre-v4 Mar 7 00:49:59.885863 kernel: CPU features: detected: Spectre-BHB Mar 7 00:49:59.885870 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 7 00:49:59.885878 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 7 00:49:59.885885 kernel: CPU features: detected: ARM erratum 1418040 Mar 7 00:49:59.885891 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 7 00:49:59.885898 kernel: alternatives: applying boot alternatives Mar 7 00:49:59.885906 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 00:49:59.885913 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 7 00:49:59.885920 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 7 00:49:59.885926 kernel: Fallback order for Node 0: 0 Mar 7 00:49:59.885933 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Mar 7 00:49:59.885940 kernel: Policy zone: Normal Mar 7 00:49:59.885947 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 7 00:49:59.885955 kernel: software IO TLB: area num 2. Mar 7 00:49:59.885962 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Mar 7 00:49:59.885969 kernel: Memory: 3882816K/4096000K available (10304K kernel code, 2180K rwdata, 8116K rodata, 39424K init, 897K bss, 213184K reserved, 0K cma-reserved) Mar 7 00:49:59.885977 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 7 00:49:59.885983 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 7 00:49:59.885991 kernel: rcu: RCU event tracing is enabled. Mar 7 00:49:59.885998 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 7 00:49:59.886005 kernel: Trampoline variant of Tasks RCU enabled. Mar 7 00:49:59.886012 kernel: Tracing variant of Tasks RCU enabled. Mar 7 00:49:59.886018 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 7 00:49:59.886025 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 7 00:49:59.886032 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 7 00:49:59.886086 kernel: GICv3: 256 SPIs implemented Mar 7 00:49:59.886093 kernel: GICv3: 0 Extended SPIs implemented Mar 7 00:49:59.886100 kernel: Root IRQ handler: gic_handle_irq Mar 7 00:49:59.886107 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Mar 7 00:49:59.886114 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Mar 7 00:49:59.886120 kernel: ITS [mem 0x08080000-0x0809ffff] Mar 7 00:49:59.886127 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Mar 7 00:49:59.886134 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Mar 7 00:49:59.886204 kernel: GICv3: using LPI property table @0x00000001000e0000 Mar 7 00:49:59.886212 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Mar 7 00:49:59.886219 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 7 00:49:59.886230 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 00:49:59.886237 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Mar 7 00:49:59.886244 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Mar 7 00:49:59.886251 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Mar 7 00:49:59.886258 kernel: Console: colour dummy device 80x25 Mar 7 00:49:59.886265 kernel: ACPI: Core revision 20230628 Mar 7 00:49:59.886272 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Mar 7 00:49:59.886279 kernel: pid_max: default: 32768 minimum: 301 Mar 7 00:49:59.886286 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 7 00:49:59.886293 kernel: landlock: Up and running. Mar 7 00:49:59.886302 kernel: SELinux: Initializing. Mar 7 00:49:59.886310 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 00:49:59.886317 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 7 00:49:59.886324 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 00:49:59.886331 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 7 00:49:59.886338 kernel: rcu: Hierarchical SRCU implementation. Mar 7 00:49:59.886346 kernel: rcu: Max phase no-delay instances is 400. Mar 7 00:49:59.886353 kernel: Platform MSI: ITS@0x8080000 domain created Mar 7 00:49:59.886360 kernel: PCI/MSI: ITS@0x8080000 domain created Mar 7 00:49:59.886368 kernel: Remapping and enabling EFI services. Mar 7 00:49:59.886375 kernel: smp: Bringing up secondary CPUs ... Mar 7 00:49:59.886383 kernel: Detected PIPT I-cache on CPU1 Mar 7 00:49:59.886390 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Mar 7 00:49:59.886397 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Mar 7 00:49:59.886404 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 7 00:49:59.886411 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Mar 7 00:49:59.886418 kernel: smp: Brought up 1 node, 2 CPUs Mar 7 00:49:59.886425 kernel: SMP: Total of 2 processors activated. Mar 7 00:49:59.886434 kernel: CPU features: detected: 32-bit EL0 Support Mar 7 00:49:59.886441 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 7 00:49:59.886448 kernel: CPU features: detected: Common not Private translations Mar 7 00:49:59.886460 kernel: CPU features: detected: CRC32 instructions Mar 7 00:49:59.886469 kernel: CPU features: detected: Enhanced Virtualization Traps Mar 7 00:49:59.886477 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 7 00:49:59.886484 kernel: CPU features: detected: LSE atomic instructions Mar 7 00:49:59.886492 kernel: CPU features: detected: Privileged Access Never Mar 7 00:49:59.886499 kernel: CPU features: detected: RAS Extension Support Mar 7 00:49:59.886508 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Mar 7 00:49:59.886516 kernel: CPU: All CPU(s) started at EL1 Mar 7 00:49:59.886523 kernel: alternatives: applying system-wide alternatives Mar 7 00:49:59.886530 kernel: devtmpfs: initialized Mar 7 00:49:59.886538 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 7 00:49:59.886545 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 7 00:49:59.886553 kernel: pinctrl core: initialized pinctrl subsystem Mar 7 00:49:59.886560 kernel: SMBIOS 3.0.0 present. Mar 7 00:49:59.886593 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Mar 7 00:49:59.886602 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 7 00:49:59.886609 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 7 00:49:59.886617 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 7 00:49:59.886624 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 7 00:49:59.886632 kernel: audit: initializing netlink subsys (disabled) Mar 7 00:49:59.886640 kernel: audit: type=2000 audit(0.018:1): state=initialized audit_enabled=0 res=1 Mar 7 00:49:59.886647 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 7 00:49:59.886655 kernel: cpuidle: using governor menu Mar 7 00:49:59.886666 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 7 00:49:59.886673 kernel: ASID allocator initialised with 32768 entries Mar 7 00:49:59.886681 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 7 00:49:59.886688 kernel: Serial: AMBA PL011 UART driver Mar 7 00:49:59.886696 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 7 00:49:59.886703 kernel: Modules: 0 pages in range for non-PLT usage Mar 7 00:49:59.886710 kernel: Modules: 509008 pages in range for PLT usage Mar 7 00:49:59.886718 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 7 00:49:59.886725 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 7 00:49:59.886734 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 7 00:49:59.886742 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 7 00:49:59.886749 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 7 00:49:59.886757 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 7 00:49:59.886764 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 7 00:49:59.886772 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 7 00:49:59.886780 kernel: ACPI: Added _OSI(Module Device) Mar 7 00:49:59.886787 kernel: ACPI: Added _OSI(Processor Device) Mar 7 00:49:59.886794 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 7 00:49:59.886803 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 7 00:49:59.886811 kernel: ACPI: Interpreter enabled Mar 7 00:49:59.886818 kernel: ACPI: Using GIC for interrupt routing Mar 7 00:49:59.886826 kernel: ACPI: MCFG table detected, 1 entries Mar 7 00:49:59.886833 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Mar 7 00:49:59.886841 kernel: printk: console [ttyAMA0] enabled Mar 7 00:49:59.886849 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 7 00:49:59.887015 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 7 00:49:59.887120 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Mar 7 00:49:59.887193 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Mar 7 00:49:59.887260 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Mar 7 00:49:59.887325 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Mar 7 00:49:59.887335 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Mar 7 00:49:59.887342 kernel: PCI host bridge to bus 0000:00 Mar 7 00:49:59.887419 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Mar 7 00:49:59.887485 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Mar 7 00:49:59.887545 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Mar 7 00:49:59.887748 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 7 00:49:59.887838 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Mar 7 00:49:59.887916 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Mar 7 00:49:59.887984 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Mar 7 00:49:59.888064 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Mar 7 00:49:59.888148 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.888216 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Mar 7 00:49:59.888296 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.888363 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Mar 7 00:49:59.888445 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.888514 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Mar 7 00:49:59.888603 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.888766 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Mar 7 00:49:59.888848 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.888918 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Mar 7 00:49:59.888995 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.889114 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Mar 7 00:49:59.890717 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.890808 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Mar 7 00:49:59.890885 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.890953 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Mar 7 00:49:59.891028 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Mar 7 00:49:59.891162 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Mar 7 00:49:59.891257 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Mar 7 00:49:59.891326 kernel: pci 0000:00:04.0: reg 0x10: [io 0x0000-0x0007] Mar 7 00:49:59.891491 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Mar 7 00:49:59.891619 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Mar 7 00:49:59.891708 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Mar 7 00:49:59.891779 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Mar 7 00:49:59.891856 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Mar 7 00:49:59.891942 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Mar 7 00:49:59.892023 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Mar 7 00:49:59.892109 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Mar 7 00:49:59.892181 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Mar 7 00:49:59.892272 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Mar 7 00:49:59.892344 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Mar 7 00:49:59.892474 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Mar 7 00:49:59.892559 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x10800000-0x10800fff] Mar 7 00:49:59.896784 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Mar 7 00:49:59.896883 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Mar 7 00:49:59.896955 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Mar 7 00:49:59.897027 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Mar 7 00:49:59.897142 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Mar 7 00:49:59.897215 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Mar 7 00:49:59.897286 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Mar 7 00:49:59.897356 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Mar 7 00:49:59.897430 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Mar 7 00:49:59.897498 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Mar 7 00:49:59.897598 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Mar 7 00:49:59.897680 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Mar 7 00:49:59.897752 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Mar 7 00:49:59.897821 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Mar 7 00:49:59.897990 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Mar 7 00:49:59.898083 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Mar 7 00:49:59.898156 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Mar 7 00:49:59.898228 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Mar 7 00:49:59.898302 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Mar 7 00:49:59.898380 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Mar 7 00:49:59.898451 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Mar 7 00:49:59.898519 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Mar 7 00:49:59.901500 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Mar 7 00:49:59.901632 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Mar 7 00:49:59.901706 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Mar 7 00:49:59.901774 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Mar 7 00:49:59.901854 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Mar 7 00:49:59.901921 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Mar 7 00:49:59.901988 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Mar 7 00:49:59.902081 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Mar 7 00:49:59.902154 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Mar 7 00:49:59.902224 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Mar 7 00:49:59.902296 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Mar 7 00:49:59.902368 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Mar 7 00:49:59.902435 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Mar 7 00:49:59.902506 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Mar 7 00:49:59.902593 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Mar 7 00:49:59.902668 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Mar 7 00:49:59.902738 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Mar 7 00:49:59.902815 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Mar 7 00:49:59.902893 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Mar 7 00:49:59.902963 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Mar 7 00:49:59.903033 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Mar 7 00:49:59.903117 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Mar 7 00:49:59.903185 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Mar 7 00:49:59.903252 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Mar 7 00:49:59.903318 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Mar 7 00:49:59.903394 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Mar 7 00:49:59.903461 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Mar 7 00:49:59.903529 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Mar 7 00:49:59.906718 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Mar 7 00:49:59.906811 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Mar 7 00:49:59.906880 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Mar 7 00:49:59.906954 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Mar 7 00:49:59.907029 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Mar 7 00:49:59.907148 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Mar 7 00:49:59.907220 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Mar 7 00:49:59.907291 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Mar 7 00:49:59.907357 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Mar 7 00:49:59.907427 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Mar 7 00:49:59.907493 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Mar 7 00:49:59.907562 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Mar 7 00:49:59.907657 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Mar 7 00:49:59.907730 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Mar 7 00:49:59.907797 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Mar 7 00:49:59.907865 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Mar 7 00:49:59.907931 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Mar 7 00:49:59.907999 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Mar 7 00:49:59.908081 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Mar 7 00:49:59.908151 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Mar 7 00:49:59.908223 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Mar 7 00:49:59.908290 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Mar 7 00:49:59.908357 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Mar 7 00:49:59.908428 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Mar 7 00:49:59.908504 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Mar 7 00:49:59.910647 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Mar 7 00:49:59.910773 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Mar 7 00:49:59.910855 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Mar 7 00:49:59.910940 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Mar 7 00:49:59.911009 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Mar 7 00:49:59.911099 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Mar 7 00:49:59.911182 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Mar 7 00:49:59.911258 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Mar 7 00:49:59.911327 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Mar 7 00:49:59.911393 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Mar 7 00:49:59.911461 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Mar 7 00:49:59.911537 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Mar 7 00:49:59.911629 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Mar 7 00:49:59.911701 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Mar 7 00:49:59.911768 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Mar 7 00:49:59.911840 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Mar 7 00:49:59.911906 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Mar 7 00:49:59.911982 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Mar 7 00:49:59.912097 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Mar 7 00:49:59.912174 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Mar 7 00:49:59.912243 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Mar 7 00:49:59.912309 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Mar 7 00:49:59.912384 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Mar 7 00:49:59.912460 kernel: pci 0000:05:00.0: BAR 1: assigned [mem 0x10800000-0x10800fff] Mar 7 00:49:59.912529 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Mar 7 00:49:59.914725 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Mar 7 00:49:59.914811 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Mar 7 00:49:59.914879 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Mar 7 00:49:59.914957 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Mar 7 00:49:59.915027 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Mar 7 00:49:59.915163 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Mar 7 00:49:59.915247 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Mar 7 00:49:59.915326 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Mar 7 00:49:59.915395 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Mar 7 00:49:59.915481 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Mar 7 00:49:59.915551 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Mar 7 00:49:59.915632 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Mar 7 00:49:59.915701 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Mar 7 00:49:59.915768 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Mar 7 00:49:59.915840 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Mar 7 00:49:59.915906 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Mar 7 00:49:59.915977 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Mar 7 00:49:59.916056 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Mar 7 00:49:59.916125 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Mar 7 00:49:59.916190 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Mar 7 00:49:59.916259 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Mar 7 00:49:59.916326 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Mar 7 00:49:59.916395 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Mar 7 00:49:59.916462 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Mar 7 00:49:59.916529 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Mar 7 00:49:59.919715 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Mar 7 00:49:59.919801 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Mar 7 00:49:59.919883 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Mar 7 00:49:59.919951 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Mar 7 00:49:59.920022 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Mar 7 00:49:59.920137 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Mar 7 00:49:59.920206 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Mar 7 00:49:59.920268 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Mar 7 00:49:59.920340 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Mar 7 00:49:59.920414 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Mar 7 00:49:59.920482 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Mar 7 00:49:59.920554 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Mar 7 00:49:59.920636 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Mar 7 00:49:59.920718 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Mar 7 00:49:59.920788 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Mar 7 00:49:59.920854 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Mar 7 00:49:59.920917 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Mar 7 00:49:59.920989 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Mar 7 00:49:59.921070 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Mar 7 00:49:59.921141 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Mar 7 00:49:59.921214 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Mar 7 00:49:59.921280 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Mar 7 00:49:59.921342 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Mar 7 00:49:59.921412 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Mar 7 00:49:59.921474 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Mar 7 00:49:59.921536 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Mar 7 00:49:59.922775 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Mar 7 00:49:59.922855 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Mar 7 00:49:59.922924 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Mar 7 00:49:59.922934 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Mar 7 00:49:59.922943 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Mar 7 00:49:59.922951 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Mar 7 00:49:59.922959 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Mar 7 00:49:59.922966 kernel: iommu: Default domain type: Translated Mar 7 00:49:59.922974 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 7 00:49:59.922982 kernel: efivars: Registered efivars operations Mar 7 00:49:59.922993 kernel: vgaarb: loaded Mar 7 00:49:59.923001 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 7 00:49:59.923009 kernel: VFS: Disk quotas dquot_6.6.0 Mar 7 00:49:59.923017 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 7 00:49:59.923025 kernel: pnp: PnP ACPI init Mar 7 00:49:59.923158 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Mar 7 00:49:59.923173 kernel: pnp: PnP ACPI: found 1 devices Mar 7 00:49:59.923181 kernel: NET: Registered PF_INET protocol family Mar 7 00:49:59.923189 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 7 00:49:59.923201 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 7 00:49:59.923209 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 7 00:49:59.923217 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 7 00:49:59.923225 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 7 00:49:59.923233 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 7 00:49:59.923241 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 00:49:59.923249 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 7 00:49:59.923257 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 7 00:49:59.923338 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Mar 7 00:49:59.923352 kernel: PCI: CLS 0 bytes, default 64 Mar 7 00:49:59.923361 kernel: kvm [1]: HYP mode not available Mar 7 00:49:59.923368 kernel: Initialise system trusted keyrings Mar 7 00:49:59.923376 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 7 00:49:59.923384 kernel: Key type asymmetric registered Mar 7 00:49:59.923392 kernel: Asymmetric key parser 'x509' registered Mar 7 00:49:59.923399 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 7 00:49:59.923407 kernel: io scheduler mq-deadline registered Mar 7 00:49:59.923417 kernel: io scheduler kyber registered Mar 7 00:49:59.923425 kernel: io scheduler bfq registered Mar 7 00:49:59.923434 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Mar 7 00:49:59.923506 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Mar 7 00:49:59.923591 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Mar 7 00:49:59.923664 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.923735 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Mar 7 00:49:59.923807 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Mar 7 00:49:59.923876 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.923947 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Mar 7 00:49:59.924014 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Mar 7 00:49:59.924093 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.924167 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Mar 7 00:49:59.924243 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Mar 7 00:49:59.924311 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.924383 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Mar 7 00:49:59.924451 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Mar 7 00:49:59.924518 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.927656 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Mar 7 00:49:59.927758 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Mar 7 00:49:59.927834 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.927907 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Mar 7 00:49:59.927977 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Mar 7 00:49:59.928083 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.928163 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Mar 7 00:49:59.928240 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Mar 7 00:49:59.928308 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.928319 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Mar 7 00:49:59.928388 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Mar 7 00:49:59.928456 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Mar 7 00:49:59.928524 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Mar 7 00:49:59.928535 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Mar 7 00:49:59.928546 kernel: ACPI: button: Power Button [PWRB] Mar 7 00:49:59.928554 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Mar 7 00:49:59.928697 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Mar 7 00:49:59.928775 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Mar 7 00:49:59.928787 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 7 00:49:59.928796 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Mar 7 00:49:59.928864 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Mar 7 00:49:59.928875 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Mar 7 00:49:59.928883 kernel: thunder_xcv, ver 1.0 Mar 7 00:49:59.928894 kernel: thunder_bgx, ver 1.0 Mar 7 00:49:59.928902 kernel: nicpf, ver 1.0 Mar 7 00:49:59.928918 kernel: nicvf, ver 1.0 Mar 7 00:49:59.929001 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 7 00:49:59.929098 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-03-07T00:49:59 UTC (1772844599) Mar 7 00:49:59.929111 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 7 00:49:59.929119 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Mar 7 00:49:59.929127 kernel: watchdog: Delayed init of the lockup detector failed: -19 Mar 7 00:49:59.929138 kernel: watchdog: Hard watchdog permanently disabled Mar 7 00:49:59.929146 kernel: NET: Registered PF_INET6 protocol family Mar 7 00:49:59.929155 kernel: Segment Routing with IPv6 Mar 7 00:49:59.929162 kernel: In-situ OAM (IOAM) with IPv6 Mar 7 00:49:59.929171 kernel: NET: Registered PF_PACKET protocol family Mar 7 00:49:59.929178 kernel: Key type dns_resolver registered Mar 7 00:49:59.929187 kernel: registered taskstats version 1 Mar 7 00:49:59.929194 kernel: Loading compiled-in X.509 certificates Mar 7 00:49:59.929202 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: e62b4e4ebcb406beff1271ecc7444548c4ab67e9' Mar 7 00:49:59.929212 kernel: Key type .fscrypt registered Mar 7 00:49:59.929220 kernel: Key type fscrypt-provisioning registered Mar 7 00:49:59.929227 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 7 00:49:59.929235 kernel: ima: Allocated hash algorithm: sha1 Mar 7 00:49:59.929243 kernel: ima: No architecture policies found Mar 7 00:49:59.929251 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 7 00:49:59.929259 kernel: clk: Disabling unused clocks Mar 7 00:49:59.929267 kernel: Freeing unused kernel memory: 39424K Mar 7 00:49:59.929274 kernel: Run /init as init process Mar 7 00:49:59.929284 kernel: with arguments: Mar 7 00:49:59.929292 kernel: /init Mar 7 00:49:59.929300 kernel: with environment: Mar 7 00:49:59.929307 kernel: HOME=/ Mar 7 00:49:59.929315 kernel: TERM=linux Mar 7 00:49:59.929325 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 00:49:59.929336 systemd[1]: Detected virtualization kvm. Mar 7 00:49:59.929344 systemd[1]: Detected architecture arm64. Mar 7 00:49:59.929354 systemd[1]: Running in initrd. Mar 7 00:49:59.929362 systemd[1]: No hostname configured, using default hostname. Mar 7 00:49:59.929370 systemd[1]: Hostname set to . Mar 7 00:49:59.929379 systemd[1]: Initializing machine ID from VM UUID. Mar 7 00:49:59.929389 systemd[1]: Queued start job for default target initrd.target. Mar 7 00:49:59.929398 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 00:49:59.929406 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 00:49:59.929415 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 7 00:49:59.929425 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 00:49:59.929434 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 7 00:49:59.929443 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 7 00:49:59.929453 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 7 00:49:59.929461 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 7 00:49:59.929470 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 00:49:59.929480 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 00:49:59.929488 systemd[1]: Reached target paths.target - Path Units. Mar 7 00:49:59.929496 systemd[1]: Reached target slices.target - Slice Units. Mar 7 00:49:59.929505 systemd[1]: Reached target swap.target - Swaps. Mar 7 00:49:59.929513 systemd[1]: Reached target timers.target - Timer Units. Mar 7 00:49:59.929521 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 00:49:59.929530 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 00:49:59.929538 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 7 00:49:59.929546 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 7 00:49:59.929557 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 00:49:59.929576 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 00:49:59.929586 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 00:49:59.929595 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 00:49:59.929604 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 7 00:49:59.929612 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 00:49:59.929620 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 7 00:49:59.929629 systemd[1]: Starting systemd-fsck-usr.service... Mar 7 00:49:59.929637 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 00:49:59.929647 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 00:49:59.929656 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 00:49:59.929664 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 7 00:49:59.929694 systemd-journald[237]: Collecting audit messages is disabled. Mar 7 00:49:59.929716 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 00:49:59.929725 systemd[1]: Finished systemd-fsck-usr.service. Mar 7 00:49:59.929734 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 7 00:49:59.929743 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 7 00:49:59.929753 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 00:49:59.929762 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 00:49:59.929770 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 00:49:59.929780 systemd-journald[237]: Journal started Mar 7 00:49:59.929800 systemd-journald[237]: Runtime Journal (/run/log/journal/54655f3653604e79b74c22d6e530746d) is 8.0M, max 76.6M, 68.6M free. Mar 7 00:49:59.913322 systemd-modules-load[238]: Inserted module 'overlay' Mar 7 00:49:59.931056 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 00:49:59.935301 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 00:49:59.940608 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 7 00:49:59.943436 systemd-modules-load[238]: Inserted module 'br_netfilter' Mar 7 00:49:59.944096 kernel: Bridge firewalling registered Mar 7 00:49:59.945817 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 00:49:59.949627 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 00:49:59.952360 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 00:49:59.957909 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 7 00:49:59.961027 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 00:49:59.969959 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 00:49:59.977976 dracut-cmdline[266]: dracut-dracut-053 Mar 7 00:49:59.983622 dracut-cmdline[266]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=9d22c40559a0d209dc0fcc2dfdd5ddf9671e6da0cc59463f610ba522f01325a6 Mar 7 00:49:59.991825 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 00:49:59.998851 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 00:50:00.025734 systemd-resolved[296]: Positive Trust Anchors: Mar 7 00:50:00.025749 systemd-resolved[296]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 00:50:00.025781 systemd-resolved[296]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 00:50:00.035664 systemd-resolved[296]: Defaulting to hostname 'linux'. Mar 7 00:50:00.037798 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 00:50:00.038427 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 00:50:00.045598 kernel: SCSI subsystem initialized Mar 7 00:50:00.049602 kernel: Loading iSCSI transport class v2.0-870. Mar 7 00:50:00.057622 kernel: iscsi: registered transport (tcp) Mar 7 00:50:00.072680 kernel: iscsi: registered transport (qla4xxx) Mar 7 00:50:00.072776 kernel: QLogic iSCSI HBA Driver Mar 7 00:50:00.120314 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 7 00:50:00.129888 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 7 00:50:00.149957 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 7 00:50:00.150089 kernel: device-mapper: uevent: version 1.0.3 Mar 7 00:50:00.150122 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 7 00:50:00.199648 kernel: raid6: neonx8 gen() 15681 MB/s Mar 7 00:50:00.216607 kernel: raid6: neonx4 gen() 15587 MB/s Mar 7 00:50:00.233636 kernel: raid6: neonx2 gen() 13149 MB/s Mar 7 00:50:00.250628 kernel: raid6: neonx1 gen() 10415 MB/s Mar 7 00:50:00.267637 kernel: raid6: int64x8 gen() 6919 MB/s Mar 7 00:50:00.284621 kernel: raid6: int64x4 gen() 7311 MB/s Mar 7 00:50:00.301617 kernel: raid6: int64x2 gen() 6106 MB/s Mar 7 00:50:00.318617 kernel: raid6: int64x1 gen() 5039 MB/s Mar 7 00:50:00.318694 kernel: raid6: using algorithm neonx8 gen() 15681 MB/s Mar 7 00:50:00.335633 kernel: raid6: .... xor() 11936 MB/s, rmw enabled Mar 7 00:50:00.335717 kernel: raid6: using neon recovery algorithm Mar 7 00:50:00.340828 kernel: xor: measuring software checksum speed Mar 7 00:50:00.340902 kernel: 8regs : 17462 MB/sec Mar 7 00:50:00.340941 kernel: 32regs : 19702 MB/sec Mar 7 00:50:00.340965 kernel: arm64_neon : 27150 MB/sec Mar 7 00:50:00.341618 kernel: xor: using function: arm64_neon (27150 MB/sec) Mar 7 00:50:00.391643 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 7 00:50:00.405583 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 7 00:50:00.413844 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 00:50:00.427268 systemd-udevd[456]: Using default interface naming scheme 'v255'. Mar 7 00:50:00.431645 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 00:50:00.439964 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 7 00:50:00.458272 dracut-pre-trigger[464]: rd.md=0: removing MD RAID activation Mar 7 00:50:00.492401 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 00:50:00.498779 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 00:50:00.548216 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 00:50:00.553740 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 7 00:50:00.580637 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 7 00:50:00.583068 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 00:50:00.584846 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 00:50:00.586395 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 00:50:00.594778 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 7 00:50:00.616600 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 7 00:50:00.654312 kernel: scsi host0: Virtio SCSI HBA Mar 7 00:50:00.670587 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 7 00:50:00.670734 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Mar 7 00:50:00.676650 kernel: ACPI: bus type USB registered Mar 7 00:50:00.679658 kernel: usbcore: registered new interface driver usbfs Mar 7 00:50:00.679776 kernel: usbcore: registered new interface driver hub Mar 7 00:50:00.681583 kernel: usbcore: registered new device driver usb Mar 7 00:50:00.683797 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 00:50:00.683932 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 00:50:00.687907 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 00:50:00.688841 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 00:50:00.689016 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 00:50:00.691413 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 00:50:00.699867 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 00:50:00.718992 kernel: sr 0:0:0:0: Power-on or device reset occurred Mar 7 00:50:00.719229 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Mar 7 00:50:00.720811 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 7 00:50:00.722599 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Mar 7 00:50:00.727469 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 00:50:00.733712 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 7 00:50:00.733897 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Mar 7 00:50:00.733984 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Mar 7 00:50:00.734111 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Mar 7 00:50:00.734197 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Mar 7 00:50:00.735670 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Mar 7 00:50:00.735807 kernel: hub 1-0:1.0: USB hub found Mar 7 00:50:00.735910 kernel: hub 1-0:1.0: 4 ports detected Mar 7 00:50:00.735998 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Mar 7 00:50:00.737276 kernel: hub 2-0:1.0: USB hub found Mar 7 00:50:00.739661 kernel: hub 2-0:1.0: 4 ports detected Mar 7 00:50:00.742810 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 7 00:50:00.749206 kernel: sd 0:0:0:1: Power-on or device reset occurred Mar 7 00:50:00.749397 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Mar 7 00:50:00.749488 kernel: sd 0:0:0:1: [sda] Write Protect is off Mar 7 00:50:00.749585 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Mar 7 00:50:00.749676 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Mar 7 00:50:00.755762 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 7 00:50:00.755822 kernel: GPT:17805311 != 80003071 Mar 7 00:50:00.755834 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 7 00:50:00.755844 kernel: GPT:17805311 != 80003071 Mar 7 00:50:00.756655 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 7 00:50:00.756698 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 00:50:00.758591 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Mar 7 00:50:00.772505 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 00:50:00.803782 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (510) Mar 7 00:50:00.805600 kernel: BTRFS: device fsid 237c8587-8110-47ef-99f9-37e4ed4d3b31 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (511) Mar 7 00:50:00.807240 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Mar 7 00:50:00.823580 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Mar 7 00:50:00.831587 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 7 00:50:00.836645 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Mar 7 00:50:00.837306 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Mar 7 00:50:00.845776 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 7 00:50:00.853157 disk-uuid[572]: Primary Header is updated. Mar 7 00:50:00.853157 disk-uuid[572]: Secondary Entries is updated. Mar 7 00:50:00.853157 disk-uuid[572]: Secondary Header is updated. Mar 7 00:50:00.859616 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 00:50:00.978835 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Mar 7 00:50:01.114890 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Mar 7 00:50:01.114955 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Mar 7 00:50:01.115142 kernel: usbcore: registered new interface driver usbhid Mar 7 00:50:01.115583 kernel: usbhid: USB HID core driver Mar 7 00:50:01.223648 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Mar 7 00:50:01.353638 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Mar 7 00:50:01.406686 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Mar 7 00:50:01.871609 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 7 00:50:01.873546 disk-uuid[573]: The operation has completed successfully. Mar 7 00:50:01.923467 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 7 00:50:01.923596 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 7 00:50:01.943848 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 7 00:50:01.949225 sh[591]: Success Mar 7 00:50:01.966338 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 7 00:50:02.028061 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 7 00:50:02.044829 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 7 00:50:02.052401 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 7 00:50:02.065854 kernel: BTRFS info (device dm-0): first mount of filesystem 237c8587-8110-47ef-99f9-37e4ed4d3b31 Mar 7 00:50:02.065916 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 7 00:50:02.065927 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 7 00:50:02.065938 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 7 00:50:02.065948 kernel: BTRFS info (device dm-0): using free space tree Mar 7 00:50:02.072585 kernel: BTRFS info (device dm-0): enabling ssd optimizations Mar 7 00:50:02.074734 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 7 00:50:02.076646 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 7 00:50:02.083869 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 7 00:50:02.088779 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 7 00:50:02.097468 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 00:50:02.097518 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 00:50:02.097540 kernel: BTRFS info (device sda6): using free space tree Mar 7 00:50:02.102616 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 00:50:02.102671 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 00:50:02.112849 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 7 00:50:02.114614 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 00:50:02.120262 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 7 00:50:02.125771 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 7 00:50:02.221533 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 00:50:02.224123 ignition[667]: Ignition 2.19.0 Mar 7 00:50:02.224137 ignition[667]: Stage: fetch-offline Mar 7 00:50:02.224181 ignition[667]: no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:02.224189 ignition[667]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:02.231240 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 00:50:02.224375 ignition[667]: parsed url from cmdline: "" Mar 7 00:50:02.234619 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 00:50:02.224378 ignition[667]: no config URL provided Mar 7 00:50:02.224383 ignition[667]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 00:50:02.224389 ignition[667]: no config at "/usr/lib/ignition/user.ign" Mar 7 00:50:02.224394 ignition[667]: failed to fetch config: resource requires networking Mar 7 00:50:02.224594 ignition[667]: Ignition finished successfully Mar 7 00:50:02.253604 systemd-networkd[778]: lo: Link UP Mar 7 00:50:02.253614 systemd-networkd[778]: lo: Gained carrier Mar 7 00:50:02.255713 systemd-networkd[778]: Enumeration completed Mar 7 00:50:02.255899 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 00:50:02.256644 systemd-networkd[778]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:02.256647 systemd-networkd[778]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 00:50:02.257958 systemd-networkd[778]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:02.257961 systemd-networkd[778]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 00:50:02.258915 systemd-networkd[778]: eth0: Link UP Mar 7 00:50:02.258919 systemd-networkd[778]: eth0: Gained carrier Mar 7 00:50:02.258927 systemd-networkd[778]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:02.261385 systemd[1]: Reached target network.target - Network. Mar 7 00:50:02.266892 systemd-networkd[778]: eth1: Link UP Mar 7 00:50:02.266896 systemd-networkd[778]: eth1: Gained carrier Mar 7 00:50:02.266907 systemd-networkd[778]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:02.270163 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 7 00:50:02.281771 ignition[781]: Ignition 2.19.0 Mar 7 00:50:02.282429 ignition[781]: Stage: fetch Mar 7 00:50:02.283151 ignition[781]: no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:02.283658 ignition[781]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:02.283783 ignition[781]: parsed url from cmdline: "" Mar 7 00:50:02.283787 ignition[781]: no config URL provided Mar 7 00:50:02.283792 ignition[781]: reading system config file "/usr/lib/ignition/user.ign" Mar 7 00:50:02.283801 ignition[781]: no config at "/usr/lib/ignition/user.ign" Mar 7 00:50:02.283822 ignition[781]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Mar 7 00:50:02.284505 ignition[781]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Mar 7 00:50:02.305700 systemd-networkd[778]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Mar 7 00:50:02.317730 systemd-networkd[778]: eth0: DHCPv4 address 188.245.50.81/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 7 00:50:02.484609 ignition[781]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Mar 7 00:50:02.488599 ignition[781]: GET result: OK Mar 7 00:50:02.488710 ignition[781]: parsing config with SHA512: 4bb939e774a0a05ea26d5af83840188cfba8b45e62bcdf0bad17c4319e1025d0631c3643032cbfab565bbde98e38b72f80165bbef437389511859faf82afd3cf Mar 7 00:50:02.493624 unknown[781]: fetched base config from "system" Mar 7 00:50:02.493636 unknown[781]: fetched base config from "system" Mar 7 00:50:02.494176 ignition[781]: fetch: fetch complete Mar 7 00:50:02.493650 unknown[781]: fetched user config from "hetzner" Mar 7 00:50:02.494186 ignition[781]: fetch: fetch passed Mar 7 00:50:02.494242 ignition[781]: Ignition finished successfully Mar 7 00:50:02.499618 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 7 00:50:02.503723 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 7 00:50:02.521406 ignition[788]: Ignition 2.19.0 Mar 7 00:50:02.521417 ignition[788]: Stage: kargs Mar 7 00:50:02.521648 ignition[788]: no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:02.521660 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:02.522737 ignition[788]: kargs: kargs passed Mar 7 00:50:02.522792 ignition[788]: Ignition finished successfully Mar 7 00:50:02.526267 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 7 00:50:02.533860 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 7 00:50:02.556307 ignition[794]: Ignition 2.19.0 Mar 7 00:50:02.556317 ignition[794]: Stage: disks Mar 7 00:50:02.556505 ignition[794]: no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:02.556515 ignition[794]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:02.557488 ignition[794]: disks: disks passed Mar 7 00:50:02.557539 ignition[794]: Ignition finished successfully Mar 7 00:50:02.560613 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 7 00:50:02.561349 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 7 00:50:02.562267 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 7 00:50:02.563328 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 00:50:02.564349 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 00:50:02.565302 systemd[1]: Reached target basic.target - Basic System. Mar 7 00:50:02.578927 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 7 00:50:02.602745 systemd-fsck[802]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Mar 7 00:50:02.607736 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 7 00:50:02.612762 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 7 00:50:02.680607 kernel: EXT4-fs (sda9): mounted filesystem 596a8ea8-9d3d-4d06-a56e-9d3ebd3cb76d r/w with ordered data mode. Quota mode: none. Mar 7 00:50:02.682804 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 7 00:50:02.686071 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 7 00:50:02.696780 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 00:50:02.701868 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 7 00:50:02.708836 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 7 00:50:02.710514 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 7 00:50:02.711996 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 00:50:02.715747 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 7 00:50:02.723747 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 7 00:50:02.729588 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (810) Mar 7 00:50:02.733899 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 00:50:02.733963 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 00:50:02.734600 kernel: BTRFS info (device sda6): using free space tree Mar 7 00:50:02.744696 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 00:50:02.744773 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 00:50:02.751814 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 00:50:02.782169 initrd-setup-root[837]: cut: /sysroot/etc/passwd: No such file or directory Mar 7 00:50:02.785177 coreos-metadata[812]: Mar 07 00:50:02.785 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Mar 7 00:50:02.787775 coreos-metadata[812]: Mar 07 00:50:02.787 INFO Fetch successful Mar 7 00:50:02.789318 coreos-metadata[812]: Mar 07 00:50:02.788 INFO wrote hostname ci-4081-3-6-n-53d1697e27 to /sysroot/etc/hostname Mar 7 00:50:02.790810 initrd-setup-root[844]: cut: /sysroot/etc/group: No such file or directory Mar 7 00:50:02.793660 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 00:50:02.797520 initrd-setup-root[852]: cut: /sysroot/etc/shadow: No such file or directory Mar 7 00:50:02.802147 initrd-setup-root[859]: cut: /sysroot/etc/gshadow: No such file or directory Mar 7 00:50:02.913460 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 7 00:50:02.919819 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 7 00:50:02.923752 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 7 00:50:02.931642 kernel: BTRFS info (device sda6): last unmount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 00:50:02.963616 ignition[927]: INFO : Ignition 2.19.0 Mar 7 00:50:02.963616 ignition[927]: INFO : Stage: mount Mar 7 00:50:02.963616 ignition[927]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:02.963616 ignition[927]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:02.967767 ignition[927]: INFO : mount: mount passed Mar 7 00:50:02.964519 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 7 00:50:02.968860 ignition[927]: INFO : Ignition finished successfully Mar 7 00:50:02.971760 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 7 00:50:02.978723 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 7 00:50:03.066244 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 7 00:50:03.073924 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 7 00:50:03.084633 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (940) Mar 7 00:50:03.086658 kernel: BTRFS info (device sda6): first mount of filesystem 6e876a94-9f11-430e-8016-2af72863cd2e Mar 7 00:50:03.086697 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 7 00:50:03.086718 kernel: BTRFS info (device sda6): using free space tree Mar 7 00:50:03.090604 kernel: BTRFS info (device sda6): enabling ssd optimizations Mar 7 00:50:03.090654 kernel: BTRFS info (device sda6): auto enabling async discard Mar 7 00:50:03.095392 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 7 00:50:03.124434 ignition[958]: INFO : Ignition 2.19.0 Mar 7 00:50:03.125164 ignition[958]: INFO : Stage: files Mar 7 00:50:03.125556 ignition[958]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:03.125556 ignition[958]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:03.126688 ignition[958]: DEBUG : files: compiled without relabeling support, skipping Mar 7 00:50:03.128293 ignition[958]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 7 00:50:03.128293 ignition[958]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 7 00:50:03.134193 ignition[958]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 7 00:50:03.135700 ignition[958]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 7 00:50:03.138198 unknown[958]: wrote ssh authorized keys file for user: core Mar 7 00:50:03.140323 ignition[958]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 7 00:50:03.141066 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 00:50:03.141066 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Mar 7 00:50:03.213733 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 7 00:50:03.299377 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 7 00:50:03.299377 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 7 00:50:03.299377 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Mar 7 00:50:03.654008 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 7 00:50:03.731860 systemd-networkd[778]: eth0: Gained IPv6LL Mar 7 00:50:03.795945 systemd-networkd[778]: eth1: Gained IPv6LL Mar 7 00:50:03.902805 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 7 00:50:03.902805 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Mar 7 00:50:03.907348 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.4-arm64.raw: attempt #1 Mar 7 00:50:04.286711 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 7 00:50:05.236930 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.4-arm64.raw" Mar 7 00:50:05.236930 ignition[958]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Mar 7 00:50:05.242743 ignition[958]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 7 00:50:05.242743 ignition[958]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 7 00:50:05.242743 ignition[958]: INFO : files: files passed Mar 7 00:50:05.242743 ignition[958]: INFO : Ignition finished successfully Mar 7 00:50:05.241272 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 7 00:50:05.250690 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 7 00:50:05.257774 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 7 00:50:05.258621 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 7 00:50:05.258741 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 7 00:50:05.286317 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 00:50:05.286317 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 7 00:50:05.289500 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 7 00:50:05.290842 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 00:50:05.292929 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 7 00:50:05.298818 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 7 00:50:05.322864 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 7 00:50:05.323685 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 7 00:50:05.325056 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 7 00:50:05.325937 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 7 00:50:05.327987 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 7 00:50:05.336854 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 7 00:50:05.349439 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 00:50:05.359806 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 7 00:50:05.373451 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 7 00:50:05.375041 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 00:50:05.376484 systemd[1]: Stopped target timers.target - Timer Units. Mar 7 00:50:05.377642 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 7 00:50:05.378307 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 7 00:50:05.379205 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 7 00:50:05.381852 systemd[1]: Stopped target basic.target - Basic System. Mar 7 00:50:05.383479 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 7 00:50:05.384585 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 7 00:50:05.385839 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 7 00:50:05.386982 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 7 00:50:05.388084 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 7 00:50:05.389236 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 7 00:50:05.390325 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 7 00:50:05.391269 systemd[1]: Stopped target swap.target - Swaps. Mar 7 00:50:05.392131 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 7 00:50:05.392248 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 7 00:50:05.393534 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 7 00:50:05.394282 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 00:50:05.395314 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 7 00:50:05.395384 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 00:50:05.396421 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 7 00:50:05.396533 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 7 00:50:05.398050 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 7 00:50:05.398161 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 7 00:50:05.399490 systemd[1]: ignition-files.service: Deactivated successfully. Mar 7 00:50:05.399595 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 7 00:50:05.400483 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 7 00:50:05.400590 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 7 00:50:05.410901 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 7 00:50:05.411940 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 7 00:50:05.412181 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 00:50:05.417747 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 7 00:50:05.418231 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 7 00:50:05.418346 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 00:50:05.419722 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 7 00:50:05.419845 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 7 00:50:05.426842 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 7 00:50:05.427013 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 7 00:50:05.436754 ignition[1010]: INFO : Ignition 2.19.0 Mar 7 00:50:05.436754 ignition[1010]: INFO : Stage: umount Mar 7 00:50:05.437779 ignition[1010]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 7 00:50:05.437779 ignition[1010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Mar 7 00:50:05.439549 ignition[1010]: INFO : umount: umount passed Mar 7 00:50:05.442793 ignition[1010]: INFO : Ignition finished successfully Mar 7 00:50:05.443384 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 7 00:50:05.445161 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 7 00:50:05.445802 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 7 00:50:05.447297 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 7 00:50:05.447352 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 7 00:50:05.447960 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 7 00:50:05.448000 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 7 00:50:05.448893 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 7 00:50:05.448931 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 7 00:50:05.450987 systemd[1]: Stopped target network.target - Network. Mar 7 00:50:05.452134 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 7 00:50:05.452189 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 7 00:50:05.452893 systemd[1]: Stopped target paths.target - Path Units. Mar 7 00:50:05.453458 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 7 00:50:05.456620 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 00:50:05.457769 systemd[1]: Stopped target slices.target - Slice Units. Mar 7 00:50:05.458894 systemd[1]: Stopped target sockets.target - Socket Units. Mar 7 00:50:05.459766 systemd[1]: iscsid.socket: Deactivated successfully. Mar 7 00:50:05.459811 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 7 00:50:05.461876 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 7 00:50:05.461948 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 7 00:50:05.463414 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 7 00:50:05.463463 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 7 00:50:05.464347 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 7 00:50:05.464387 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 7 00:50:05.465344 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 7 00:50:05.466252 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 7 00:50:05.467363 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 7 00:50:05.467463 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 7 00:50:05.468564 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 7 00:50:05.468843 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 7 00:50:05.472687 systemd-networkd[778]: eth1: DHCPv6 lease lost Mar 7 00:50:05.474812 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 7 00:50:05.474921 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 7 00:50:05.477947 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 7 00:50:05.478081 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 00:50:05.478847 systemd-networkd[778]: eth0: DHCPv6 lease lost Mar 7 00:50:05.480882 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 7 00:50:05.482658 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 7 00:50:05.485788 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 7 00:50:05.485830 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 7 00:50:05.490816 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 7 00:50:05.491376 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 7 00:50:05.491470 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 7 00:50:05.494225 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 7 00:50:05.494283 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 7 00:50:05.498420 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 7 00:50:05.498483 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 7 00:50:05.500274 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 00:50:05.511982 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 7 00:50:05.512127 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 7 00:50:05.514963 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 7 00:50:05.515112 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 00:50:05.516879 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 7 00:50:05.516942 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 7 00:50:05.518128 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 7 00:50:05.518159 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 00:50:05.519216 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 7 00:50:05.519262 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 7 00:50:05.520642 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 7 00:50:05.520685 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 7 00:50:05.522226 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 7 00:50:05.522276 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 7 00:50:05.528800 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 7 00:50:05.529336 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 7 00:50:05.529388 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 00:50:05.531283 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 7 00:50:05.531326 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 00:50:05.539040 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 7 00:50:05.539162 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 7 00:50:05.540382 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 7 00:50:05.542173 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 7 00:50:05.576058 systemd[1]: Switching root. Mar 7 00:50:05.618666 systemd-journald[237]: Journal stopped Mar 7 00:50:06.528219 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). Mar 7 00:50:06.528279 kernel: SELinux: policy capability network_peer_controls=1 Mar 7 00:50:06.528291 kernel: SELinux: policy capability open_perms=1 Mar 7 00:50:06.528305 kernel: SELinux: policy capability extended_socket_class=1 Mar 7 00:50:06.528318 kernel: SELinux: policy capability always_check_network=0 Mar 7 00:50:06.528331 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 7 00:50:06.528341 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 7 00:50:06.528355 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 7 00:50:06.528365 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 7 00:50:06.528375 kernel: audit: type=1403 audit(1772844605.786:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 7 00:50:06.528386 systemd[1]: Successfully loaded SELinux policy in 34.728ms. Mar 7 00:50:06.528410 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.683ms. Mar 7 00:50:06.528421 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 7 00:50:06.528434 systemd[1]: Detected virtualization kvm. Mar 7 00:50:06.528445 systemd[1]: Detected architecture arm64. Mar 7 00:50:06.528455 systemd[1]: Detected first boot. Mar 7 00:50:06.528466 systemd[1]: Hostname set to . Mar 7 00:50:06.528477 systemd[1]: Initializing machine ID from VM UUID. Mar 7 00:50:06.528487 zram_generator::config[1052]: No configuration found. Mar 7 00:50:06.528498 systemd[1]: Populated /etc with preset unit settings. Mar 7 00:50:06.528509 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 7 00:50:06.528521 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 7 00:50:06.528532 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 7 00:50:06.528543 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 7 00:50:06.528553 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 7 00:50:06.528564 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 7 00:50:06.532907 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 7 00:50:06.532921 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 7 00:50:06.532932 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 7 00:50:06.532950 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 7 00:50:06.532961 systemd[1]: Created slice user.slice - User and Session Slice. Mar 7 00:50:06.532972 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 7 00:50:06.532983 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 7 00:50:06.533025 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 7 00:50:06.533042 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 7 00:50:06.533054 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 7 00:50:06.533065 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 7 00:50:06.533076 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Mar 7 00:50:06.533090 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 7 00:50:06.533101 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 7 00:50:06.533119 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 7 00:50:06.533129 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 7 00:50:06.533139 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 7 00:50:06.533150 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 7 00:50:06.533162 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 7 00:50:06.533174 systemd[1]: Reached target slices.target - Slice Units. Mar 7 00:50:06.533185 systemd[1]: Reached target swap.target - Swaps. Mar 7 00:50:06.533196 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 7 00:50:06.533206 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 7 00:50:06.533217 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 7 00:50:06.533228 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 7 00:50:06.533238 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 7 00:50:06.533248 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 7 00:50:06.533259 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 7 00:50:06.533271 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 7 00:50:06.533282 systemd[1]: Mounting media.mount - External Media Directory... Mar 7 00:50:06.533293 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 7 00:50:06.533304 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 7 00:50:06.533314 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 7 00:50:06.533325 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 7 00:50:06.533335 systemd[1]: Reached target machines.target - Containers. Mar 7 00:50:06.533350 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 7 00:50:06.533363 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 00:50:06.533377 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 7 00:50:06.533390 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 7 00:50:06.533407 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 00:50:06.533419 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 00:50:06.533430 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 00:50:06.533442 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 7 00:50:06.533453 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 00:50:06.533464 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 7 00:50:06.533475 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 7 00:50:06.533485 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 7 00:50:06.533496 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 7 00:50:06.533506 systemd[1]: Stopped systemd-fsck-usr.service. Mar 7 00:50:06.533517 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 7 00:50:06.533529 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 7 00:50:06.533540 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 7 00:50:06.533551 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 7 00:50:06.533562 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 7 00:50:06.536598 systemd[1]: verity-setup.service: Deactivated successfully. Mar 7 00:50:06.536621 systemd[1]: Stopped verity-setup.service. Mar 7 00:50:06.536632 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 7 00:50:06.536643 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 7 00:50:06.536661 systemd[1]: Mounted media.mount - External Media Directory. Mar 7 00:50:06.536672 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 7 00:50:06.536683 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 7 00:50:06.536694 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 7 00:50:06.536704 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 7 00:50:06.536715 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 7 00:50:06.536727 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 7 00:50:06.536766 systemd-journald[1119]: Collecting audit messages is disabled. Mar 7 00:50:06.536795 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 00:50:06.536805 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 00:50:06.536817 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 00:50:06.536827 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 00:50:06.536838 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 7 00:50:06.536853 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 7 00:50:06.536864 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 7 00:50:06.536875 systemd-journald[1119]: Journal started Mar 7 00:50:06.536897 systemd-journald[1119]: Runtime Journal (/run/log/journal/54655f3653604e79b74c22d6e530746d) is 8.0M, max 76.6M, 68.6M free. Mar 7 00:50:06.271163 systemd[1]: Queued start job for default target multi-user.target. Mar 7 00:50:06.293975 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 7 00:50:06.294479 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 7 00:50:06.543209 kernel: fuse: init (API version 7.39) Mar 7 00:50:06.543266 systemd[1]: Started systemd-journald.service - Journal Service. Mar 7 00:50:06.550562 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 7 00:50:06.552619 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 7 00:50:06.553815 kernel: loop: module loaded Mar 7 00:50:06.554389 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 7 00:50:06.563835 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 7 00:50:06.568595 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 7 00:50:06.569334 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 7 00:50:06.569440 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 7 00:50:06.571275 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 7 00:50:06.579319 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 7 00:50:06.589653 kernel: ACPI: bus type drm_connector registered Mar 7 00:50:06.592751 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 7 00:50:06.593392 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 00:50:06.596747 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 7 00:50:06.602955 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 7 00:50:06.603619 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 00:50:06.604796 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 7 00:50:06.610756 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 00:50:06.616271 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 7 00:50:06.620521 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 7 00:50:06.623940 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 00:50:06.624097 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 00:50:06.624901 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 00:50:06.625078 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 00:50:06.625811 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 7 00:50:06.626502 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 7 00:50:06.633023 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 7 00:50:06.644938 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 00:50:06.650744 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 7 00:50:06.670753 systemd-journald[1119]: Time spent on flushing to /var/log/journal/54655f3653604e79b74c22d6e530746d is 97.848ms for 1126 entries. Mar 7 00:50:06.670753 systemd-journald[1119]: System Journal (/var/log/journal/54655f3653604e79b74c22d6e530746d) is 8.0M, max 584.8M, 576.8M free. Mar 7 00:50:06.780240 systemd-journald[1119]: Received client request to flush runtime journal. Mar 7 00:50:06.780296 kernel: loop0: detected capacity change from 0 to 114328 Mar 7 00:50:06.780324 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 7 00:50:06.780337 kernel: loop1: detected capacity change from 0 to 200864 Mar 7 00:50:06.686332 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 7 00:50:06.687281 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 7 00:50:06.703291 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 7 00:50:06.734644 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 00:50:06.746141 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 7 00:50:06.759134 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 7 00:50:06.784637 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 7 00:50:06.799630 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 7 00:50:06.802613 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 7 00:50:06.804187 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 7 00:50:06.817155 kernel: loop2: detected capacity change from 0 to 114432 Mar 7 00:50:06.817875 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 7 00:50:06.821109 udevadm[1180]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Mar 7 00:50:06.863600 kernel: loop3: detected capacity change from 0 to 8 Mar 7 00:50:06.873720 systemd-tmpfiles[1186]: ACLs are not supported, ignoring. Mar 7 00:50:06.873735 systemd-tmpfiles[1186]: ACLs are not supported, ignoring. Mar 7 00:50:06.886777 kernel: loop4: detected capacity change from 0 to 114328 Mar 7 00:50:06.885615 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 7 00:50:06.898478 kernel: loop5: detected capacity change from 0 to 200864 Mar 7 00:50:06.916594 kernel: loop6: detected capacity change from 0 to 114432 Mar 7 00:50:06.929643 kernel: loop7: detected capacity change from 0 to 8 Mar 7 00:50:06.930977 (sd-merge)[1191]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Mar 7 00:50:06.932140 (sd-merge)[1191]: Merged extensions into '/usr'. Mar 7 00:50:06.939810 systemd[1]: Reloading requested from client PID 1163 ('systemd-sysext') (unit systemd-sysext.service)... Mar 7 00:50:06.939846 systemd[1]: Reloading... Mar 7 00:50:07.092596 zram_generator::config[1219]: No configuration found. Mar 7 00:50:07.146658 ldconfig[1158]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 7 00:50:07.222495 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 00:50:07.269668 systemd[1]: Reloading finished in 328 ms. Mar 7 00:50:07.312395 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 7 00:50:07.314664 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 7 00:50:07.329200 systemd[1]: Starting ensure-sysext.service... Mar 7 00:50:07.333776 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 7 00:50:07.341138 systemd[1]: Reloading requested from client PID 1255 ('systemctl') (unit ensure-sysext.service)... Mar 7 00:50:07.341151 systemd[1]: Reloading... Mar 7 00:50:07.366134 systemd-tmpfiles[1256]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 7 00:50:07.368334 systemd-tmpfiles[1256]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 7 00:50:07.372241 systemd-tmpfiles[1256]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 7 00:50:07.373868 systemd-tmpfiles[1256]: ACLs are not supported, ignoring. Mar 7 00:50:07.373955 systemd-tmpfiles[1256]: ACLs are not supported, ignoring. Mar 7 00:50:07.378320 systemd-tmpfiles[1256]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 00:50:07.378335 systemd-tmpfiles[1256]: Skipping /boot Mar 7 00:50:07.387973 systemd-tmpfiles[1256]: Detected autofs mount point /boot during canonicalization of boot. Mar 7 00:50:07.388037 systemd-tmpfiles[1256]: Skipping /boot Mar 7 00:50:07.421598 zram_generator::config[1283]: No configuration found. Mar 7 00:50:07.526349 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 00:50:07.573137 systemd[1]: Reloading finished in 231 ms. Mar 7 00:50:07.595289 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 7 00:50:07.609461 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 7 00:50:07.629430 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 00:50:07.633869 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 7 00:50:07.644210 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 7 00:50:07.650923 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 7 00:50:07.659853 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 7 00:50:07.670029 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 7 00:50:07.675048 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 00:50:07.677410 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 00:50:07.680854 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 00:50:07.683807 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 00:50:07.684943 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 00:50:07.689528 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 7 00:50:07.692463 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 7 00:50:07.698200 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 00:50:07.698360 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 00:50:07.699777 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 00:50:07.699896 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 00:50:07.711398 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 00:50:07.717177 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 00:50:07.720105 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 00:50:07.720771 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 00:50:07.722665 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 7 00:50:07.724608 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 00:50:07.725687 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 00:50:07.733808 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 00:50:07.743142 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 7 00:50:07.748607 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 00:50:07.750678 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 00:50:07.752408 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 7 00:50:07.762423 systemd[1]: Finished ensure-sysext.service. Mar 7 00:50:07.768661 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 00:50:07.769657 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 00:50:07.775812 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 00:50:07.781947 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 7 00:50:07.786164 augenrules[1357]: No rules Mar 7 00:50:07.786950 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 7 00:50:07.790289 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 00:50:07.795364 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 00:50:07.795499 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 00:50:07.796871 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 7 00:50:07.797079 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 7 00:50:07.802386 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 7 00:50:07.818885 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 00:50:07.819872 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 00:50:07.821646 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 7 00:50:07.824628 systemd-udevd[1333]: Using default interface naming scheme 'v255'. Mar 7 00:50:07.825087 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 00:50:07.825122 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 7 00:50:07.858635 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 7 00:50:07.863901 systemd-resolved[1332]: Positive Trust Anchors: Mar 7 00:50:07.863917 systemd-resolved[1332]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 7 00:50:07.863949 systemd-resolved[1332]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 7 00:50:07.871474 systemd-resolved[1332]: Using system hostname 'ci-4081-3-6-n-53d1697e27'. Mar 7 00:50:07.875917 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 7 00:50:07.876604 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 7 00:50:07.878738 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 7 00:50:07.916398 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 7 00:50:07.917650 systemd[1]: Reached target time-set.target - System Time Set. Mar 7 00:50:07.949362 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Mar 7 00:50:07.954886 systemd-networkd[1380]: lo: Link UP Mar 7 00:50:07.954898 systemd-networkd[1380]: lo: Gained carrier Mar 7 00:50:07.993296 systemd-networkd[1380]: Enumeration completed Mar 7 00:50:07.993393 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 7 00:50:07.994742 systemd[1]: Reached target network.target - Network. Mar 7 00:50:07.996633 systemd-networkd[1380]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:07.996637 systemd-networkd[1380]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 00:50:07.998756 systemd-networkd[1380]: eth1: Link UP Mar 7 00:50:07.998764 systemd-networkd[1380]: eth1: Gained carrier Mar 7 00:50:07.998780 systemd-networkd[1380]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:08.004892 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 7 00:50:08.015770 systemd-networkd[1380]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:08.032720 systemd-networkd[1380]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Mar 7 00:50:08.033546 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:08.041642 kernel: mousedev: PS/2 mouse device common for all mice Mar 7 00:50:08.053638 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1390) Mar 7 00:50:08.119894 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Mar 7 00:50:08.119969 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Mar 7 00:50:08.120026 kernel: [drm] features: -context_init Mar 7 00:50:08.122509 systemd-networkd[1380]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:08.122522 systemd-networkd[1380]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 7 00:50:08.123585 kernel: [drm] number of scanouts: 1 Mar 7 00:50:08.123640 kernel: [drm] number of cap sets: 0 Mar 7 00:50:08.123887 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:08.124125 systemd-networkd[1380]: eth0: Link UP Mar 7 00:50:08.124134 systemd-networkd[1380]: eth0: Gained carrier Mar 7 00:50:08.124151 systemd-networkd[1380]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 7 00:50:08.128621 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Mar 7 00:50:08.129672 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Mar 7 00:50:08.129796 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 7 00:50:08.130837 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:08.138011 kernel: Console: switching to colour frame buffer device 160x50 Mar 7 00:50:08.141188 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 7 00:50:08.149290 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Mar 7 00:50:08.176970 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 7 00:50:08.182257 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 7 00:50:08.184115 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 7 00:50:08.184158 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 7 00:50:08.184553 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 7 00:50:08.185792 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 7 00:50:08.200947 systemd-networkd[1380]: eth0: DHCPv4 address 188.245.50.81/32, gateway 172.31.1.1 acquired from 172.31.1.1 Mar 7 00:50:08.203705 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:08.205205 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:08.207880 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Mar 7 00:50:08.218415 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 7 00:50:08.219404 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 7 00:50:08.222391 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 7 00:50:08.224629 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 7 00:50:08.235083 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 7 00:50:08.236773 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 7 00:50:08.237042 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 7 00:50:08.241557 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 7 00:50:08.255615 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 7 00:50:08.316024 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 7 00:50:08.337048 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 7 00:50:08.346869 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 7 00:50:08.373134 lvm[1438]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 00:50:08.404669 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 7 00:50:08.406249 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 7 00:50:08.407793 systemd[1]: Reached target sysinit.target - System Initialization. Mar 7 00:50:08.408900 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 7 00:50:08.410086 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 7 00:50:08.411690 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 7 00:50:08.412787 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 7 00:50:08.413969 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 7 00:50:08.414775 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 7 00:50:08.414814 systemd[1]: Reached target paths.target - Path Units. Mar 7 00:50:08.415304 systemd[1]: Reached target timers.target - Timer Units. Mar 7 00:50:08.418642 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 7 00:50:08.420911 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 7 00:50:08.426976 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 7 00:50:08.429239 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 7 00:50:08.430659 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 7 00:50:08.431491 systemd[1]: Reached target sockets.target - Socket Units. Mar 7 00:50:08.432369 systemd[1]: Reached target basic.target - Basic System. Mar 7 00:50:08.432941 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 7 00:50:08.432970 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 7 00:50:08.435698 systemd[1]: Starting containerd.service - containerd container runtime... Mar 7 00:50:08.439872 lvm[1442]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 7 00:50:08.441781 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 7 00:50:08.444777 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 7 00:50:08.448807 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 7 00:50:08.454756 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 7 00:50:08.455513 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 7 00:50:08.459784 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 7 00:50:08.466684 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 7 00:50:08.468736 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Mar 7 00:50:08.475400 jq[1446]: false Mar 7 00:50:08.483789 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 7 00:50:08.488857 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 7 00:50:08.494787 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 7 00:50:08.497930 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 7 00:50:08.498429 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 7 00:50:08.501103 dbus-daemon[1445]: [system] SELinux support is enabled Mar 7 00:50:08.512902 systemd[1]: Starting update-engine.service - Update Engine... Mar 7 00:50:08.517807 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 7 00:50:08.518947 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 7 00:50:08.521553 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 7 00:50:08.524656 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 7 00:50:08.524861 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 7 00:50:08.530067 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 7 00:50:08.530232 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 7 00:50:08.535508 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 7 00:50:08.535562 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 7 00:50:08.537306 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 7 00:50:08.537325 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 7 00:50:08.554742 update_engine[1456]: I20260307 00:50:08.553802 1456 main.cc:92] Flatcar Update Engine starting Mar 7 00:50:08.556274 extend-filesystems[1448]: Found loop4 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found loop5 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found loop6 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found loop7 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda1 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda2 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda3 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found usr Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda4 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda6 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda7 Mar 7 00:50:08.556274 extend-filesystems[1448]: Found sda9 Mar 7 00:50:08.556274 extend-filesystems[1448]: Checking size of /dev/sda9 Mar 7 00:50:08.582921 coreos-metadata[1444]: Mar 07 00:50:08.563 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Mar 7 00:50:08.582921 coreos-metadata[1444]: Mar 07 00:50:08.564 INFO Fetch successful Mar 7 00:50:08.582921 coreos-metadata[1444]: Mar 07 00:50:08.564 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Mar 7 00:50:08.582921 coreos-metadata[1444]: Mar 07 00:50:08.565 INFO Fetch successful Mar 7 00:50:08.583112 update_engine[1456]: I20260307 00:50:08.559618 1456 update_check_scheduler.cc:74] Next update check in 7m23s Mar 7 00:50:08.559483 systemd[1]: Started update-engine.service - Update Engine. Mar 7 00:50:08.573769 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 7 00:50:08.592040 extend-filesystems[1448]: Resized partition /dev/sda9 Mar 7 00:50:08.596424 systemd[1]: motdgen.service: Deactivated successfully. Mar 7 00:50:08.599652 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 7 00:50:08.601121 jq[1460]: true Mar 7 00:50:08.602077 tar[1462]: linux-arm64/LICENSE Mar 7 00:50:08.602077 tar[1462]: linux-arm64/helm Mar 7 00:50:08.605776 extend-filesystems[1489]: resize2fs 1.47.1 (20-May-2024) Mar 7 00:50:08.617111 (ntainerd)[1484]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 7 00:50:08.619578 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Mar 7 00:50:08.621598 jq[1490]: true Mar 7 00:50:08.708358 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 7 00:50:08.713498 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 7 00:50:08.724206 systemd-logind[1455]: New seat seat0. Mar 7 00:50:08.732389 systemd-logind[1455]: Watching system buttons on /dev/input/event0 (Power Button) Mar 7 00:50:08.732413 systemd-logind[1455]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Mar 7 00:50:08.732671 systemd[1]: Started systemd-logind.service - User Login Management. Mar 7 00:50:08.753352 bash[1513]: Updated "/home/core/.ssh/authorized_keys" Mar 7 00:50:08.754599 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 7 00:50:08.765593 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1379) Mar 7 00:50:08.767877 systemd[1]: Starting sshkeys.service... Mar 7 00:50:08.786589 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Mar 7 00:50:08.805190 extend-filesystems[1489]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Mar 7 00:50:08.805190 extend-filesystems[1489]: old_desc_blocks = 1, new_desc_blocks = 5 Mar 7 00:50:08.805190 extend-filesystems[1489]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Mar 7 00:50:08.822340 extend-filesystems[1448]: Resized filesystem in /dev/sda9 Mar 7 00:50:08.822340 extend-filesystems[1448]: Found sr0 Mar 7 00:50:08.808474 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 7 00:50:08.808684 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 7 00:50:08.834201 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 7 00:50:08.848018 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 7 00:50:08.898581 coreos-metadata[1525]: Mar 07 00:50:08.896 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Mar 7 00:50:08.901829 coreos-metadata[1525]: Mar 07 00:50:08.901 INFO Fetch successful Mar 7 00:50:08.906702 unknown[1525]: wrote ssh authorized keys file for user: core Mar 7 00:50:08.934841 containerd[1484]: time="2026-03-07T00:50:08.934663120Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Mar 7 00:50:08.942812 update-ssh-keys[1530]: Updated "/home/core/.ssh/authorized_keys" Mar 7 00:50:08.945220 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 7 00:50:08.950912 systemd[1]: Finished sshkeys.service. Mar 7 00:50:08.991221 locksmithd[1478]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 7 00:50:08.998858 containerd[1484]: time="2026-03-07T00:50:08.998609240Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.003778 containerd[1484]: time="2026-03-07T00:50:09.003727560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 7 00:50:09.003901 containerd[1484]: time="2026-03-07T00:50:09.003886440Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 7 00:50:09.004377 containerd[1484]: time="2026-03-07T00:50:09.003950080Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 7 00:50:09.004377 containerd[1484]: time="2026-03-07T00:50:09.004199400Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 7 00:50:09.004377 containerd[1484]: time="2026-03-07T00:50:09.004221360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.004377 containerd[1484]: time="2026-03-07T00:50:09.004291200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 00:50:09.004377 containerd[1484]: time="2026-03-07T00:50:09.004307360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.005765480Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.005795520Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.005811560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.005823360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.005937080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006233 containerd[1484]: time="2026-03-07T00:50:09.006179880Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006544 containerd[1484]: time="2026-03-07T00:50:09.006521080Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 7 00:50:09.006636 containerd[1484]: time="2026-03-07T00:50:09.006618720Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 7 00:50:09.006805 containerd[1484]: time="2026-03-07T00:50:09.006785720Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 7 00:50:09.006919 containerd[1484]: time="2026-03-07T00:50:09.006901680Z" level=info msg="metadata content store policy set" policy=shared Mar 7 00:50:09.014921 containerd[1484]: time="2026-03-07T00:50:09.014751480Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 7 00:50:09.014921 containerd[1484]: time="2026-03-07T00:50:09.014815120Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 7 00:50:09.014921 containerd[1484]: time="2026-03-07T00:50:09.014831960Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 7 00:50:09.014921 containerd[1484]: time="2026-03-07T00:50:09.014849160Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 7 00:50:09.014921 containerd[1484]: time="2026-03-07T00:50:09.014863000Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 7 00:50:09.015271 containerd[1484]: time="2026-03-07T00:50:09.015252320Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 7 00:50:09.015910 containerd[1484]: time="2026-03-07T00:50:09.015875360Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 7 00:50:09.016078 containerd[1484]: time="2026-03-07T00:50:09.016054560Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 7 00:50:09.016110 containerd[1484]: time="2026-03-07T00:50:09.016079480Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 7 00:50:09.016110 containerd[1484]: time="2026-03-07T00:50:09.016093120Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 7 00:50:09.016110 containerd[1484]: time="2026-03-07T00:50:09.016106840Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016120080Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016133760Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016147240Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016161280Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016173840Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016185 containerd[1484]: time="2026-03-07T00:50:09.016185640Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016197000Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016217080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016232640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016244280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016258120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016283 containerd[1484]: time="2026-03-07T00:50:09.016271160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016284640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016297800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016309920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016321640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016340520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016352840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016363680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016386 containerd[1484]: time="2026-03-07T00:50:09.016374920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016521 containerd[1484]: time="2026-03-07T00:50:09.016393880Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 7 00:50:09.016521 containerd[1484]: time="2026-03-07T00:50:09.016413960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016521 containerd[1484]: time="2026-03-07T00:50:09.016426800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.016521 containerd[1484]: time="2026-03-07T00:50:09.016437200Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018072920Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018121640Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018135920Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018147640Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018157280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018175600Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018186600Z" level=info msg="NRI interface is disabled by configuration." Mar 7 00:50:09.020855 containerd[1484]: time="2026-03-07T00:50:09.018196560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.018471240Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.018540840Z" level=info msg="Connect containerd service" Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.018664160Z" level=info msg="using legacy CRI server" Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.018672240Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.018756720Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 7 00:50:09.021125 containerd[1484]: time="2026-03-07T00:50:09.019424240Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 7 00:50:09.022939 containerd[1484]: time="2026-03-07T00:50:09.022892120Z" level=info msg="Start subscribing containerd event" Mar 7 00:50:09.023105 containerd[1484]: time="2026-03-07T00:50:09.023088440Z" level=info msg="Start recovering state" Mar 7 00:50:09.023224 containerd[1484]: time="2026-03-07T00:50:09.023209120Z" level=info msg="Start event monitor" Mar 7 00:50:09.023350 containerd[1484]: time="2026-03-07T00:50:09.023332560Z" level=info msg="Start snapshots syncer" Mar 7 00:50:09.023620 containerd[1484]: time="2026-03-07T00:50:09.023603480Z" level=info msg="Start cni network conf syncer for default" Mar 7 00:50:09.023671 containerd[1484]: time="2026-03-07T00:50:09.023660640Z" level=info msg="Start streaming server" Mar 7 00:50:09.028910 containerd[1484]: time="2026-03-07T00:50:09.028867240Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 7 00:50:09.029463 containerd[1484]: time="2026-03-07T00:50:09.029439600Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 7 00:50:09.029716 systemd[1]: Started containerd.service - containerd container runtime. Mar 7 00:50:09.031234 containerd[1484]: time="2026-03-07T00:50:09.030714600Z" level=info msg="containerd successfully booted in 0.097157s" Mar 7 00:50:09.207389 sshd_keygen[1477]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 7 00:50:09.236066 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 7 00:50:09.247849 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 7 00:50:09.254437 systemd[1]: issuegen.service: Deactivated successfully. Mar 7 00:50:09.254693 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 7 00:50:09.262179 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 7 00:50:09.279301 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 7 00:50:09.287010 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 7 00:50:09.288969 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Mar 7 00:50:09.290045 systemd[1]: Reached target getty.target - Login Prompts. Mar 7 00:50:09.333669 tar[1462]: linux-arm64/README.md Mar 7 00:50:09.347035 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 7 00:50:09.491789 systemd-networkd[1380]: eth1: Gained IPv6LL Mar 7 00:50:09.492923 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:09.497338 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 7 00:50:09.498992 systemd[1]: Reached target network-online.target - Network is Online. Mar 7 00:50:09.509952 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:50:09.514482 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 7 00:50:09.538647 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 7 00:50:10.196214 systemd-networkd[1380]: eth0: Gained IPv6LL Mar 7 00:50:10.196701 systemd-timesyncd[1360]: Network configuration changed, trying to establish connection. Mar 7 00:50:10.259823 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:50:10.261391 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 7 00:50:10.263198 (kubelet)[1575]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:50:10.267609 systemd[1]: Startup finished in 757ms (kernel) + 6.095s (initrd) + 4.514s (userspace) = 11.368s. Mar 7 00:50:10.723796 kubelet[1575]: E0307 00:50:10.723723 1575 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:50:10.727053 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:50:10.727311 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:50:20.978039 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 7 00:50:20.988314 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:50:21.114838 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:50:21.120045 (kubelet)[1594]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:50:21.164348 kubelet[1594]: E0307 00:50:21.164275 1594 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:50:21.168896 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:50:21.169189 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:50:31.397930 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 7 00:50:31.414971 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:50:31.539876 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:50:31.555184 (kubelet)[1609]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:50:31.603148 kubelet[1609]: E0307 00:50:31.603065 1609 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:50:31.607459 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:50:31.607760 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:50:40.493421 systemd-timesyncd[1360]: Contacted time server 78.47.56.71:123 (2.flatcar.pool.ntp.org). Mar 7 00:50:40.493511 systemd-timesyncd[1360]: Initial clock synchronization to Sat 2026-03-07 00:50:40.787016 UTC. Mar 7 00:50:41.397929 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 7 00:50:41.406066 systemd[1]: Started sshd@0-188.245.50.81:22-20.161.92.111:35712.service - OpenSSH per-connection server daemon (20.161.92.111:35712). Mar 7 00:50:41.650060 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 7 00:50:41.660526 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:50:41.777122 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:50:41.782219 (kubelet)[1628]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:50:41.829885 kubelet[1628]: E0307 00:50:41.829840 1628 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:50:41.832903 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:50:41.833120 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:50:42.017330 sshd[1618]: Accepted publickey for core from 20.161.92.111 port 35712 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:42.020504 sshd[1618]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:42.031591 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 7 00:50:42.043093 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 7 00:50:42.049466 systemd-logind[1455]: New session 1 of user core. Mar 7 00:50:42.056238 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 7 00:50:42.062986 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 7 00:50:42.081235 (systemd)[1637]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 7 00:50:42.190443 systemd[1637]: Queued start job for default target default.target. Mar 7 00:50:42.199854 systemd[1637]: Created slice app.slice - User Application Slice. Mar 7 00:50:42.199908 systemd[1637]: Reached target paths.target - Paths. Mar 7 00:50:42.199940 systemd[1637]: Reached target timers.target - Timers. Mar 7 00:50:42.202441 systemd[1637]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 7 00:50:42.218145 systemd[1637]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 7 00:50:42.218464 systemd[1637]: Reached target sockets.target - Sockets. Mar 7 00:50:42.218510 systemd[1637]: Reached target basic.target - Basic System. Mar 7 00:50:42.218703 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 7 00:50:42.220697 systemd[1637]: Reached target default.target - Main User Target. Mar 7 00:50:42.221034 systemd[1637]: Startup finished in 132ms. Mar 7 00:50:42.226930 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 7 00:50:42.680523 systemd[1]: Started sshd@1-188.245.50.81:22-20.161.92.111:35726.service - OpenSSH per-connection server daemon (20.161.92.111:35726). Mar 7 00:50:43.283342 sshd[1648]: Accepted publickey for core from 20.161.92.111 port 35726 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:43.284942 sshd[1648]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:43.291634 systemd-logind[1455]: New session 2 of user core. Mar 7 00:50:43.302017 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 7 00:50:43.714869 sshd[1648]: pam_unix(sshd:session): session closed for user core Mar 7 00:50:43.720535 systemd[1]: sshd@1-188.245.50.81:22-20.161.92.111:35726.service: Deactivated successfully. Mar 7 00:50:43.722819 systemd[1]: session-2.scope: Deactivated successfully. Mar 7 00:50:43.723570 systemd-logind[1455]: Session 2 logged out. Waiting for processes to exit. Mar 7 00:50:43.725260 systemd-logind[1455]: Removed session 2. Mar 7 00:50:43.828122 systemd[1]: Started sshd@2-188.245.50.81:22-20.161.92.111:35736.service - OpenSSH per-connection server daemon (20.161.92.111:35736). Mar 7 00:50:44.425520 sshd[1655]: Accepted publickey for core from 20.161.92.111 port 35736 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:44.428139 sshd[1655]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:44.432767 systemd-logind[1455]: New session 3 of user core. Mar 7 00:50:44.440921 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 7 00:50:44.851095 sshd[1655]: pam_unix(sshd:session): session closed for user core Mar 7 00:50:44.857963 systemd[1]: sshd@2-188.245.50.81:22-20.161.92.111:35736.service: Deactivated successfully. Mar 7 00:50:44.860470 systemd[1]: session-3.scope: Deactivated successfully. Mar 7 00:50:44.861493 systemd-logind[1455]: Session 3 logged out. Waiting for processes to exit. Mar 7 00:50:44.862814 systemd-logind[1455]: Removed session 3. Mar 7 00:50:44.960936 systemd[1]: Started sshd@3-188.245.50.81:22-20.161.92.111:35746.service - OpenSSH per-connection server daemon (20.161.92.111:35746). Mar 7 00:50:45.551990 sshd[1662]: Accepted publickey for core from 20.161.92.111 port 35746 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:45.554880 sshd[1662]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:45.560085 systemd-logind[1455]: New session 4 of user core. Mar 7 00:50:45.567023 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 7 00:50:45.984976 sshd[1662]: pam_unix(sshd:session): session closed for user core Mar 7 00:50:45.991750 systemd[1]: sshd@3-188.245.50.81:22-20.161.92.111:35746.service: Deactivated successfully. Mar 7 00:50:45.993965 systemd[1]: session-4.scope: Deactivated successfully. Mar 7 00:50:45.996647 systemd-logind[1455]: Session 4 logged out. Waiting for processes to exit. Mar 7 00:50:45.998105 systemd-logind[1455]: Removed session 4. Mar 7 00:50:46.100130 systemd[1]: Started sshd@4-188.245.50.81:22-20.161.92.111:35760.service - OpenSSH per-connection server daemon (20.161.92.111:35760). Mar 7 00:50:46.695365 sshd[1669]: Accepted publickey for core from 20.161.92.111 port 35760 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:46.697649 sshd[1669]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:46.704876 systemd-logind[1455]: New session 5 of user core. Mar 7 00:50:46.711976 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 7 00:50:47.038631 sudo[1672]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 7 00:50:47.038927 sudo[1672]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 00:50:47.059631 sudo[1672]: pam_unix(sudo:session): session closed for user root Mar 7 00:50:47.156843 sshd[1669]: pam_unix(sshd:session): session closed for user core Mar 7 00:50:47.163364 systemd[1]: sshd@4-188.245.50.81:22-20.161.92.111:35760.service: Deactivated successfully. Mar 7 00:50:47.165472 systemd[1]: session-5.scope: Deactivated successfully. Mar 7 00:50:47.166285 systemd-logind[1455]: Session 5 logged out. Waiting for processes to exit. Mar 7 00:50:47.167668 systemd-logind[1455]: Removed session 5. Mar 7 00:50:47.264092 systemd[1]: Started sshd@5-188.245.50.81:22-20.161.92.111:35770.service - OpenSSH per-connection server daemon (20.161.92.111:35770). Mar 7 00:50:47.861644 sshd[1677]: Accepted publickey for core from 20.161.92.111 port 35770 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:47.863361 sshd[1677]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:47.869338 systemd-logind[1455]: New session 6 of user core. Mar 7 00:50:47.875887 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 7 00:50:48.194654 sudo[1681]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 7 00:50:48.194961 sudo[1681]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 00:50:48.200828 sudo[1681]: pam_unix(sudo:session): session closed for user root Mar 7 00:50:48.208219 sudo[1680]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Mar 7 00:50:48.208891 sudo[1680]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 00:50:48.232220 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Mar 7 00:50:48.236183 auditctl[1684]: No rules Mar 7 00:50:48.237744 systemd[1]: audit-rules.service: Deactivated successfully. Mar 7 00:50:48.238035 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Mar 7 00:50:48.246530 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Mar 7 00:50:48.275230 augenrules[1702]: No rules Mar 7 00:50:48.276979 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Mar 7 00:50:48.280032 sudo[1680]: pam_unix(sudo:session): session closed for user root Mar 7 00:50:48.375912 sshd[1677]: pam_unix(sshd:session): session closed for user core Mar 7 00:50:48.380753 systemd[1]: sshd@5-188.245.50.81:22-20.161.92.111:35770.service: Deactivated successfully. Mar 7 00:50:48.383043 systemd[1]: session-6.scope: Deactivated successfully. Mar 7 00:50:48.386991 systemd-logind[1455]: Session 6 logged out. Waiting for processes to exit. Mar 7 00:50:48.388359 systemd-logind[1455]: Removed session 6. Mar 7 00:50:48.493094 systemd[1]: Started sshd@6-188.245.50.81:22-20.161.92.111:35778.service - OpenSSH per-connection server daemon (20.161.92.111:35778). Mar 7 00:50:49.097541 sshd[1710]: Accepted publickey for core from 20.161.92.111 port 35778 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:50:49.100093 sshd[1710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:50:49.105163 systemd-logind[1455]: New session 7 of user core. Mar 7 00:50:49.112934 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 7 00:50:49.433628 sudo[1713]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 7 00:50:49.433906 sudo[1713]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 7 00:50:49.743263 (dockerd)[1729]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 7 00:50:49.743681 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 7 00:50:50.000838 dockerd[1729]: time="2026-03-07T00:50:50.000125071Z" level=info msg="Starting up" Mar 7 00:50:50.085256 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2767700744-merged.mount: Deactivated successfully. Mar 7 00:50:50.106541 dockerd[1729]: time="2026-03-07T00:50:50.106493568Z" level=info msg="Loading containers: start." Mar 7 00:50:50.244609 kernel: Initializing XFRM netlink socket Mar 7 00:50:50.318891 systemd-networkd[1380]: docker0: Link UP Mar 7 00:50:50.334322 dockerd[1729]: time="2026-03-07T00:50:50.334233653Z" level=info msg="Loading containers: done." Mar 7 00:50:50.351645 dockerd[1729]: time="2026-03-07T00:50:50.351502574Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 7 00:50:50.352230 dockerd[1729]: time="2026-03-07T00:50:50.351857328Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Mar 7 00:50:50.352230 dockerd[1729]: time="2026-03-07T00:50:50.352016021Z" level=info msg="Daemon has completed initialization" Mar 7 00:50:50.397992 dockerd[1729]: time="2026-03-07T00:50:50.397820592Z" level=info msg="API listen on /run/docker.sock" Mar 7 00:50:50.398793 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 7 00:50:50.895326 containerd[1484]: time="2026-03-07T00:50:50.895280115Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\"" Mar 7 00:50:51.081640 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1831618818-merged.mount: Deactivated successfully. Mar 7 00:50:51.519187 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1782247163.mount: Deactivated successfully. Mar 7 00:50:51.898076 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 7 00:50:51.903920 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:50:52.041994 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:50:52.061189 (kubelet)[1927]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:50:52.107827 kubelet[1927]: E0307 00:50:52.107760 1927 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:50:52.111512 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:50:52.111680 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:50:52.728894 containerd[1484]: time="2026-03-07T00:50:52.728811686Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:52.731440 containerd[1484]: time="2026-03-07T00:50:52.731373603Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.5: active requests=0, bytes read=24583350" Mar 7 00:50:52.733510 containerd[1484]: time="2026-03-07T00:50:52.733434238Z" level=info msg="ImageCreate event name:\"sha256:3299c3f36446e899e7d38f97cdbd93a12ace0457ebca8f6d94ab33d86f9740bd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:52.738423 containerd[1484]: time="2026-03-07T00:50:52.738329354Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:52.739666 containerd[1484]: time="2026-03-07T00:50:52.739329499Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.5\" with image id \"sha256:3299c3f36446e899e7d38f97cdbd93a12ace0457ebca8f6d94ab33d86f9740bd\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c548633fcd3b4aad59b70815be4c8be54a0fddaddc3fcffa9371eedb0e96417a\", size \"24579851\" in 1.844002374s" Mar 7 00:50:52.739666 containerd[1484]: time="2026-03-07T00:50:52.739368304Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.5\" returns image reference \"sha256:3299c3f36446e899e7d38f97cdbd93a12ace0457ebca8f6d94ab33d86f9740bd\"" Mar 7 00:50:52.740995 containerd[1484]: time="2026-03-07T00:50:52.740960532Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\"" Mar 7 00:50:53.900682 update_engine[1456]: I20260307 00:50:53.900069 1456 update_attempter.cc:509] Updating boot flags... Mar 7 00:50:53.955616 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1951) Mar 7 00:50:54.025622 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1952) Mar 7 00:50:54.274365 containerd[1484]: time="2026-03-07T00:50:54.274186444Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:54.281718 containerd[1484]: time="2026-03-07T00:50:54.281146565Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.5: active requests=0, bytes read=19139661" Mar 7 00:50:54.284613 containerd[1484]: time="2026-03-07T00:50:54.282984021Z" level=info msg="ImageCreate event name:\"sha256:be20fbe989d9e759458cc8dbbc6e6c4a17e5d6f9db86b2a6cf4e3dfba0fe86e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:54.288950 containerd[1484]: time="2026-03-07T00:50:54.288892177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:54.292019 containerd[1484]: time="2026-03-07T00:50:54.291954873Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.5\" with image id \"sha256:be20fbe989d9e759458cc8dbbc6e6c4a17e5d6f9db86b2a6cf4e3dfba0fe86e5\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f0426100c873816560c520d542fa28999a98dad909edd04365f3b0eead790da3\", size \"20724045\" in 1.55095299s" Mar 7 00:50:54.292232 containerd[1484]: time="2026-03-07T00:50:54.292195540Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.5\" returns image reference \"sha256:be20fbe989d9e759458cc8dbbc6e6c4a17e5d6f9db86b2a6cf4e3dfba0fe86e5\"" Mar 7 00:50:54.293082 containerd[1484]: time="2026-03-07T00:50:54.293046412Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\"" Mar 7 00:50:55.319700 containerd[1484]: time="2026-03-07T00:50:55.319639478Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:55.321509 containerd[1484]: time="2026-03-07T00:50:55.321465884Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.5: active requests=0, bytes read=14195564" Mar 7 00:50:55.323598 containerd[1484]: time="2026-03-07T00:50:55.321827359Z" level=info msg="ImageCreate event name:\"sha256:4addcfb720a81f20ddfad093c4a397bb9f3d99b798f610f0ecc83cafd7f0a3bd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:55.325133 containerd[1484]: time="2026-03-07T00:50:55.325098622Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:55.326400 containerd[1484]: time="2026-03-07T00:50:55.326363344Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.5\" with image id \"sha256:4addcfb720a81f20ddfad093c4a397bb9f3d99b798f610f0ecc83cafd7f0a3bd\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b67b0d627c8e99ffa362bd4d9a60ca9a6c449e363a5f88d2aa8c224bd84ca51d\", size \"15779966\" in 1.033163514s" Mar 7 00:50:55.326511 containerd[1484]: time="2026-03-07T00:50:55.326492632Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.5\" returns image reference \"sha256:4addcfb720a81f20ddfad093c4a397bb9f3d99b798f610f0ecc83cafd7f0a3bd\"" Mar 7 00:50:55.327637 containerd[1484]: time="2026-03-07T00:50:55.327606747Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\"" Mar 7 00:50:56.218951 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount628410357.mount: Deactivated successfully. Mar 7 00:50:56.511675 containerd[1484]: time="2026-03-07T00:50:56.511461013Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:56.513027 containerd[1484]: time="2026-03-07T00:50:56.512969934Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.5: active requests=0, bytes read=22697114" Mar 7 00:50:56.513796 containerd[1484]: time="2026-03-07T00:50:56.513716053Z" level=info msg="ImageCreate event name:\"sha256:8167398c8957d56adceac5bd6436d6ac238c546a5f5c92e450a1c380c0aa7d5d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:56.516953 containerd[1484]: time="2026-03-07T00:50:56.516899833Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:56.517765 containerd[1484]: time="2026-03-07T00:50:56.517710912Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.5\" with image id \"sha256:8167398c8957d56adceac5bd6436d6ac238c546a5f5c92e450a1c380c0aa7d5d\", repo tag \"registry.k8s.io/kube-proxy:v1.34.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:8a22a3bf452d07af3b5a3064b089d2ad6579d5dd3b850386e05cc0f36dc3f4cf\", size \"22696107\" in 1.189948414s" Mar 7 00:50:56.517765 containerd[1484]: time="2026-03-07T00:50:56.517750507Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.5\" returns image reference \"sha256:8167398c8957d56adceac5bd6436d6ac238c546a5f5c92e450a1c380c0aa7d5d\"" Mar 7 00:50:56.518617 containerd[1484]: time="2026-03-07T00:50:56.518581284Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Mar 7 00:50:57.027881 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount945364479.mount: Deactivated successfully. Mar 7 00:50:58.105236 containerd[1484]: time="2026-03-07T00:50:58.105175777Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.106619 containerd[1484]: time="2026-03-07T00:50:58.106585706Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=20395498" Mar 7 00:50:58.107634 containerd[1484]: time="2026-03-07T00:50:58.107589822Z" level=info msg="ImageCreate event name:\"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.112029 containerd[1484]: time="2026-03-07T00:50:58.111950424Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.113755 containerd[1484]: time="2026-03-07T00:50:58.113407652Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"20392204\" in 1.594626515s" Mar 7 00:50:58.113755 containerd[1484]: time="2026-03-07T00:50:58.113452381Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:138784d87c9c50f8e59412544da4cf4928d61ccbaf93b9f5898a3ba406871bfc\"" Mar 7 00:50:58.114286 containerd[1484]: time="2026-03-07T00:50:58.114261881Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Mar 7 00:50:58.579523 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3295096155.mount: Deactivated successfully. Mar 7 00:50:58.589941 containerd[1484]: time="2026-03-07T00:50:58.588456799Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.590345 containerd[1484]: time="2026-03-07T00:50:58.590297155Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=268729" Mar 7 00:50:58.591392 containerd[1484]: time="2026-03-07T00:50:58.591328253Z" level=info msg="ImageCreate event name:\"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.594452 containerd[1484]: time="2026-03-07T00:50:58.594381152Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:58.596251 containerd[1484]: time="2026-03-07T00:50:58.596022797Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"267939\" in 481.654193ms" Mar 7 00:50:58.596251 containerd[1484]: time="2026-03-07T00:50:58.596067125Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:d7b100cd9a77ba782c5e428c8dd5a1df4a1e79d4cb6294acd7d01290ab3babbd\"" Mar 7 00:50:58.596676 containerd[1484]: time="2026-03-07T00:50:58.596625555Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\"" Mar 7 00:50:59.136896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3980549195.mount: Deactivated successfully. Mar 7 00:50:59.896648 containerd[1484]: time="2026-03-07T00:50:59.896552158Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.5-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:59.898530 containerd[1484]: time="2026-03-07T00:50:59.898459228Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.5-0: active requests=0, bytes read=21125601" Mar 7 00:50:59.899795 containerd[1484]: time="2026-03-07T00:50:59.899712254Z" level=info msg="ImageCreate event name:\"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:59.902998 containerd[1484]: time="2026-03-07T00:50:59.902958435Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:50:59.905366 containerd[1484]: time="2026-03-07T00:50:59.904943965Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.5-0\" with image id \"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\", repo tag \"registry.k8s.io/etcd:3.6.5-0\", repo digest \"registry.k8s.io/etcd@sha256:042ef9c02799eb9303abf1aa99b09f09d94b8ee3ba0c2dd3f42dc4e1d3dce534\", size \"21136588\" in 1.308284687s" Mar 7 00:50:59.905366 containerd[1484]: time="2026-03-07T00:50:59.904977717Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.5-0\" returns image reference \"sha256:2c5f0dedd21c25ec3a6709934d22152d53ec50fe57b72d29e4450655e3d14d42\"" Mar 7 00:51:02.148930 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Mar 7 00:51:02.157780 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:51:02.282809 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:02.287728 (kubelet)[2117]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 7 00:51:02.347597 kubelet[2117]: E0307 00:51:02.346993 2117 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 7 00:51:02.350143 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 7 00:51:02.350468 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 7 00:51:07.444623 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:07.455302 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:51:07.491217 systemd[1]: Reloading requested from client PID 2132 ('systemctl') (unit session-7.scope)... Mar 7 00:51:07.491234 systemd[1]: Reloading... Mar 7 00:51:07.635890 zram_generator::config[2172]: No configuration found. Mar 7 00:51:07.727024 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 00:51:07.799944 systemd[1]: Reloading finished in 308 ms. Mar 7 00:51:07.858927 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 7 00:51:07.859019 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 7 00:51:07.859266 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:07.868093 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:51:08.002492 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:08.016200 (kubelet)[2220]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 00:51:08.064169 kubelet[2220]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 00:51:08.064169 kubelet[2220]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 00:51:08.065296 kubelet[2220]: I0307 00:51:08.065216 2220 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 00:51:09.685508 kubelet[2220]: I0307 00:51:09.685462 2220 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 7 00:51:09.685508 kubelet[2220]: I0307 00:51:09.685501 2220 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 00:51:09.686008 kubelet[2220]: I0307 00:51:09.685533 2220 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 7 00:51:09.686008 kubelet[2220]: I0307 00:51:09.685541 2220 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 00:51:09.686008 kubelet[2220]: I0307 00:51:09.685927 2220 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 00:51:09.694468 kubelet[2220]: E0307 00:51:09.694411 2220 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://188.245.50.81:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 7 00:51:09.697049 kubelet[2220]: I0307 00:51:09.696010 2220 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 00:51:09.702049 kubelet[2220]: E0307 00:51:09.701834 2220 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 00:51:09.702729 kubelet[2220]: I0307 00:51:09.702595 2220 server.go:1400] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Mar 7 00:51:09.705798 kubelet[2220]: I0307 00:51:09.705096 2220 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 7 00:51:09.705798 kubelet[2220]: I0307 00:51:09.705319 2220 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 00:51:09.705798 kubelet[2220]: I0307 00:51:09.705340 2220 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-53d1697e27","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 7 00:51:09.705798 kubelet[2220]: I0307 00:51:09.705479 2220 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 00:51:09.706023 kubelet[2220]: I0307 00:51:09.705487 2220 container_manager_linux.go:306] "Creating device plugin manager" Mar 7 00:51:09.706023 kubelet[2220]: I0307 00:51:09.705610 2220 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 7 00:51:09.707776 kubelet[2220]: I0307 00:51:09.707537 2220 state_mem.go:36] "Initialized new in-memory state store" Mar 7 00:51:09.709185 kubelet[2220]: I0307 00:51:09.709161 2220 kubelet.go:475] "Attempting to sync node with API server" Mar 7 00:51:09.709294 kubelet[2220]: I0307 00:51:09.709283 2220 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 00:51:09.709392 kubelet[2220]: I0307 00:51:09.709382 2220 kubelet.go:387] "Adding apiserver pod source" Mar 7 00:51:09.709451 kubelet[2220]: I0307 00:51:09.709442 2220 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 00:51:09.710760 kubelet[2220]: E0307 00:51:09.710731 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://188.245.50.81:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 00:51:09.710989 kubelet[2220]: E0307 00:51:09.710958 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://188.245.50.81:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-53d1697e27&limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 00:51:09.711981 kubelet[2220]: I0307 00:51:09.711963 2220 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 00:51:09.713247 kubelet[2220]: I0307 00:51:09.712671 2220 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 00:51:09.713247 kubelet[2220]: I0307 00:51:09.712703 2220 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 7 00:51:09.713247 kubelet[2220]: W0307 00:51:09.712745 2220 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 7 00:51:09.716901 kubelet[2220]: I0307 00:51:09.716884 2220 server.go:1262] "Started kubelet" Mar 7 00:51:09.717790 kubelet[2220]: I0307 00:51:09.717758 2220 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 00:51:09.718608 kubelet[2220]: I0307 00:51:09.718593 2220 server.go:310] "Adding debug handlers to kubelet server" Mar 7 00:51:09.721247 kubelet[2220]: I0307 00:51:09.720434 2220 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 00:51:09.721247 kubelet[2220]: I0307 00:51:09.720515 2220 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 7 00:51:09.721247 kubelet[2220]: I0307 00:51:09.720865 2220 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 00:51:09.722288 kubelet[2220]: E0307 00:51:09.720998 2220 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://188.245.50.81:6443/api/v1/namespaces/default/events\": dial tcp 188.245.50.81:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-6-n-53d1697e27.189a68d58c6af0cb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-6-n-53d1697e27,UID:ci-4081-3-6-n-53d1697e27,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-53d1697e27,},FirstTimestamp:2026-03-07 00:51:09.716852939 +0000 UTC m=+1.695034444,LastTimestamp:2026-03-07 00:51:09.716852939 +0000 UTC m=+1.695034444,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-53d1697e27,}" Mar 7 00:51:09.725068 kubelet[2220]: I0307 00:51:09.725043 2220 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 00:51:09.725387 kubelet[2220]: I0307 00:51:09.725369 2220 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 00:51:09.727707 kubelet[2220]: E0307 00:51:09.727684 2220 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 7 00:51:09.727943 kubelet[2220]: E0307 00:51:09.727926 2220 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-53d1697e27\" not found" Mar 7 00:51:09.728031 kubelet[2220]: I0307 00:51:09.728019 2220 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 7 00:51:09.728280 kubelet[2220]: I0307 00:51:09.728264 2220 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 7 00:51:09.728404 kubelet[2220]: I0307 00:51:09.728394 2220 reconciler.go:29] "Reconciler: start to sync state" Mar 7 00:51:09.728957 kubelet[2220]: E0307 00:51:09.728934 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://188.245.50.81:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 7 00:51:09.729433 kubelet[2220]: E0307 00:51:09.729404 2220 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://188.245.50.81:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-53d1697e27?timeout=10s\": dial tcp 188.245.50.81:6443: connect: connection refused" interval="200ms" Mar 7 00:51:09.729821 kubelet[2220]: I0307 00:51:09.729802 2220 factory.go:223] Registration of the systemd container factory successfully Mar 7 00:51:09.730004 kubelet[2220]: I0307 00:51:09.729985 2220 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 00:51:09.731092 kubelet[2220]: I0307 00:51:09.731062 2220 factory.go:223] Registration of the containerd container factory successfully Mar 7 00:51:09.732014 kubelet[2220]: I0307 00:51:09.731964 2220 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 7 00:51:09.751550 kubelet[2220]: I0307 00:51:09.751463 2220 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 7 00:51:09.751550 kubelet[2220]: I0307 00:51:09.751502 2220 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 7 00:51:09.751550 kubelet[2220]: I0307 00:51:09.751532 2220 kubelet.go:2428] "Starting kubelet main sync loop" Mar 7 00:51:09.751824 kubelet[2220]: E0307 00:51:09.751594 2220 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 00:51:09.756153 kubelet[2220]: E0307 00:51:09.756088 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://188.245.50.81:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 7 00:51:09.767730 kubelet[2220]: I0307 00:51:09.767679 2220 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 00:51:09.767730 kubelet[2220]: I0307 00:51:09.767704 2220 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 00:51:09.767730 kubelet[2220]: I0307 00:51:09.767723 2220 state_mem.go:36] "Initialized new in-memory state store" Mar 7 00:51:09.772159 kubelet[2220]: I0307 00:51:09.772095 2220 policy_none.go:49] "None policy: Start" Mar 7 00:51:09.772159 kubelet[2220]: I0307 00:51:09.772133 2220 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 7 00:51:09.772159 kubelet[2220]: I0307 00:51:09.772145 2220 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 7 00:51:09.773979 kubelet[2220]: I0307 00:51:09.773955 2220 policy_none.go:47] "Start" Mar 7 00:51:09.779486 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 7 00:51:09.792642 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 7 00:51:09.796318 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 7 00:51:09.806650 kubelet[2220]: E0307 00:51:09.805847 2220 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 00:51:09.806650 kubelet[2220]: I0307 00:51:09.806127 2220 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 00:51:09.806650 kubelet[2220]: I0307 00:51:09.806144 2220 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 00:51:09.806650 kubelet[2220]: I0307 00:51:09.806663 2220 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 00:51:09.810399 kubelet[2220]: E0307 00:51:09.810359 2220 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 00:51:09.810760 kubelet[2220]: E0307 00:51:09.810730 2220 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-6-n-53d1697e27\" not found" Mar 7 00:51:09.865367 systemd[1]: Created slice kubepods-burstable-podd8f73cc52477f3d4b53b63fe459959f3.slice - libcontainer container kubepods-burstable-podd8f73cc52477f3d4b53b63fe459959f3.slice. Mar 7 00:51:09.884062 kubelet[2220]: E0307 00:51:09.883604 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.888246 systemd[1]: Created slice kubepods-burstable-pod94e92bee46617f89a0f97cade04f43bc.slice - libcontainer container kubepods-burstable-pod94e92bee46617f89a0f97cade04f43bc.slice. Mar 7 00:51:09.903356 kubelet[2220]: E0307 00:51:09.902775 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.905771 systemd[1]: Created slice kubepods-burstable-podf5d6fc8459c7d92c6c3fae424330ef0e.slice - libcontainer container kubepods-burstable-podf5d6fc8459c7d92c6c3fae424330ef0e.slice. Mar 7 00:51:09.908515 kubelet[2220]: E0307 00:51:09.908406 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.908867 kubelet[2220]: I0307 00:51:09.908820 2220 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.909664 kubelet[2220]: E0307 00:51:09.909622 2220 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://188.245.50.81:6443/api/v1/nodes\": dial tcp 188.245.50.81:6443: connect: connection refused" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929335 kubelet[2220]: I0307 00:51:09.929247 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929335 kubelet[2220]: I0307 00:51:09.929320 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929558 kubelet[2220]: I0307 00:51:09.929360 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929558 kubelet[2220]: I0307 00:51:09.929390 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929558 kubelet[2220]: I0307 00:51:09.929440 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f5d6fc8459c7d92c6c3fae424330ef0e-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-53d1697e27\" (UID: \"f5d6fc8459c7d92c6c3fae424330ef0e\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929558 kubelet[2220]: I0307 00:51:09.929474 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929558 kubelet[2220]: I0307 00:51:09.929504 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929789 kubelet[2220]: I0307 00:51:09.929534 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.929789 kubelet[2220]: I0307 00:51:09.929601 2220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:09.930367 kubelet[2220]: E0307 00:51:09.930305 2220 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://188.245.50.81:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-53d1697e27?timeout=10s\": dial tcp 188.245.50.81:6443: connect: connection refused" interval="400ms" Mar 7 00:51:10.112982 kubelet[2220]: I0307 00:51:10.112817 2220 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:10.114779 kubelet[2220]: E0307 00:51:10.114732 2220 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://188.245.50.81:6443/api/v1/nodes\": dial tcp 188.245.50.81:6443: connect: connection refused" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:10.188098 containerd[1484]: time="2026-03-07T00:51:10.188021174Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-53d1697e27,Uid:d8f73cc52477f3d4b53b63fe459959f3,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:10.205304 containerd[1484]: time="2026-03-07T00:51:10.205193247Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-53d1697e27,Uid:94e92bee46617f89a0f97cade04f43bc,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:10.211990 containerd[1484]: time="2026-03-07T00:51:10.211918222Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-53d1697e27,Uid:f5d6fc8459c7d92c6c3fae424330ef0e,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:10.331591 kubelet[2220]: E0307 00:51:10.331178 2220 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://188.245.50.81:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-53d1697e27?timeout=10s\": dial tcp 188.245.50.81:6443: connect: connection refused" interval="800ms" Mar 7 00:51:10.518206 kubelet[2220]: I0307 00:51:10.517856 2220 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:10.518206 kubelet[2220]: E0307 00:51:10.518167 2220 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://188.245.50.81:6443/api/v1/nodes\": dial tcp 188.245.50.81:6443: connect: connection refused" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:10.675815 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1934949700.mount: Deactivated successfully. Mar 7 00:51:10.682520 containerd[1484]: time="2026-03-07T00:51:10.682461619Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 00:51:10.685974 containerd[1484]: time="2026-03-07T00:51:10.685928426Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Mar 7 00:51:10.686819 containerd[1484]: time="2026-03-07T00:51:10.686783679Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 00:51:10.688827 containerd[1484]: time="2026-03-07T00:51:10.688331701Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 00:51:10.690247 containerd[1484]: time="2026-03-07T00:51:10.690206093Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 00:51:10.691221 containerd[1484]: time="2026-03-07T00:51:10.691162663Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 00:51:10.691221 containerd[1484]: time="2026-03-07T00:51:10.691200332Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 7 00:51:10.694607 kubelet[2220]: E0307 00:51:10.693087 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://188.245.50.81:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-53d1697e27&limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 7 00:51:10.697728 containerd[1484]: time="2026-03-07T00:51:10.697652659Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 7 00:51:10.701970 containerd[1484]: time="2026-03-07T00:51:10.701915034Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 496.596572ms" Mar 7 00:51:10.704540 containerd[1484]: time="2026-03-07T00:51:10.704496485Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 516.385002ms" Mar 7 00:51:10.705343 containerd[1484]: time="2026-03-07T00:51:10.705313629Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 493.300735ms" Mar 7 00:51:10.823740 containerd[1484]: time="2026-03-07T00:51:10.822795180Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:10.823740 containerd[1484]: time="2026-03-07T00:51:10.822842456Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:10.823740 containerd[1484]: time="2026-03-07T00:51:10.822853865Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.823740 containerd[1484]: time="2026-03-07T00:51:10.822924759Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.828101 containerd[1484]: time="2026-03-07T00:51:10.827670663Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:10.828101 containerd[1484]: time="2026-03-07T00:51:10.827740036Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:10.828101 containerd[1484]: time="2026-03-07T00:51:10.827751085Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.828101 containerd[1484]: time="2026-03-07T00:51:10.827841794Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.843008 containerd[1484]: time="2026-03-07T00:51:10.840497338Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:10.843008 containerd[1484]: time="2026-03-07T00:51:10.841617874Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:10.843008 containerd[1484]: time="2026-03-07T00:51:10.841649938Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.843988 containerd[1484]: time="2026-03-07T00:51:10.843953137Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:10.847008 systemd[1]: Started cri-containerd-b57aa2baf4ae34eb9e0d5897e04eeed91e8cb89056fb32d605a7a8e3246205e8.scope - libcontainer container b57aa2baf4ae34eb9e0d5897e04eeed91e8cb89056fb32d605a7a8e3246205e8. Mar 7 00:51:10.855523 systemd[1]: Started cri-containerd-6d36d780723906d304f985b6428052b72c8b2c9f2cc1358828a0432b8d633f35.scope - libcontainer container 6d36d780723906d304f985b6428052b72c8b2c9f2cc1358828a0432b8d633f35. Mar 7 00:51:10.881974 systemd[1]: Started cri-containerd-3059296bc86bc86254d53572f83772e50ea86ce700adb1654d59132e6884163f.scope - libcontainer container 3059296bc86bc86254d53572f83772e50ea86ce700adb1654d59132e6884163f. Mar 7 00:51:10.902467 containerd[1484]: time="2026-03-07T00:51:10.901551320Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-53d1697e27,Uid:d8f73cc52477f3d4b53b63fe459959f3,Namespace:kube-system,Attempt:0,} returns sandbox id \"b57aa2baf4ae34eb9e0d5897e04eeed91e8cb89056fb32d605a7a8e3246205e8\"" Mar 7 00:51:10.914488 containerd[1484]: time="2026-03-07T00:51:10.914322793Z" level=info msg="CreateContainer within sandbox \"b57aa2baf4ae34eb9e0d5897e04eeed91e8cb89056fb32d605a7a8e3246205e8\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 7 00:51:10.931077 containerd[1484]: time="2026-03-07T00:51:10.931038237Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-53d1697e27,Uid:94e92bee46617f89a0f97cade04f43bc,Namespace:kube-system,Attempt:0,} returns sandbox id \"6d36d780723906d304f985b6428052b72c8b2c9f2cc1358828a0432b8d633f35\"" Mar 7 00:51:10.932322 containerd[1484]: time="2026-03-07T00:51:10.932205088Z" level=info msg="CreateContainer within sandbox \"b57aa2baf4ae34eb9e0d5897e04eeed91e8cb89056fb32d605a7a8e3246205e8\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9acf00759608ce64828f326a18feef104186947d337848eb7504d83558d016fb\"" Mar 7 00:51:10.933455 containerd[1484]: time="2026-03-07T00:51:10.933410608Z" level=info msg="StartContainer for \"9acf00759608ce64828f326a18feef104186947d337848eb7504d83558d016fb\"" Mar 7 00:51:10.940158 containerd[1484]: time="2026-03-07T00:51:10.940121013Z" level=info msg="CreateContainer within sandbox \"6d36d780723906d304f985b6428052b72c8b2c9f2cc1358828a0432b8d633f35\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 7 00:51:10.959028 containerd[1484]: time="2026-03-07T00:51:10.958933218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-53d1697e27,Uid:f5d6fc8459c7d92c6c3fae424330ef0e,Namespace:kube-system,Attempt:0,} returns sandbox id \"3059296bc86bc86254d53572f83772e50ea86ce700adb1654d59132e6884163f\"" Mar 7 00:51:10.965435 containerd[1484]: time="2026-03-07T00:51:10.965267215Z" level=info msg="CreateContainer within sandbox \"3059296bc86bc86254d53572f83772e50ea86ce700adb1654d59132e6884163f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 7 00:51:10.970148 systemd[1]: Started cri-containerd-9acf00759608ce64828f326a18feef104186947d337848eb7504d83558d016fb.scope - libcontainer container 9acf00759608ce64828f326a18feef104186947d337848eb7504d83558d016fb. Mar 7 00:51:10.973310 containerd[1484]: time="2026-03-07T00:51:10.972871342Z" level=info msg="CreateContainer within sandbox \"6d36d780723906d304f985b6428052b72c8b2c9f2cc1358828a0432b8d633f35\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"913cd677ad467f5408705567b08720511c9a971e7aa9dee4f18fffcadaba7f52\"" Mar 7 00:51:10.974289 containerd[1484]: time="2026-03-07T00:51:10.974264245Z" level=info msg="StartContainer for \"913cd677ad467f5408705567b08720511c9a971e7aa9dee4f18fffcadaba7f52\"" Mar 7 00:51:10.990081 containerd[1484]: time="2026-03-07T00:51:10.989943538Z" level=info msg="CreateContainer within sandbox \"3059296bc86bc86254d53572f83772e50ea86ce700adb1654d59132e6884163f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"a311943d7242671555783e27edcdd8824600cbe167037ae2dc091aeee073a7d9\"" Mar 7 00:51:10.991521 containerd[1484]: time="2026-03-07T00:51:10.991406055Z" level=info msg="StartContainer for \"a311943d7242671555783e27edcdd8824600cbe167037ae2dc091aeee073a7d9\"" Mar 7 00:51:11.015774 kubelet[2220]: E0307 00:51:11.015555 2220 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://188.245.50.81:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 188.245.50.81:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 7 00:51:11.022409 containerd[1484]: time="2026-03-07T00:51:11.022042721Z" level=info msg="StartContainer for \"9acf00759608ce64828f326a18feef104186947d337848eb7504d83558d016fb\" returns successfully" Mar 7 00:51:11.041339 systemd[1]: Started cri-containerd-913cd677ad467f5408705567b08720511c9a971e7aa9dee4f18fffcadaba7f52.scope - libcontainer container 913cd677ad467f5408705567b08720511c9a971e7aa9dee4f18fffcadaba7f52. Mar 7 00:51:11.042495 systemd[1]: Started cri-containerd-a311943d7242671555783e27edcdd8824600cbe167037ae2dc091aeee073a7d9.scope - libcontainer container a311943d7242671555783e27edcdd8824600cbe167037ae2dc091aeee073a7d9. Mar 7 00:51:11.117279 containerd[1484]: time="2026-03-07T00:51:11.116410581Z" level=info msg="StartContainer for \"913cd677ad467f5408705567b08720511c9a971e7aa9dee4f18fffcadaba7f52\" returns successfully" Mar 7 00:51:11.117279 containerd[1484]: time="2026-03-07T00:51:11.116530421Z" level=info msg="StartContainer for \"a311943d7242671555783e27edcdd8824600cbe167037ae2dc091aeee073a7d9\" returns successfully" Mar 7 00:51:11.132156 kubelet[2220]: E0307 00:51:11.132101 2220 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://188.245.50.81:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-53d1697e27?timeout=10s\": dial tcp 188.245.50.81:6443: connect: connection refused" interval="1.6s" Mar 7 00:51:11.320523 kubelet[2220]: I0307 00:51:11.320488 2220 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:11.776886 kubelet[2220]: E0307 00:51:11.776793 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:11.781781 kubelet[2220]: E0307 00:51:11.781710 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:11.793696 kubelet[2220]: E0307 00:51:11.791964 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:12.788791 kubelet[2220]: E0307 00:51:12.788755 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:12.789319 kubelet[2220]: E0307 00:51:12.789116 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:12.790764 kubelet[2220]: E0307 00:51:12.790640 2220 kubelet.go:3216] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.712492 kubelet[2220]: I0307 00:51:13.712455 2220 apiserver.go:52] "Watching apiserver" Mar 7 00:51:13.735475 kubelet[2220]: E0307 00:51:13.735430 2220 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-3-6-n-53d1697e27\" not found" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.802778 kubelet[2220]: I0307 00:51:13.802653 2220 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.828841 kubelet[2220]: I0307 00:51:13.828608 2220 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 7 00:51:13.831623 kubelet[2220]: I0307 00:51:13.829739 2220 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.888916 kubelet[2220]: E0307 00:51:13.888678 2220 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.888916 kubelet[2220]: I0307 00:51:13.888712 2220 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.890951 kubelet[2220]: E0307 00:51:13.890917 2220 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-53d1697e27\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.891281 kubelet[2220]: I0307 00:51:13.891077 2220 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:13.893198 kubelet[2220]: E0307 00:51:13.893156 2220 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:15.966165 systemd[1]: Reloading requested from client PID 2507 ('systemctl') (unit session-7.scope)... Mar 7 00:51:15.966180 systemd[1]: Reloading... Mar 7 00:51:16.095652 zram_generator::config[2552]: No configuration found. Mar 7 00:51:16.219561 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 7 00:51:16.302992 systemd[1]: Reloading finished in 336 ms. Mar 7 00:51:16.349773 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:51:16.369437 systemd[1]: kubelet.service: Deactivated successfully. Mar 7 00:51:16.370020 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:16.370200 systemd[1]: kubelet.service: Consumed 2.125s CPU time, 122.3M memory peak, 0B memory swap peak. Mar 7 00:51:16.378479 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 7 00:51:16.518019 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 7 00:51:16.519105 (kubelet)[2593]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 7 00:51:16.569757 kubelet[2593]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 7 00:51:16.569757 kubelet[2593]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 7 00:51:16.569757 kubelet[2593]: I0307 00:51:16.569493 2593 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 7 00:51:16.577908 kubelet[2593]: I0307 00:51:16.577771 2593 server.go:529] "Kubelet version" kubeletVersion="v1.34.4" Mar 7 00:51:16.577908 kubelet[2593]: I0307 00:51:16.577893 2593 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 7 00:51:16.578072 kubelet[2593]: I0307 00:51:16.577927 2593 watchdog_linux.go:95] "Systemd watchdog is not enabled" Mar 7 00:51:16.578072 kubelet[2593]: I0307 00:51:16.577936 2593 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 7 00:51:16.579469 kubelet[2593]: I0307 00:51:16.578416 2593 server.go:956] "Client rotation is on, will bootstrap in background" Mar 7 00:51:16.583514 kubelet[2593]: I0307 00:51:16.583461 2593 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 7 00:51:16.589935 kubelet[2593]: I0307 00:51:16.589140 2593 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 7 00:51:16.593506 kubelet[2593]: E0307 00:51:16.593476 2593 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 7 00:51:16.593623 kubelet[2593]: I0307 00:51:16.593535 2593 server.go:1400] "CRI implementation should be updated to support RuntimeConfig. Falling back to using cgroupDriver from kubelet config." Mar 7 00:51:16.596720 kubelet[2593]: I0307 00:51:16.596697 2593 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Mar 7 00:51:16.596958 kubelet[2593]: I0307 00:51:16.596929 2593 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 7 00:51:16.597104 kubelet[2593]: I0307 00:51:16.596960 2593 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-53d1697e27","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 7 00:51:16.597189 kubelet[2593]: I0307 00:51:16.597107 2593 topology_manager.go:138] "Creating topology manager with none policy" Mar 7 00:51:16.597189 kubelet[2593]: I0307 00:51:16.597116 2593 container_manager_linux.go:306] "Creating device plugin manager" Mar 7 00:51:16.597189 kubelet[2593]: I0307 00:51:16.597138 2593 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Mar 7 00:51:16.597323 kubelet[2593]: I0307 00:51:16.597314 2593 state_mem.go:36] "Initialized new in-memory state store" Mar 7 00:51:16.597476 kubelet[2593]: I0307 00:51:16.597466 2593 kubelet.go:475] "Attempting to sync node with API server" Mar 7 00:51:16.597510 kubelet[2593]: I0307 00:51:16.597481 2593 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 7 00:51:16.598073 kubelet[2593]: I0307 00:51:16.597505 2593 kubelet.go:387] "Adding apiserver pod source" Mar 7 00:51:16.598073 kubelet[2593]: I0307 00:51:16.598034 2593 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 7 00:51:16.599223 kubelet[2593]: I0307 00:51:16.599200 2593 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Mar 7 00:51:16.599864 kubelet[2593]: I0307 00:51:16.599836 2593 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 7 00:51:16.600021 kubelet[2593]: I0307 00:51:16.599875 2593 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Mar 7 00:51:16.605475 kubelet[2593]: I0307 00:51:16.605448 2593 server.go:1262] "Started kubelet" Mar 7 00:51:16.610289 kubelet[2593]: I0307 00:51:16.610265 2593 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 7 00:51:16.617109 kubelet[2593]: I0307 00:51:16.613770 2593 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 7 00:51:16.617109 kubelet[2593]: I0307 00:51:16.614620 2593 server.go:310] "Adding debug handlers to kubelet server" Mar 7 00:51:16.626850 kubelet[2593]: I0307 00:51:16.626781 2593 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 7 00:51:16.627049 kubelet[2593]: I0307 00:51:16.627036 2593 server_v1.go:49] "podresources" method="list" useActivePods=true Mar 7 00:51:16.627805 kubelet[2593]: I0307 00:51:16.627759 2593 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 7 00:51:16.632106 kubelet[2593]: I0307 00:51:16.632080 2593 volume_manager.go:313] "Starting Kubelet Volume Manager" Mar 7 00:51:16.632387 kubelet[2593]: E0307 00:51:16.632369 2593 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-53d1697e27\" not found" Mar 7 00:51:16.635631 kubelet[2593]: I0307 00:51:16.634316 2593 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 7 00:51:16.652040 kubelet[2593]: I0307 00:51:16.648943 2593 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 7 00:51:16.652040 kubelet[2593]: I0307 00:51:16.649099 2593 reconciler.go:29] "Reconciler: start to sync state" Mar 7 00:51:16.659033 kubelet[2593]: I0307 00:51:16.658305 2593 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Mar 7 00:51:16.659670 kubelet[2593]: I0307 00:51:16.659578 2593 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Mar 7 00:51:16.659670 kubelet[2593]: I0307 00:51:16.659604 2593 status_manager.go:244] "Starting to sync pod status with apiserver" Mar 7 00:51:16.659670 kubelet[2593]: I0307 00:51:16.659624 2593 kubelet.go:2428] "Starting kubelet main sync loop" Mar 7 00:51:16.659670 kubelet[2593]: E0307 00:51:16.659666 2593 kubelet.go:2452] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 7 00:51:16.666635 kubelet[2593]: I0307 00:51:16.665199 2593 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 7 00:51:16.669833 kubelet[2593]: I0307 00:51:16.669374 2593 factory.go:223] Registration of the containerd container factory successfully Mar 7 00:51:16.669833 kubelet[2593]: I0307 00:51:16.669390 2593 factory.go:223] Registration of the systemd container factory successfully Mar 7 00:51:16.671143 kubelet[2593]: E0307 00:51:16.671123 2593 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 7 00:51:16.741701 kubelet[2593]: I0307 00:51:16.741676 2593 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 7 00:51:16.741895 kubelet[2593]: I0307 00:51:16.741880 2593 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 7 00:51:16.741967 kubelet[2593]: I0307 00:51:16.741959 2593 state_mem.go:36] "Initialized new in-memory state store" Mar 7 00:51:16.742161 kubelet[2593]: I0307 00:51:16.742147 2593 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 7 00:51:16.742236 kubelet[2593]: I0307 00:51:16.742212 2593 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 7 00:51:16.742290 kubelet[2593]: I0307 00:51:16.742283 2593 policy_none.go:49] "None policy: Start" Mar 7 00:51:16.742352 kubelet[2593]: I0307 00:51:16.742344 2593 memory_manager.go:187] "Starting memorymanager" policy="None" Mar 7 00:51:16.742407 kubelet[2593]: I0307 00:51:16.742398 2593 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Mar 7 00:51:16.742688 kubelet[2593]: I0307 00:51:16.742672 2593 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Mar 7 00:51:16.742821 kubelet[2593]: I0307 00:51:16.742809 2593 policy_none.go:47] "Start" Mar 7 00:51:16.750454 kubelet[2593]: E0307 00:51:16.750417 2593 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 7 00:51:16.752143 kubelet[2593]: I0307 00:51:16.752123 2593 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 7 00:51:16.752423 kubelet[2593]: I0307 00:51:16.752319 2593 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 7 00:51:16.752839 kubelet[2593]: I0307 00:51:16.752817 2593 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 7 00:51:16.755719 kubelet[2593]: E0307 00:51:16.755699 2593 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 7 00:51:16.760438 kubelet[2593]: I0307 00:51:16.760405 2593 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.760673 kubelet[2593]: I0307 00:51:16.760660 2593 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.761156 kubelet[2593]: I0307 00:51:16.761132 2593 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.851053 kubelet[2593]: I0307 00:51:16.850880 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.851053 kubelet[2593]: I0307 00:51:16.850946 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.851053 kubelet[2593]: I0307 00:51:16.850979 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f5d6fc8459c7d92c6c3fae424330ef0e-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-53d1697e27\" (UID: \"f5d6fc8459c7d92c6c3fae424330ef0e\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.851053 kubelet[2593]: I0307 00:51:16.851003 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.852292 kubelet[2593]: I0307 00:51:16.852003 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.852292 kubelet[2593]: I0307 00:51:16.852075 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.852292 kubelet[2593]: I0307 00:51:16.852107 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.852292 kubelet[2593]: I0307 00:51:16.852139 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/94e92bee46617f89a0f97cade04f43bc-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-53d1697e27\" (UID: \"94e92bee46617f89a0f97cade04f43bc\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.852292 kubelet[2593]: I0307 00:51:16.852165 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d8f73cc52477f3d4b53b63fe459959f3-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" (UID: \"d8f73cc52477f3d4b53b63fe459959f3\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.862204 kubelet[2593]: I0307 00:51:16.861633 2593 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.871804 kubelet[2593]: I0307 00:51:16.871697 2593 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.873007 kubelet[2593]: I0307 00:51:16.872731 2593 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-53d1697e27" Mar 7 00:51:16.965117 sudo[2629]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 7 00:51:16.965525 sudo[2629]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 7 00:51:17.428361 sudo[2629]: pam_unix(sudo:session): session closed for user root Mar 7 00:51:17.605632 kubelet[2593]: I0307 00:51:17.605560 2593 apiserver.go:52] "Watching apiserver" Mar 7 00:51:17.649199 kubelet[2593]: I0307 00:51:17.649141 2593 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 7 00:51:17.712214 kubelet[2593]: I0307 00:51:17.712106 2593 kubelet.go:3220] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:17.729452 kubelet[2593]: E0307 00:51:17.729411 2593 kubelet.go:3222] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-53d1697e27\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" Mar 7 00:51:17.761094 kubelet[2593]: I0307 00:51:17.761023 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-6-n-53d1697e27" podStartSLOduration=1.761005621 podStartE2EDuration="1.761005621s" podCreationTimestamp="2026-03-07 00:51:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:17.759411236 +0000 UTC m=+1.233955478" watchObservedRunningTime="2026-03-07 00:51:17.761005621 +0000 UTC m=+1.235549823" Mar 7 00:51:17.761260 kubelet[2593]: I0307 00:51:17.761151 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-53d1697e27" podStartSLOduration=1.761146588 podStartE2EDuration="1.761146588s" podCreationTimestamp="2026-03-07 00:51:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:17.746477364 +0000 UTC m=+1.221021566" watchObservedRunningTime="2026-03-07 00:51:17.761146588 +0000 UTC m=+1.235690790" Mar 7 00:51:17.788981 kubelet[2593]: I0307 00:51:17.788909 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-6-n-53d1697e27" podStartSLOduration=1.788890691 podStartE2EDuration="1.788890691s" podCreationTimestamp="2026-03-07 00:51:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:17.772807433 +0000 UTC m=+1.247351675" watchObservedRunningTime="2026-03-07 00:51:17.788890691 +0000 UTC m=+1.263434853" Mar 7 00:51:20.166388 sudo[1713]: pam_unix(sudo:session): session closed for user root Mar 7 00:51:20.262654 sshd[1710]: pam_unix(sshd:session): session closed for user core Mar 7 00:51:20.270232 systemd-logind[1455]: Session 7 logged out. Waiting for processes to exit. Mar 7 00:51:20.270681 systemd[1]: sshd@6-188.245.50.81:22-20.161.92.111:35778.service: Deactivated successfully. Mar 7 00:51:20.273144 systemd[1]: session-7.scope: Deactivated successfully. Mar 7 00:51:20.273635 systemd[1]: session-7.scope: Consumed 10.920s CPU time, 152.8M memory peak, 0B memory swap peak. Mar 7 00:51:20.275265 systemd-logind[1455]: Removed session 7. Mar 7 00:51:21.159534 kubelet[2593]: I0307 00:51:21.159291 2593 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 7 00:51:21.161009 kubelet[2593]: I0307 00:51:21.160508 2593 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 7 00:51:21.161101 containerd[1484]: time="2026-03-07T00:51:21.159837077Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 7 00:51:22.127244 systemd[1]: Created slice kubepods-besteffort-pod6c452799_bfc5_44e7_8666_979c49488982.slice - libcontainer container kubepods-besteffort-pod6c452799_bfc5_44e7_8666_979c49488982.slice. Mar 7 00:51:22.142667 systemd[1]: Created slice kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice - libcontainer container kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice. Mar 7 00:51:22.186632 kubelet[2593]: I0307 00:51:22.186591 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6c452799-bfc5-44e7-8666-979c49488982-xtables-lock\") pod \"kube-proxy-m6h8c\" (UID: \"6c452799-bfc5-44e7-8666-979c49488982\") " pod="kube-system/kube-proxy-m6h8c" Mar 7 00:51:22.186632 kubelet[2593]: I0307 00:51:22.186630 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbsx\" (UniqueName: \"kubernetes.io/projected/6c452799-bfc5-44e7-8666-979c49488982-kube-api-access-hqbsx\") pod \"kube-proxy-m6h8c\" (UID: \"6c452799-bfc5-44e7-8666-979c49488982\") " pod="kube-system/kube-proxy-m6h8c" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186651 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-bpf-maps\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186665 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cni-path\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186699 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-lib-modules\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186715 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/6c452799-bfc5-44e7-8666-979c49488982-kube-proxy\") pod \"kube-proxy-m6h8c\" (UID: \"6c452799-bfc5-44e7-8666-979c49488982\") " pod="kube-system/kube-proxy-m6h8c" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186735 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6c452799-bfc5-44e7-8666-979c49488982-lib-modules\") pod \"kube-proxy-m6h8c\" (UID: \"6c452799-bfc5-44e7-8666-979c49488982\") " pod="kube-system/kube-proxy-m6h8c" Mar 7 00:51:22.187035 kubelet[2593]: I0307 00:51:22.186800 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-hostproc\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187194 kubelet[2593]: I0307 00:51:22.186819 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-cgroup\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187194 kubelet[2593]: I0307 00:51:22.186832 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2b37e447-bc99-4e82-8d7e-5c64e610107f-clustermesh-secrets\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187194 kubelet[2593]: I0307 00:51:22.186852 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzq4\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-kube-api-access-swzq4\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187194 kubelet[2593]: I0307 00:51:22.186869 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-xtables-lock\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187194 kubelet[2593]: I0307 00:51:22.186883 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-config-path\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187300 kubelet[2593]: I0307 00:51:22.186897 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-net\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187300 kubelet[2593]: I0307 00:51:22.186912 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-kernel\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187300 kubelet[2593]: I0307 00:51:22.186926 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-run\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187300 kubelet[2593]: I0307 00:51:22.186945 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-etc-cni-netd\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.187300 kubelet[2593]: I0307 00:51:22.186960 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-hubble-tls\") pod \"cilium-g7nh5\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " pod="kube-system/cilium-g7nh5" Mar 7 00:51:22.362239 systemd[1]: Created slice kubepods-besteffort-pod93c1ce27_65f7_44f0_9c8f_520ddcd0e3d6.slice - libcontainer container kubepods-besteffort-pod93c1ce27_65f7_44f0_9c8f_520ddcd0e3d6.slice. Mar 7 00:51:22.388740 kubelet[2593]: I0307 00:51:22.388536 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-cilium-config-path\") pod \"cilium-operator-6f9c7c5859-9k24n\" (UID: \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\") " pod="kube-system/cilium-operator-6f9c7c5859-9k24n" Mar 7 00:51:22.388740 kubelet[2593]: I0307 00:51:22.388676 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj7gr\" (UniqueName: \"kubernetes.io/projected/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-kube-api-access-sj7gr\") pod \"cilium-operator-6f9c7c5859-9k24n\" (UID: \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\") " pod="kube-system/cilium-operator-6f9c7c5859-9k24n" Mar 7 00:51:22.442324 containerd[1484]: time="2026-03-07T00:51:22.441825225Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-m6h8c,Uid:6c452799-bfc5-44e7-8666-979c49488982,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:22.448978 containerd[1484]: time="2026-03-07T00:51:22.448936216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-g7nh5,Uid:2b37e447-bc99-4e82-8d7e-5c64e610107f,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:22.474828 containerd[1484]: time="2026-03-07T00:51:22.473942001Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:22.474828 containerd[1484]: time="2026-03-07T00:51:22.474007272Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:22.474828 containerd[1484]: time="2026-03-07T00:51:22.474018157Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.474828 containerd[1484]: time="2026-03-07T00:51:22.474100516Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.482992 containerd[1484]: time="2026-03-07T00:51:22.482874050Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:22.482992 containerd[1484]: time="2026-03-07T00:51:22.482950806Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:22.482992 containerd[1484]: time="2026-03-07T00:51:22.482967935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.483364 containerd[1484]: time="2026-03-07T00:51:22.483281082Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.499797 systemd[1]: Started cri-containerd-3dc96eb6fe217fd4f1fed7b7f3e8236a234e53aef301ae1709967107bb7703df.scope - libcontainer container 3dc96eb6fe217fd4f1fed7b7f3e8236a234e53aef301ae1709967107bb7703df. Mar 7 00:51:22.517859 systemd[1]: Started cri-containerd-8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467.scope - libcontainer container 8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467. Mar 7 00:51:22.549045 containerd[1484]: time="2026-03-07T00:51:22.548988368Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-m6h8c,Uid:6c452799-bfc5-44e7-8666-979c49488982,Namespace:kube-system,Attempt:0,} returns sandbox id \"3dc96eb6fe217fd4f1fed7b7f3e8236a234e53aef301ae1709967107bb7703df\"" Mar 7 00:51:22.558328 containerd[1484]: time="2026-03-07T00:51:22.558283589Z" level=info msg="CreateContainer within sandbox \"3dc96eb6fe217fd4f1fed7b7f3e8236a234e53aef301ae1709967107bb7703df\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 7 00:51:22.560196 containerd[1484]: time="2026-03-07T00:51:22.559910556Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-g7nh5,Uid:2b37e447-bc99-4e82-8d7e-5c64e610107f,Namespace:kube-system,Attempt:0,} returns sandbox id \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\"" Mar 7 00:51:22.562312 containerd[1484]: time="2026-03-07T00:51:22.562189830Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 7 00:51:22.580820 containerd[1484]: time="2026-03-07T00:51:22.580764183Z" level=info msg="CreateContainer within sandbox \"3dc96eb6fe217fd4f1fed7b7f3e8236a234e53aef301ae1709967107bb7703df\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8be8a615423bbbb1cd58d93b5b3d73a50fa2a13fdcad65b791c80f43bb60e7ec\"" Mar 7 00:51:22.582382 containerd[1484]: time="2026-03-07T00:51:22.582099773Z" level=info msg="StartContainer for \"8be8a615423bbbb1cd58d93b5b3d73a50fa2a13fdcad65b791c80f43bb60e7ec\"" Mar 7 00:51:22.607789 systemd[1]: Started cri-containerd-8be8a615423bbbb1cd58d93b5b3d73a50fa2a13fdcad65b791c80f43bb60e7ec.scope - libcontainer container 8be8a615423bbbb1cd58d93b5b3d73a50fa2a13fdcad65b791c80f43bb60e7ec. Mar 7 00:51:22.640247 containerd[1484]: time="2026-03-07T00:51:22.640124599Z" level=info msg="StartContainer for \"8be8a615423bbbb1cd58d93b5b3d73a50fa2a13fdcad65b791c80f43bb60e7ec\" returns successfully" Mar 7 00:51:22.673543 containerd[1484]: time="2026-03-07T00:51:22.673490963Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6f9c7c5859-9k24n,Uid:93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:22.703306 containerd[1484]: time="2026-03-07T00:51:22.703205247Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:22.703591 containerd[1484]: time="2026-03-07T00:51:22.703308055Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:22.703591 containerd[1484]: time="2026-03-07T00:51:22.703321782Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.703591 containerd[1484]: time="2026-03-07T00:51:22.703421789Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:22.720753 systemd[1]: Started cri-containerd-e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69.scope - libcontainer container e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69. Mar 7 00:51:22.748653 kubelet[2593]: I0307 00:51:22.747848 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-m6h8c" podStartSLOduration=0.747833119 podStartE2EDuration="747.833119ms" podCreationTimestamp="2026-03-07 00:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:22.745929222 +0000 UTC m=+6.220473504" watchObservedRunningTime="2026-03-07 00:51:22.747833119 +0000 UTC m=+6.222377321" Mar 7 00:51:22.772458 containerd[1484]: time="2026-03-07T00:51:22.772415904Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6f9c7c5859-9k24n,Uid:93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6,Namespace:kube-system,Attempt:0,} returns sandbox id \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\"" Mar 7 00:51:26.054264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1665155285.mount: Deactivated successfully. Mar 7 00:51:27.523132 containerd[1484]: time="2026-03-07T00:51:27.523058739Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:51:27.525896 containerd[1484]: time="2026-03-07T00:51:27.525828190Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Mar 7 00:51:27.527053 containerd[1484]: time="2026-03-07T00:51:27.527024787Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:51:27.529332 containerd[1484]: time="2026-03-07T00:51:27.529303058Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 4.967060405s" Mar 7 00:51:27.529443 containerd[1484]: time="2026-03-07T00:51:27.529424983Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Mar 7 00:51:27.532653 containerd[1484]: time="2026-03-07T00:51:27.532622230Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 7 00:51:27.537348 containerd[1484]: time="2026-03-07T00:51:27.537208304Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 7 00:51:27.553815 containerd[1484]: time="2026-03-07T00:51:27.553768350Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\"" Mar 7 00:51:27.556238 containerd[1484]: time="2026-03-07T00:51:27.554500617Z" level=info msg="StartContainer for \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\"" Mar 7 00:51:27.583103 systemd[1]: run-containerd-runc-k8s.io-2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e-runc.tX8TkD.mount: Deactivated successfully. Mar 7 00:51:27.592933 systemd[1]: Started cri-containerd-2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e.scope - libcontainer container 2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e. Mar 7 00:51:27.624381 containerd[1484]: time="2026-03-07T00:51:27.624269208Z" level=info msg="StartContainer for \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\" returns successfully" Mar 7 00:51:27.638082 systemd[1]: cri-containerd-2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e.scope: Deactivated successfully. Mar 7 00:51:27.813009 containerd[1484]: time="2026-03-07T00:51:27.812734091Z" level=info msg="shim disconnected" id=2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e namespace=k8s.io Mar 7 00:51:27.813009 containerd[1484]: time="2026-03-07T00:51:27.812794473Z" level=warning msg="cleaning up after shim disconnected" id=2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e namespace=k8s.io Mar 7 00:51:27.813009 containerd[1484]: time="2026-03-07T00:51:27.812805437Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:51:28.551474 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e-rootfs.mount: Deactivated successfully. Mar 7 00:51:28.764348 containerd[1484]: time="2026-03-07T00:51:28.763731254Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 7 00:51:28.790523 containerd[1484]: time="2026-03-07T00:51:28.790455024Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\"" Mar 7 00:51:28.792585 containerd[1484]: time="2026-03-07T00:51:28.792090273Z" level=info msg="StartContainer for \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\"" Mar 7 00:51:28.826950 systemd[1]: Started cri-containerd-ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238.scope - libcontainer container ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238. Mar 7 00:51:28.860703 containerd[1484]: time="2026-03-07T00:51:28.860647226Z" level=info msg="StartContainer for \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\" returns successfully" Mar 7 00:51:28.877662 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 7 00:51:28.877937 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 7 00:51:28.878005 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 7 00:51:28.886760 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 7 00:51:28.888545 systemd[1]: cri-containerd-ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238.scope: Deactivated successfully. Mar 7 00:51:28.913655 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 7 00:51:28.924131 containerd[1484]: time="2026-03-07T00:51:28.924027620Z" level=info msg="shim disconnected" id=ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238 namespace=k8s.io Mar 7 00:51:28.924131 containerd[1484]: time="2026-03-07T00:51:28.924090962Z" level=warning msg="cleaning up after shim disconnected" id=ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238 namespace=k8s.io Mar 7 00:51:28.924131 containerd[1484]: time="2026-03-07T00:51:28.924099965Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:51:29.552329 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238-rootfs.mount: Deactivated successfully. Mar 7 00:51:29.633399 containerd[1484]: time="2026-03-07T00:51:29.633323117Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:51:29.635844 containerd[1484]: time="2026-03-07T00:51:29.635787773Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Mar 7 00:51:29.636986 containerd[1484]: time="2026-03-07T00:51:29.636956280Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 7 00:51:29.639046 containerd[1484]: time="2026-03-07T00:51:29.639011481Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.106240717s" Mar 7 00:51:29.639046 containerd[1484]: time="2026-03-07T00:51:29.639048333Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Mar 7 00:51:29.644378 containerd[1484]: time="2026-03-07T00:51:29.644340086Z" level=info msg="CreateContainer within sandbox \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 7 00:51:29.661444 containerd[1484]: time="2026-03-07T00:51:29.661369808Z" level=info msg="CreateContainer within sandbox \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\"" Mar 7 00:51:29.663727 containerd[1484]: time="2026-03-07T00:51:29.662891793Z" level=info msg="StartContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\"" Mar 7 00:51:29.695817 systemd[1]: Started cri-containerd-98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76.scope - libcontainer container 98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76. Mar 7 00:51:29.723349 containerd[1484]: time="2026-03-07T00:51:29.723140513Z" level=info msg="StartContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" returns successfully" Mar 7 00:51:29.774850 containerd[1484]: time="2026-03-07T00:51:29.774688791Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 7 00:51:29.806667 containerd[1484]: time="2026-03-07T00:51:29.805080700Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\"" Mar 7 00:51:29.808482 containerd[1484]: time="2026-03-07T00:51:29.808433571Z" level=info msg="StartContainer for \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\"" Mar 7 00:51:29.812774 kubelet[2593]: I0307 00:51:29.812701 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6f9c7c5859-9k24n" podStartSLOduration=0.946346724 podStartE2EDuration="7.812153643s" podCreationTimestamp="2026-03-07 00:51:22 +0000 UTC" firstStartedPulling="2026-03-07 00:51:22.774391915 +0000 UTC m=+6.248936077" lastFinishedPulling="2026-03-07 00:51:29.640198754 +0000 UTC m=+13.114742996" observedRunningTime="2026-03-07 00:51:29.81048349 +0000 UTC m=+13.285027692" watchObservedRunningTime="2026-03-07 00:51:29.812153643 +0000 UTC m=+13.286697845" Mar 7 00:51:29.845450 systemd[1]: Started cri-containerd-818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909.scope - libcontainer container 818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909. Mar 7 00:51:29.887622 containerd[1484]: time="2026-03-07T00:51:29.886590704Z" level=info msg="StartContainer for \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\" returns successfully" Mar 7 00:51:29.893735 systemd[1]: cri-containerd-818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909.scope: Deactivated successfully. Mar 7 00:51:29.969601 containerd[1484]: time="2026-03-07T00:51:29.969102560Z" level=info msg="shim disconnected" id=818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909 namespace=k8s.io Mar 7 00:51:29.969873 containerd[1484]: time="2026-03-07T00:51:29.969849368Z" level=warning msg="cleaning up after shim disconnected" id=818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909 namespace=k8s.io Mar 7 00:51:29.969954 containerd[1484]: time="2026-03-07T00:51:29.969941198Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:51:30.782366 containerd[1484]: time="2026-03-07T00:51:30.782120299Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 7 00:51:30.801665 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2601912196.mount: Deactivated successfully. Mar 7 00:51:30.804597 containerd[1484]: time="2026-03-07T00:51:30.804431629Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\"" Mar 7 00:51:30.806489 containerd[1484]: time="2026-03-07T00:51:30.805616363Z" level=info msg="StartContainer for \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\"" Mar 7 00:51:30.834763 systemd[1]: Started cri-containerd-e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a.scope - libcontainer container e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a. Mar 7 00:51:30.862450 systemd[1]: cri-containerd-e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a.scope: Deactivated successfully. Mar 7 00:51:30.863694 containerd[1484]: time="2026-03-07T00:51:30.863178592Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice/cri-containerd-e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a.scope/memory.events\": no such file or directory" Mar 7 00:51:30.868548 containerd[1484]: time="2026-03-07T00:51:30.868262998Z" level=info msg="StartContainer for \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\" returns successfully" Mar 7 00:51:30.899813 containerd[1484]: time="2026-03-07T00:51:30.899644794Z" level=info msg="shim disconnected" id=e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a namespace=k8s.io Mar 7 00:51:30.899813 containerd[1484]: time="2026-03-07T00:51:30.899735023Z" level=warning msg="cleaning up after shim disconnected" id=e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a namespace=k8s.io Mar 7 00:51:30.899813 containerd[1484]: time="2026-03-07T00:51:30.899751908Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:51:31.551564 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a-rootfs.mount: Deactivated successfully. Mar 7 00:51:31.792375 containerd[1484]: time="2026-03-07T00:51:31.792299242Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 7 00:51:31.810082 containerd[1484]: time="2026-03-07T00:51:31.809019526Z" level=info msg="CreateContainer within sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\"" Mar 7 00:51:31.810082 containerd[1484]: time="2026-03-07T00:51:31.809867781Z" level=info msg="StartContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\"" Mar 7 00:51:31.846949 systemd[1]: Started cri-containerd-90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a.scope - libcontainer container 90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a. Mar 7 00:51:31.879738 containerd[1484]: time="2026-03-07T00:51:31.879592932Z" level=info msg="StartContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" returns successfully" Mar 7 00:51:32.022959 kubelet[2593]: I0307 00:51:32.022925 2593 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Mar 7 00:51:32.067138 systemd[1]: Created slice kubepods-burstable-podb984189f_a243_427b_8749_6dc025ee22a4.slice - libcontainer container kubepods-burstable-podb984189f_a243_427b_8749_6dc025ee22a4.slice. Mar 7 00:51:32.077612 systemd[1]: Created slice kubepods-burstable-pod6b5b66a4_cfeb_4f2b_bf10_5fd09c9cc78f.slice - libcontainer container kubepods-burstable-pod6b5b66a4_cfeb_4f2b_bf10_5fd09c9cc78f.slice. Mar 7 00:51:32.159220 kubelet[2593]: I0307 00:51:32.159170 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f-config-volume\") pod \"coredns-66bc5c9577-v2ktj\" (UID: \"6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f\") " pod="kube-system/coredns-66bc5c9577-v2ktj" Mar 7 00:51:32.159220 kubelet[2593]: I0307 00:51:32.159222 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b984189f-a243-427b-8749-6dc025ee22a4-config-volume\") pod \"coredns-66bc5c9577-5lxpj\" (UID: \"b984189f-a243-427b-8749-6dc025ee22a4\") " pod="kube-system/coredns-66bc5c9577-5lxpj" Mar 7 00:51:32.159472 kubelet[2593]: I0307 00:51:32.159240 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxml5\" (UniqueName: \"kubernetes.io/projected/b984189f-a243-427b-8749-6dc025ee22a4-kube-api-access-xxml5\") pod \"coredns-66bc5c9577-5lxpj\" (UID: \"b984189f-a243-427b-8749-6dc025ee22a4\") " pod="kube-system/coredns-66bc5c9577-5lxpj" Mar 7 00:51:32.159472 kubelet[2593]: I0307 00:51:32.159259 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jthxm\" (UniqueName: \"kubernetes.io/projected/6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f-kube-api-access-jthxm\") pod \"coredns-66bc5c9577-v2ktj\" (UID: \"6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f\") " pod="kube-system/coredns-66bc5c9577-v2ktj" Mar 7 00:51:32.376192 containerd[1484]: time="2026-03-07T00:51:32.375853011Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-5lxpj,Uid:b984189f-a243-427b-8749-6dc025ee22a4,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:32.383424 containerd[1484]: time="2026-03-07T00:51:32.383150634Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-v2ktj,Uid:6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f,Namespace:kube-system,Attempt:0,}" Mar 7 00:51:34.175090 systemd-networkd[1380]: cilium_host: Link UP Mar 7 00:51:34.176057 systemd-networkd[1380]: cilium_net: Link UP Mar 7 00:51:34.177738 systemd-networkd[1380]: cilium_net: Gained carrier Mar 7 00:51:34.177923 systemd-networkd[1380]: cilium_host: Gained carrier Mar 7 00:51:34.178094 systemd-networkd[1380]: cilium_net: Gained IPv6LL Mar 7 00:51:34.178250 systemd-networkd[1380]: cilium_host: Gained IPv6LL Mar 7 00:51:34.286734 systemd-networkd[1380]: cilium_vxlan: Link UP Mar 7 00:51:34.286743 systemd-networkd[1380]: cilium_vxlan: Gained carrier Mar 7 00:51:34.599886 kernel: NET: Registered PF_ALG protocol family Mar 7 00:51:35.359138 systemd-networkd[1380]: lxc_health: Link UP Mar 7 00:51:35.374732 systemd-networkd[1380]: lxc_health: Gained carrier Mar 7 00:51:35.927127 systemd-networkd[1380]: lxcba0c9d521cd0: Link UP Mar 7 00:51:35.932896 kernel: eth0: renamed from tmp7ea57 Mar 7 00:51:35.938783 systemd-networkd[1380]: lxcba0c9d521cd0: Gained carrier Mar 7 00:51:35.954803 systemd-networkd[1380]: lxcad6778bb697e: Link UP Mar 7 00:51:35.962179 kernel: eth0: renamed from tmp9ac49 Mar 7 00:51:35.967270 systemd-networkd[1380]: lxcad6778bb697e: Gained carrier Mar 7 00:51:36.275734 systemd-networkd[1380]: cilium_vxlan: Gained IPv6LL Mar 7 00:51:36.474500 kubelet[2593]: I0307 00:51:36.474442 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-g7nh5" podStartSLOduration=9.505047695 podStartE2EDuration="14.474426325s" podCreationTimestamp="2026-03-07 00:51:22 +0000 UTC" firstStartedPulling="2026-03-07 00:51:22.561678829 +0000 UTC m=+6.036223031" lastFinishedPulling="2026-03-07 00:51:27.531057419 +0000 UTC m=+11.005601661" observedRunningTime="2026-03-07 00:51:32.812490828 +0000 UTC m=+16.287035070" watchObservedRunningTime="2026-03-07 00:51:36.474426325 +0000 UTC m=+19.948970527" Mar 7 00:51:37.107794 systemd-networkd[1380]: lxcad6778bb697e: Gained IPv6LL Mar 7 00:51:37.173879 systemd-networkd[1380]: lxc_health: Gained IPv6LL Mar 7 00:51:37.619896 systemd-networkd[1380]: lxcba0c9d521cd0: Gained IPv6LL Mar 7 00:51:39.968278 containerd[1484]: time="2026-03-07T00:51:39.960893353Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:39.968278 containerd[1484]: time="2026-03-07T00:51:39.960952406Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:39.968278 containerd[1484]: time="2026-03-07T00:51:39.960976571Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:39.968278 containerd[1484]: time="2026-03-07T00:51:39.961066671Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:39.988369 containerd[1484]: time="2026-03-07T00:51:39.987950443Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:51:39.988369 containerd[1484]: time="2026-03-07T00:51:39.988065308Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:51:39.988369 containerd[1484]: time="2026-03-07T00:51:39.988079511Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:39.988369 containerd[1484]: time="2026-03-07T00:51:39.988198256Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:51:39.990792 systemd[1]: Started cri-containerd-9ac49ed3c64bc77007042fc0ec5d8144e0346fbf7e7bba796c510a4d3e86ed55.scope - libcontainer container 9ac49ed3c64bc77007042fc0ec5d8144e0346fbf7e7bba796c510a4d3e86ed55. Mar 7 00:51:40.028803 systemd[1]: Started cri-containerd-7ea577ca88d9efa263210e1d2ab70eeef926d2009b0234f88024c9d161998ffa.scope - libcontainer container 7ea577ca88d9efa263210e1d2ab70eeef926d2009b0234f88024c9d161998ffa. Mar 7 00:51:40.060195 containerd[1484]: time="2026-03-07T00:51:40.060052614Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-v2ktj,Uid:6b5b66a4-cfeb-4f2b-bf10-5fd09c9cc78f,Namespace:kube-system,Attempt:0,} returns sandbox id \"9ac49ed3c64bc77007042fc0ec5d8144e0346fbf7e7bba796c510a4d3e86ed55\"" Mar 7 00:51:40.068272 containerd[1484]: time="2026-03-07T00:51:40.067521118Z" level=info msg="CreateContainer within sandbox \"9ac49ed3c64bc77007042fc0ec5d8144e0346fbf7e7bba796c510a4d3e86ed55\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 00:51:40.095343 containerd[1484]: time="2026-03-07T00:51:40.095010039Z" level=info msg="CreateContainer within sandbox \"9ac49ed3c64bc77007042fc0ec5d8144e0346fbf7e7bba796c510a4d3e86ed55\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9a6bb02121ec7747f063bc6a016fc05b64bd896ef85560e56e45f27068a5189c\"" Mar 7 00:51:40.096721 containerd[1484]: time="2026-03-07T00:51:40.096328472Z" level=info msg="StartContainer for \"9a6bb02121ec7747f063bc6a016fc05b64bd896ef85560e56e45f27068a5189c\"" Mar 7 00:51:40.102544 containerd[1484]: time="2026-03-07T00:51:40.102500627Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-5lxpj,Uid:b984189f-a243-427b-8749-6dc025ee22a4,Namespace:kube-system,Attempt:0,} returns sandbox id \"7ea577ca88d9efa263210e1d2ab70eeef926d2009b0234f88024c9d161998ffa\"" Mar 7 00:51:40.111602 containerd[1484]: time="2026-03-07T00:51:40.111429513Z" level=info msg="CreateContainer within sandbox \"7ea577ca88d9efa263210e1d2ab70eeef926d2009b0234f88024c9d161998ffa\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 7 00:51:40.130112 containerd[1484]: time="2026-03-07T00:51:40.130040879Z" level=info msg="CreateContainer within sandbox \"7ea577ca88d9efa263210e1d2ab70eeef926d2009b0234f88024c9d161998ffa\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ff046882d2142558b1dba0860af67eb14e585ee706b9ee55ef38829163f0e35c\"" Mar 7 00:51:40.132616 containerd[1484]: time="2026-03-07T00:51:40.132445856Z" level=info msg="StartContainer for \"ff046882d2142558b1dba0860af67eb14e585ee706b9ee55ef38829163f0e35c\"" Mar 7 00:51:40.146081 systemd[1]: Started cri-containerd-9a6bb02121ec7747f063bc6a016fc05b64bd896ef85560e56e45f27068a5189c.scope - libcontainer container 9a6bb02121ec7747f063bc6a016fc05b64bd896ef85560e56e45f27068a5189c. Mar 7 00:51:40.182028 systemd[1]: Started cri-containerd-ff046882d2142558b1dba0860af67eb14e585ee706b9ee55ef38829163f0e35c.scope - libcontainer container ff046882d2142558b1dba0860af67eb14e585ee706b9ee55ef38829163f0e35c. Mar 7 00:51:40.203438 containerd[1484]: time="2026-03-07T00:51:40.203347150Z" level=info msg="StartContainer for \"9a6bb02121ec7747f063bc6a016fc05b64bd896ef85560e56e45f27068a5189c\" returns successfully" Mar 7 00:51:40.221410 containerd[1484]: time="2026-03-07T00:51:40.221250690Z" level=info msg="StartContainer for \"ff046882d2142558b1dba0860af67eb14e585ee706b9ee55ef38829163f0e35c\" returns successfully" Mar 7 00:51:40.857301 kubelet[2593]: I0307 00:51:40.856133 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-v2ktj" podStartSLOduration=18.856110622 podStartE2EDuration="18.856110622s" podCreationTimestamp="2026-03-07 00:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:40.830685167 +0000 UTC m=+24.305229369" watchObservedRunningTime="2026-03-07 00:51:40.856110622 +0000 UTC m=+24.330654904" Mar 7 00:51:44.775668 kubelet[2593]: I0307 00:51:44.775395 2593 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 7 00:51:44.801604 kubelet[2593]: I0307 00:51:44.799971 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-5lxpj" podStartSLOduration=22.799951292 podStartE2EDuration="22.799951292s" podCreationTimestamp="2026-03-07 00:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:51:40.877563616 +0000 UTC m=+24.352107858" watchObservedRunningTime="2026-03-07 00:51:44.799951292 +0000 UTC m=+28.274495494" Mar 7 00:53:35.783977 systemd[1]: Started sshd@7-188.245.50.81:22-20.161.92.111:35474.service - OpenSSH per-connection server daemon (20.161.92.111:35474). Mar 7 00:53:36.378255 sshd[4003]: Accepted publickey for core from 20.161.92.111 port 35474 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:36.380727 sshd[4003]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:36.387225 systemd-logind[1455]: New session 8 of user core. Mar 7 00:53:36.392817 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 7 00:53:36.895262 sshd[4003]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:36.900660 systemd[1]: sshd@7-188.245.50.81:22-20.161.92.111:35474.service: Deactivated successfully. Mar 7 00:53:36.904264 systemd[1]: session-8.scope: Deactivated successfully. Mar 7 00:53:36.905244 systemd-logind[1455]: Session 8 logged out. Waiting for processes to exit. Mar 7 00:53:36.906410 systemd-logind[1455]: Removed session 8. Mar 7 00:53:42.006964 systemd[1]: Started sshd@8-188.245.50.81:22-20.161.92.111:38892.service - OpenSSH per-connection server daemon (20.161.92.111:38892). Mar 7 00:53:42.598223 sshd[4017]: Accepted publickey for core from 20.161.92.111 port 38892 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:42.600755 sshd[4017]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:42.608526 systemd-logind[1455]: New session 9 of user core. Mar 7 00:53:42.614755 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 7 00:53:43.099976 sshd[4017]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:43.104608 systemd[1]: sshd@8-188.245.50.81:22-20.161.92.111:38892.service: Deactivated successfully. Mar 7 00:53:43.107142 systemd[1]: session-9.scope: Deactivated successfully. Mar 7 00:53:43.108531 systemd-logind[1455]: Session 9 logged out. Waiting for processes to exit. Mar 7 00:53:43.109685 systemd-logind[1455]: Removed session 9. Mar 7 00:53:48.214873 systemd[1]: Started sshd@9-188.245.50.81:22-20.161.92.111:38894.service - OpenSSH per-connection server daemon (20.161.92.111:38894). Mar 7 00:53:48.797401 sshd[4031]: Accepted publickey for core from 20.161.92.111 port 38894 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:48.800980 sshd[4031]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:48.806143 systemd-logind[1455]: New session 10 of user core. Mar 7 00:53:48.816126 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 7 00:53:49.290224 sshd[4031]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:49.295250 systemd[1]: sshd@9-188.245.50.81:22-20.161.92.111:38894.service: Deactivated successfully. Mar 7 00:53:49.297254 systemd[1]: session-10.scope: Deactivated successfully. Mar 7 00:53:49.299156 systemd-logind[1455]: Session 10 logged out. Waiting for processes to exit. Mar 7 00:53:49.300221 systemd-logind[1455]: Removed session 10. Mar 7 00:53:49.401308 systemd[1]: Started sshd@10-188.245.50.81:22-20.161.92.111:38902.service - OpenSSH per-connection server daemon (20.161.92.111:38902). Mar 7 00:53:49.984008 sshd[4044]: Accepted publickey for core from 20.161.92.111 port 38902 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:49.985876 sshd[4044]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:49.991620 systemd-logind[1455]: New session 11 of user core. Mar 7 00:53:50.001966 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 7 00:53:50.512039 sshd[4044]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:50.519211 systemd[1]: sshd@10-188.245.50.81:22-20.161.92.111:38902.service: Deactivated successfully. Mar 7 00:53:50.522004 systemd[1]: session-11.scope: Deactivated successfully. Mar 7 00:53:50.524318 systemd-logind[1455]: Session 11 logged out. Waiting for processes to exit. Mar 7 00:53:50.525982 systemd-logind[1455]: Removed session 11. Mar 7 00:53:50.624050 systemd[1]: Started sshd@11-188.245.50.81:22-20.161.92.111:60970.service - OpenSSH per-connection server daemon (20.161.92.111:60970). Mar 7 00:53:51.205941 sshd[4054]: Accepted publickey for core from 20.161.92.111 port 60970 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:51.208013 sshd[4054]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:51.214424 systemd-logind[1455]: New session 12 of user core. Mar 7 00:53:51.222916 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 7 00:53:51.705642 sshd[4054]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:51.710188 systemd[1]: sshd@11-188.245.50.81:22-20.161.92.111:60970.service: Deactivated successfully. Mar 7 00:53:51.713907 systemd[1]: session-12.scope: Deactivated successfully. Mar 7 00:53:51.715733 systemd-logind[1455]: Session 12 logged out. Waiting for processes to exit. Mar 7 00:53:51.717274 systemd-logind[1455]: Removed session 12. Mar 7 00:53:56.826037 systemd[1]: Started sshd@12-188.245.50.81:22-20.161.92.111:60982.service - OpenSSH per-connection server daemon (20.161.92.111:60982). Mar 7 00:53:57.410914 sshd[4069]: Accepted publickey for core from 20.161.92.111 port 60982 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:57.413027 sshd[4069]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:57.417429 systemd-logind[1455]: New session 13 of user core. Mar 7 00:53:57.428067 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 7 00:53:57.901054 sshd[4069]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:57.906477 systemd[1]: sshd@12-188.245.50.81:22-20.161.92.111:60982.service: Deactivated successfully. Mar 7 00:53:57.908522 systemd[1]: session-13.scope: Deactivated successfully. Mar 7 00:53:57.910501 systemd-logind[1455]: Session 13 logged out. Waiting for processes to exit. Mar 7 00:53:57.911462 systemd-logind[1455]: Removed session 13. Mar 7 00:53:58.011936 systemd[1]: Started sshd@13-188.245.50.81:22-20.161.92.111:60988.service - OpenSSH per-connection server daemon (20.161.92.111:60988). Mar 7 00:53:58.595114 sshd[4082]: Accepted publickey for core from 20.161.92.111 port 60988 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:58.598479 sshd[4082]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:58.604792 systemd-logind[1455]: New session 14 of user core. Mar 7 00:53:58.611803 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 7 00:53:59.127516 sshd[4082]: pam_unix(sshd:session): session closed for user core Mar 7 00:53:59.134402 systemd[1]: sshd@13-188.245.50.81:22-20.161.92.111:60988.service: Deactivated successfully. Mar 7 00:53:59.137439 systemd[1]: session-14.scope: Deactivated successfully. Mar 7 00:53:59.139403 systemd-logind[1455]: Session 14 logged out. Waiting for processes to exit. Mar 7 00:53:59.140828 systemd-logind[1455]: Removed session 14. Mar 7 00:53:59.232634 systemd[1]: Started sshd@14-188.245.50.81:22-20.161.92.111:60998.service - OpenSSH per-connection server daemon (20.161.92.111:60998). Mar 7 00:53:59.839919 sshd[4092]: Accepted publickey for core from 20.161.92.111 port 60998 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:53:59.842699 sshd[4092]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:53:59.848507 systemd-logind[1455]: New session 15 of user core. Mar 7 00:53:59.857995 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 7 00:54:00.904499 sshd[4092]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:00.911458 systemd[1]: sshd@14-188.245.50.81:22-20.161.92.111:60998.service: Deactivated successfully. Mar 7 00:54:00.914400 systemd[1]: session-15.scope: Deactivated successfully. Mar 7 00:54:00.916045 systemd-logind[1455]: Session 15 logged out. Waiting for processes to exit. Mar 7 00:54:00.918250 systemd-logind[1455]: Removed session 15. Mar 7 00:54:01.015972 systemd[1]: Started sshd@15-188.245.50.81:22-20.161.92.111:49392.service - OpenSSH per-connection server daemon (20.161.92.111:49392). Mar 7 00:54:01.598623 sshd[4109]: Accepted publickey for core from 20.161.92.111 port 49392 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:01.600908 sshd[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:01.606438 systemd-logind[1455]: New session 16 of user core. Mar 7 00:54:01.615866 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 7 00:54:02.211616 sshd[4109]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:02.216737 systemd[1]: sshd@15-188.245.50.81:22-20.161.92.111:49392.service: Deactivated successfully. Mar 7 00:54:02.220454 systemd[1]: session-16.scope: Deactivated successfully. Mar 7 00:54:02.222301 systemd-logind[1455]: Session 16 logged out. Waiting for processes to exit. Mar 7 00:54:02.223703 systemd-logind[1455]: Removed session 16. Mar 7 00:54:02.322087 systemd[1]: Started sshd@16-188.245.50.81:22-20.161.92.111:49408.service - OpenSSH per-connection server daemon (20.161.92.111:49408). Mar 7 00:54:02.907563 sshd[4122]: Accepted publickey for core from 20.161.92.111 port 49408 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:02.908938 sshd[4122]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:02.913901 systemd-logind[1455]: New session 17 of user core. Mar 7 00:54:02.925863 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 7 00:54:03.392910 sshd[4122]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:03.397640 systemd[1]: sshd@16-188.245.50.81:22-20.161.92.111:49408.service: Deactivated successfully. Mar 7 00:54:03.399464 systemd[1]: session-17.scope: Deactivated successfully. Mar 7 00:54:03.402857 systemd-logind[1455]: Session 17 logged out. Waiting for processes to exit. Mar 7 00:54:03.404086 systemd-logind[1455]: Removed session 17. Mar 7 00:54:08.498949 systemd[1]: Started sshd@17-188.245.50.81:22-20.161.92.111:49424.service - OpenSSH per-connection server daemon (20.161.92.111:49424). Mar 7 00:54:09.082304 sshd[4137]: Accepted publickey for core from 20.161.92.111 port 49424 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:09.084437 sshd[4137]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:09.093864 systemd-logind[1455]: New session 18 of user core. Mar 7 00:54:09.103896 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 7 00:54:09.567161 sshd[4137]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:09.573122 systemd[1]: sshd@17-188.245.50.81:22-20.161.92.111:49424.service: Deactivated successfully. Mar 7 00:54:09.577424 systemd[1]: session-18.scope: Deactivated successfully. Mar 7 00:54:09.578469 systemd-logind[1455]: Session 18 logged out. Waiting for processes to exit. Mar 7 00:54:09.580536 systemd-logind[1455]: Removed session 18. Mar 7 00:54:14.678897 systemd[1]: Started sshd@18-188.245.50.81:22-20.161.92.111:44858.service - OpenSSH per-connection server daemon (20.161.92.111:44858). Mar 7 00:54:15.266766 sshd[4151]: Accepted publickey for core from 20.161.92.111 port 44858 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:15.268937 sshd[4151]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:15.275899 systemd-logind[1455]: New session 19 of user core. Mar 7 00:54:15.285181 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 7 00:54:15.755113 sshd[4151]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:15.760062 systemd[1]: sshd@18-188.245.50.81:22-20.161.92.111:44858.service: Deactivated successfully. Mar 7 00:54:15.764093 systemd[1]: session-19.scope: Deactivated successfully. Mar 7 00:54:15.764887 systemd-logind[1455]: Session 19 logged out. Waiting for processes to exit. Mar 7 00:54:15.766121 systemd-logind[1455]: Removed session 19. Mar 7 00:54:15.865020 systemd[1]: Started sshd@19-188.245.50.81:22-20.161.92.111:44860.service - OpenSSH per-connection server daemon (20.161.92.111:44860). Mar 7 00:54:16.457860 sshd[4164]: Accepted publickey for core from 20.161.92.111 port 44860 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:16.460482 sshd[4164]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:16.465628 systemd-logind[1455]: New session 20 of user core. Mar 7 00:54:16.471845 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 7 00:54:18.742272 containerd[1484]: time="2026-03-07T00:54:18.742223016Z" level=info msg="StopContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" with timeout 30 (s)" Mar 7 00:54:18.743246 containerd[1484]: time="2026-03-07T00:54:18.742952101Z" level=info msg="Stop container \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" with signal terminated" Mar 7 00:54:18.770396 containerd[1484]: time="2026-03-07T00:54:18.770334749Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 7 00:54:18.778703 systemd[1]: cri-containerd-98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76.scope: Deactivated successfully. Mar 7 00:54:18.788221 containerd[1484]: time="2026-03-07T00:54:18.788176089Z" level=info msg="StopContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" with timeout 2 (s)" Mar 7 00:54:18.790974 containerd[1484]: time="2026-03-07T00:54:18.790935939Z" level=info msg="Stop container \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" with signal terminated" Mar 7 00:54:18.800393 systemd-networkd[1380]: lxc_health: Link DOWN Mar 7 00:54:18.800874 systemd-networkd[1380]: lxc_health: Lost carrier Mar 7 00:54:18.820146 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76-rootfs.mount: Deactivated successfully. Mar 7 00:54:18.824949 systemd[1]: cri-containerd-90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a.scope: Deactivated successfully. Mar 7 00:54:18.825337 systemd[1]: cri-containerd-90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a.scope: Consumed 7.487s CPU time. Mar 7 00:54:18.836443 containerd[1484]: time="2026-03-07T00:54:18.836324976Z" level=info msg="shim disconnected" id=98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76 namespace=k8s.io Mar 7 00:54:18.836443 containerd[1484]: time="2026-03-07T00:54:18.836384820Z" level=warning msg="cleaning up after shim disconnected" id=98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76 namespace=k8s.io Mar 7 00:54:18.836443 containerd[1484]: time="2026-03-07T00:54:18.836392861Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:18.857081 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a-rootfs.mount: Deactivated successfully. Mar 7 00:54:18.865341 containerd[1484]: time="2026-03-07T00:54:18.865267880Z" level=info msg="shim disconnected" id=90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a namespace=k8s.io Mar 7 00:54:18.865545 containerd[1484]: time="2026-03-07T00:54:18.865463252Z" level=warning msg="cleaning up after shim disconnected" id=90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a namespace=k8s.io Mar 7 00:54:18.865545 containerd[1484]: time="2026-03-07T00:54:18.865480373Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:18.866030 containerd[1484]: time="2026-03-07T00:54:18.865986604Z" level=info msg="StopContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" returns successfully" Mar 7 00:54:18.866659 containerd[1484]: time="2026-03-07T00:54:18.866626084Z" level=info msg="StopPodSandbox for \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\"" Mar 7 00:54:18.866742 containerd[1484]: time="2026-03-07T00:54:18.866661526Z" level=info msg="Container to stop \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.868499 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69-shm.mount: Deactivated successfully. Mar 7 00:54:18.876957 systemd[1]: cri-containerd-e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69.scope: Deactivated successfully. Mar 7 00:54:18.888409 containerd[1484]: time="2026-03-07T00:54:18.888368904Z" level=info msg="StopContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" returns successfully" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889410728Z" level=info msg="StopPodSandbox for \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\"" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889450371Z" level=info msg="Container to stop \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889461651Z" level=info msg="Container to stop \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889470652Z" level=info msg="Container to stop \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889480053Z" level=info msg="Container to stop \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.889625 containerd[1484]: time="2026-03-07T00:54:18.889488893Z" level=info msg="Container to stop \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 7 00:54:18.892826 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467-shm.mount: Deactivated successfully. Mar 7 00:54:18.899311 systemd[1]: cri-containerd-8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467.scope: Deactivated successfully. Mar 7 00:54:18.923178 containerd[1484]: time="2026-03-07T00:54:18.923102845Z" level=info msg="shim disconnected" id=e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69 namespace=k8s.io Mar 7 00:54:18.923178 containerd[1484]: time="2026-03-07T00:54:18.923169769Z" level=warning msg="cleaning up after shim disconnected" id=e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69 namespace=k8s.io Mar 7 00:54:18.923178 containerd[1484]: time="2026-03-07T00:54:18.923182170Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:18.931603 containerd[1484]: time="2026-03-07T00:54:18.931300830Z" level=info msg="shim disconnected" id=8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467 namespace=k8s.io Mar 7 00:54:18.931603 containerd[1484]: time="2026-03-07T00:54:18.931542085Z" level=warning msg="cleaning up after shim disconnected" id=8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467 namespace=k8s.io Mar 7 00:54:18.931603 containerd[1484]: time="2026-03-07T00:54:18.931552246Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:18.943088 containerd[1484]: time="2026-03-07T00:54:18.942898545Z" level=info msg="TearDown network for sandbox \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\" successfully" Mar 7 00:54:18.943088 containerd[1484]: time="2026-03-07T00:54:18.942941188Z" level=info msg="StopPodSandbox for \"e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69\" returns successfully" Mar 7 00:54:18.952528 containerd[1484]: time="2026-03-07T00:54:18.951984865Z" level=info msg="TearDown network for sandbox \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" successfully" Mar 7 00:54:18.952528 containerd[1484]: time="2026-03-07T00:54:18.952018467Z" level=info msg="StopPodSandbox for \"8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467\" returns successfully" Mar 7 00:54:19.059661 kubelet[2593]: I0307 00:54:19.059141 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj7gr\" (UniqueName: \"kubernetes.io/projected/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-kube-api-access-sj7gr\") pod \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\" (UID: \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\") " Mar 7 00:54:19.059661 kubelet[2593]: I0307 00:54:19.059216 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-cilium-config-path\") pod \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\" (UID: \"93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6\") " Mar 7 00:54:19.064622 kubelet[2593]: I0307 00:54:19.064510 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-kube-api-access-sj7gr" (OuterVolumeSpecName: "kube-api-access-sj7gr") pod "93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6" (UID: "93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6"). InnerVolumeSpecName "kube-api-access-sj7gr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 7 00:54:19.064622 kubelet[2593]: I0307 00:54:19.064592 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6" (UID: "93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 00:54:19.160515 kubelet[2593]: I0307 00:54:19.160363 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-cgroup\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.160912 kubelet[2593]: I0307 00:54:19.160647 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-xtables-lock\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.160912 kubelet[2593]: I0307 00:54:19.160464 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.160912 kubelet[2593]: I0307 00:54:19.160776 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.161686 kubelet[2593]: I0307 00:54:19.161180 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-hubble-tls\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.161855 kubelet[2593]: I0307 00:54:19.161819 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-bpf-maps\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.161976 kubelet[2593]: I0307 00:54:19.161957 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-hostproc\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162134 kubelet[2593]: I0307 00:54:19.162111 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-kernel\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162220 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-lib-modules\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162250 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-etc-cni-netd\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162276 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cni-path\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162305 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2b37e447-bc99-4e82-8d7e-5c64e610107f-clustermesh-secrets\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162350 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-run\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.162619 kubelet[2593]: I0307 00:54:19.162384 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-config-path\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162416 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swzq4\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-kube-api-access-swzq4\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162441 2593 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-net\") pod \"2b37e447-bc99-4e82-8d7e-5c64e610107f\" (UID: \"2b37e447-bc99-4e82-8d7e-5c64e610107f\") " Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162500 2593 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sj7gr\" (UniqueName: \"kubernetes.io/projected/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-kube-api-access-sj7gr\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162515 2593 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-cgroup\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162529 2593 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-xtables-lock\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.163083 kubelet[2593]: I0307 00:54:19.162544 2593 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6-cilium-config-path\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.163748 kubelet[2593]: I0307 00:54:19.163332 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163748 kubelet[2593]: I0307 00:54:19.163375 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163748 kubelet[2593]: I0307 00:54:19.163393 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-hostproc" (OuterVolumeSpecName: "hostproc") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163748 kubelet[2593]: I0307 00:54:19.163408 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163748 kubelet[2593]: I0307 00:54:19.163427 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163982 kubelet[2593]: I0307 00:54:19.163448 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.163982 kubelet[2593]: I0307 00:54:19.163467 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cni-path" (OuterVolumeSpecName: "cni-path") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.164625 kubelet[2593]: I0307 00:54:19.164600 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 7 00:54:19.167198 kubelet[2593]: I0307 00:54:19.167173 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b37e447-bc99-4e82-8d7e-5c64e610107f-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 7 00:54:19.168838 kubelet[2593]: I0307 00:54:19.168218 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 7 00:54:19.168838 kubelet[2593]: I0307 00:54:19.168282 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 7 00:54:19.169399 kubelet[2593]: I0307 00:54:19.169375 2593 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-kube-api-access-swzq4" (OuterVolumeSpecName: "kube-api-access-swzq4") pod "2b37e447-bc99-4e82-8d7e-5c64e610107f" (UID: "2b37e447-bc99-4e82-8d7e-5c64e610107f"). InnerVolumeSpecName "kube-api-access-swzq4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 7 00:54:19.249602 kubelet[2593]: I0307 00:54:19.249466 2593 scope.go:117] "RemoveContainer" containerID="90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a" Mar 7 00:54:19.253594 containerd[1484]: time="2026-03-07T00:54:19.252904040Z" level=info msg="RemoveContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\"" Mar 7 00:54:19.258892 systemd[1]: Removed slice kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice - libcontainer container kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice. Mar 7 00:54:19.258993 systemd[1]: kubepods-burstable-pod2b37e447_bc99_4e82_8d7e_5c64e610107f.slice: Consumed 7.574s CPU time. Mar 7 00:54:19.262744 containerd[1484]: time="2026-03-07T00:54:19.262582398Z" level=info msg="RemoveContainer for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" returns successfully" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263516 2593 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-etc-cni-netd\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263538 2593 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cni-path\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263547 2593 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2b37e447-bc99-4e82-8d7e-5c64e610107f-clustermesh-secrets\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263555 2593 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-run\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263563 2593 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2b37e447-bc99-4e82-8d7e-5c64e610107f-cilium-config-path\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263594 2593 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-swzq4\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-kube-api-access-swzq4\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263603 2593 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-net\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264006 kubelet[2593]: I0307 00:54:19.263611 2593 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2b37e447-bc99-4e82-8d7e-5c64e610107f-hubble-tls\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264300 kubelet[2593]: I0307 00:54:19.263702 2593 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-bpf-maps\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264300 kubelet[2593]: I0307 00:54:19.263715 2593 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-hostproc\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264300 kubelet[2593]: I0307 00:54:19.263723 2593 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-host-proc-sys-kernel\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264300 kubelet[2593]: I0307 00:54:19.263731 2593 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b37e447-bc99-4e82-8d7e-5c64e610107f-lib-modules\") on node \"ci-4081-3-6-n-53d1697e27\" DevicePath \"\"" Mar 7 00:54:19.264300 kubelet[2593]: I0307 00:54:19.264117 2593 scope.go:117] "RemoveContainer" containerID="e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a" Mar 7 00:54:19.265014 systemd[1]: Removed slice kubepods-besteffort-pod93c1ce27_65f7_44f0_9c8f_520ddcd0e3d6.slice - libcontainer container kubepods-besteffort-pod93c1ce27_65f7_44f0_9c8f_520ddcd0e3d6.slice. Mar 7 00:54:19.267107 containerd[1484]: time="2026-03-07T00:54:19.267070995Z" level=info msg="RemoveContainer for \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\"" Mar 7 00:54:19.274088 containerd[1484]: time="2026-03-07T00:54:19.274021704Z" level=info msg="RemoveContainer for \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\" returns successfully" Mar 7 00:54:19.275773 kubelet[2593]: I0307 00:54:19.275239 2593 scope.go:117] "RemoveContainer" containerID="818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909" Mar 7 00:54:19.278840 containerd[1484]: time="2026-03-07T00:54:19.278786118Z" level=info msg="RemoveContainer for \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\"" Mar 7 00:54:19.285558 containerd[1484]: time="2026-03-07T00:54:19.285396767Z" level=info msg="RemoveContainer for \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\" returns successfully" Mar 7 00:54:19.287077 kubelet[2593]: I0307 00:54:19.287033 2593 scope.go:117] "RemoveContainer" containerID="ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238" Mar 7 00:54:19.291787 containerd[1484]: time="2026-03-07T00:54:19.291444140Z" level=info msg="RemoveContainer for \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\"" Mar 7 00:54:19.296683 containerd[1484]: time="2026-03-07T00:54:19.296629100Z" level=info msg="RemoveContainer for \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\" returns successfully" Mar 7 00:54:19.297014 kubelet[2593]: I0307 00:54:19.296991 2593 scope.go:117] "RemoveContainer" containerID="2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e" Mar 7 00:54:19.298598 containerd[1484]: time="2026-03-07T00:54:19.298190317Z" level=info msg="RemoveContainer for \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\"" Mar 7 00:54:19.306773 containerd[1484]: time="2026-03-07T00:54:19.306730764Z" level=info msg="RemoveContainer for \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\" returns successfully" Mar 7 00:54:19.308674 kubelet[2593]: I0307 00:54:19.308655 2593 scope.go:117] "RemoveContainer" containerID="90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a" Mar 7 00:54:19.309135 containerd[1484]: time="2026-03-07T00:54:19.309097270Z" level=error msg="ContainerStatus for \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\": not found" Mar 7 00:54:19.309460 kubelet[2593]: E0307 00:54:19.309438 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\": not found" containerID="90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a" Mar 7 00:54:19.309597 kubelet[2593]: I0307 00:54:19.309534 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a"} err="failed to get container status \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\": rpc error: code = NotFound desc = an error occurred when try to find container \"90a3f4c536d37ce19af1261263e670a4d565dc4b0f6624119f53a3045f96871a\": not found" Mar 7 00:54:19.309719 kubelet[2593]: I0307 00:54:19.309659 2593 scope.go:117] "RemoveContainer" containerID="e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a" Mar 7 00:54:19.311748 containerd[1484]: time="2026-03-07T00:54:19.311713472Z" level=error msg="ContainerStatus for \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\": not found" Mar 7 00:54:19.312121 kubelet[2593]: E0307 00:54:19.312094 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\": not found" containerID="e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a" Mar 7 00:54:19.312205 kubelet[2593]: I0307 00:54:19.312123 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a"} err="failed to get container status \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\": rpc error: code = NotFound desc = an error occurred when try to find container \"e5e3978a0a7622a824eb671e7f7da09b5791f5b7207f710c2a4ef9831822263a\": not found" Mar 7 00:54:19.312205 kubelet[2593]: I0307 00:54:19.312142 2593 scope.go:117] "RemoveContainer" containerID="818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909" Mar 7 00:54:19.312693 containerd[1484]: time="2026-03-07T00:54:19.312399354Z" level=error msg="ContainerStatus for \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\": not found" Mar 7 00:54:19.312867 kubelet[2593]: E0307 00:54:19.312771 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\": not found" containerID="818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909" Mar 7 00:54:19.312867 kubelet[2593]: I0307 00:54:19.312802 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909"} err="failed to get container status \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\": rpc error: code = NotFound desc = an error occurred when try to find container \"818158973cda8d534f16d53c4d341d0fd6282b373bc7af6433ff9e0ce0ac3909\": not found" Mar 7 00:54:19.312867 kubelet[2593]: I0307 00:54:19.312818 2593 scope.go:117] "RemoveContainer" containerID="ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238" Mar 7 00:54:19.314113 containerd[1484]: time="2026-03-07T00:54:19.314025054Z" level=error msg="ContainerStatus for \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\": not found" Mar 7 00:54:19.314305 kubelet[2593]: E0307 00:54:19.314285 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\": not found" containerID="ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238" Mar 7 00:54:19.314424 kubelet[2593]: I0307 00:54:19.314407 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238"} err="failed to get container status \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\": rpc error: code = NotFound desc = an error occurred when try to find container \"ea2fdc37519629cfc1fd6943a552851621bf87d22bc2cad8c4e1f2c835a6b238\": not found" Mar 7 00:54:19.314716 kubelet[2593]: I0307 00:54:19.314477 2593 scope.go:117] "RemoveContainer" containerID="2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e" Mar 7 00:54:19.314789 containerd[1484]: time="2026-03-07T00:54:19.314657413Z" level=error msg="ContainerStatus for \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\": not found" Mar 7 00:54:19.314819 kubelet[2593]: E0307 00:54:19.314770 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\": not found" containerID="2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e" Mar 7 00:54:19.314819 kubelet[2593]: I0307 00:54:19.314797 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e"} err="failed to get container status \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\": rpc error: code = NotFound desc = an error occurred when try to find container \"2589ef421a46d144a6f0c616fc1ab5c25d8c4bdd79adb96d6491a0f7eea6e46e\": not found" Mar 7 00:54:19.314819 kubelet[2593]: I0307 00:54:19.314812 2593 scope.go:117] "RemoveContainer" containerID="98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76" Mar 7 00:54:19.316150 containerd[1484]: time="2026-03-07T00:54:19.316038699Z" level=info msg="RemoveContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\"" Mar 7 00:54:19.319380 containerd[1484]: time="2026-03-07T00:54:19.319334062Z" level=info msg="RemoveContainer for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" returns successfully" Mar 7 00:54:19.319662 kubelet[2593]: I0307 00:54:19.319639 2593 scope.go:117] "RemoveContainer" containerID="98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76" Mar 7 00:54:19.319975 containerd[1484]: time="2026-03-07T00:54:19.319937939Z" level=error msg="ContainerStatus for \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\": not found" Mar 7 00:54:19.320250 kubelet[2593]: E0307 00:54:19.320226 2593 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\": not found" containerID="98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76" Mar 7 00:54:19.320323 kubelet[2593]: I0307 00:54:19.320252 2593 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76"} err="failed to get container status \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\": rpc error: code = NotFound desc = an error occurred when try to find container \"98ba02a2ec7e3397dfdf7d79f5c4e11bf219cfd72ad981358f31ed22319b9f76\": not found" Mar 7 00:54:19.730592 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e04264c348f10eda03383d941d02a96ece9546819aab50f3cd08dcfad36b8b69-rootfs.mount: Deactivated successfully. Mar 7 00:54:19.730694 systemd[1]: var-lib-kubelet-pods-93c1ce27\x2d65f7\x2d44f0\x2d9c8f\x2d520ddcd0e3d6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dsj7gr.mount: Deactivated successfully. Mar 7 00:54:19.730757 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8c17f8c134fdc677c7bc83b596136fa46ea8ee2466a4090c0107ad1c85c5f467-rootfs.mount: Deactivated successfully. Mar 7 00:54:19.730812 systemd[1]: var-lib-kubelet-pods-2b37e447\x2dbc99\x2d4e82\x2d8d7e\x2d5c64e610107f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dswzq4.mount: Deactivated successfully. Mar 7 00:54:19.730866 systemd[1]: var-lib-kubelet-pods-2b37e447\x2dbc99\x2d4e82\x2d8d7e\x2d5c64e610107f-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 7 00:54:19.730916 systemd[1]: var-lib-kubelet-pods-2b37e447\x2dbc99\x2d4e82\x2d8d7e\x2d5c64e610107f-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 7 00:54:20.665118 kubelet[2593]: I0307 00:54:20.664110 2593 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b37e447-bc99-4e82-8d7e-5c64e610107f" path="/var/lib/kubelet/pods/2b37e447-bc99-4e82-8d7e-5c64e610107f/volumes" Mar 7 00:54:20.665118 kubelet[2593]: I0307 00:54:20.664859 2593 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6" path="/var/lib/kubelet/pods/93c1ce27-65f7-44f0-9c8f-520ddcd0e3d6/volumes" Mar 7 00:54:20.754562 sshd[4164]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:20.759250 systemd[1]: sshd@19-188.245.50.81:22-20.161.92.111:44860.service: Deactivated successfully. Mar 7 00:54:20.763121 systemd[1]: session-20.scope: Deactivated successfully. Mar 7 00:54:20.763386 systemd[1]: session-20.scope: Consumed 1.305s CPU time. Mar 7 00:54:20.765395 systemd-logind[1455]: Session 20 logged out. Waiting for processes to exit. Mar 7 00:54:20.767303 systemd-logind[1455]: Removed session 20. Mar 7 00:54:20.869981 systemd[1]: Started sshd@20-188.245.50.81:22-20.161.92.111:54906.service - OpenSSH per-connection server daemon (20.161.92.111:54906). Mar 7 00:54:21.455313 sshd[4323]: Accepted publickey for core from 20.161.92.111 port 54906 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:21.457614 sshd[4323]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:21.463455 systemd-logind[1455]: New session 21 of user core. Mar 7 00:54:21.469869 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 7 00:54:21.808077 kubelet[2593]: E0307 00:54:21.807941 2593 kubelet.go:3012] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 7 00:54:23.329873 systemd[1]: Created slice kubepods-burstable-pod6240e3fe_f131_4159_88b5_7680d47abfd6.slice - libcontainer container kubepods-burstable-pod6240e3fe_f131_4159_88b5_7680d47abfd6.slice. Mar 7 00:54:23.365116 sshd[4323]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:23.372742 systemd-logind[1455]: Session 21 logged out. Waiting for processes to exit. Mar 7 00:54:23.373031 systemd[1]: sshd@20-188.245.50.81:22-20.161.92.111:54906.service: Deactivated successfully. Mar 7 00:54:23.375383 systemd[1]: session-21.scope: Deactivated successfully. Mar 7 00:54:23.376708 systemd[1]: session-21.scope: Consumed 1.424s CPU time. Mar 7 00:54:23.378385 systemd-logind[1455]: Removed session 21. Mar 7 00:54:23.490238 systemd[1]: Started sshd@21-188.245.50.81:22-20.161.92.111:54920.service - OpenSSH per-connection server daemon (20.161.92.111:54920). Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491147 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-bpf-maps\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491195 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-cni-path\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491215 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-xtables-lock\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491239 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6240e3fe-f131-4159-88b5-7680d47abfd6-cilium-config-path\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491263 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-lib-modules\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492145 kubelet[2593]: I0307 00:54:23.491285 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-cilium-run\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492786 kubelet[2593]: I0307 00:54:23.491330 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6240e3fe-f131-4159-88b5-7680d47abfd6-clustermesh-secrets\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492786 kubelet[2593]: I0307 00:54:23.491356 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-host-proc-sys-kernel\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492786 kubelet[2593]: I0307 00:54:23.491379 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-host-proc-sys-net\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492786 kubelet[2593]: I0307 00:54:23.491398 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-hostproc\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492786 kubelet[2593]: I0307 00:54:23.491416 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6240e3fe-f131-4159-88b5-7680d47abfd6-hubble-tls\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492952 kubelet[2593]: I0307 00:54:23.491437 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlwhg\" (UniqueName: \"kubernetes.io/projected/6240e3fe-f131-4159-88b5-7680d47abfd6-kube-api-access-rlwhg\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492952 kubelet[2593]: I0307 00:54:23.491465 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-cilium-cgroup\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492952 kubelet[2593]: I0307 00:54:23.491487 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6240e3fe-f131-4159-88b5-7680d47abfd6-etc-cni-netd\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.492952 kubelet[2593]: I0307 00:54:23.491510 2593 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6240e3fe-f131-4159-88b5-7680d47abfd6-cilium-ipsec-secrets\") pod \"cilium-kj5sn\" (UID: \"6240e3fe-f131-4159-88b5-7680d47abfd6\") " pod="kube-system/cilium-kj5sn" Mar 7 00:54:23.639041 containerd[1484]: time="2026-03-07T00:54:23.637949383Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kj5sn,Uid:6240e3fe-f131-4159-88b5-7680d47abfd6,Namespace:kube-system,Attempt:0,}" Mar 7 00:54:23.664753 containerd[1484]: time="2026-03-07T00:54:23.664412788Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 7 00:54:23.664753 containerd[1484]: time="2026-03-07T00:54:23.664478112Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 7 00:54:23.664753 containerd[1484]: time="2026-03-07T00:54:23.664495233Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:54:23.664753 containerd[1484]: time="2026-03-07T00:54:23.664669964Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 7 00:54:23.684910 systemd[1]: Started cri-containerd-c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b.scope - libcontainer container c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b. Mar 7 00:54:23.711104 containerd[1484]: time="2026-03-07T00:54:23.710685985Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kj5sn,Uid:6240e3fe-f131-4159-88b5-7680d47abfd6,Namespace:kube-system,Attempt:0,} returns sandbox id \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\"" Mar 7 00:54:23.720146 containerd[1484]: time="2026-03-07T00:54:23.719991323Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 7 00:54:23.733382 containerd[1484]: time="2026-03-07T00:54:23.733209625Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64\"" Mar 7 00:54:23.734243 containerd[1484]: time="2026-03-07T00:54:23.734028196Z" level=info msg="StartContainer for \"eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64\"" Mar 7 00:54:23.764807 systemd[1]: Started cri-containerd-eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64.scope - libcontainer container eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64. Mar 7 00:54:23.794152 containerd[1484]: time="2026-03-07T00:54:23.793536695Z" level=info msg="StartContainer for \"eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64\" returns successfully" Mar 7 00:54:23.807641 systemd[1]: cri-containerd-eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64.scope: Deactivated successfully. Mar 7 00:54:23.838242 containerd[1484]: time="2026-03-07T00:54:23.837910213Z" level=info msg="shim disconnected" id=eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64 namespace=k8s.io Mar 7 00:54:23.838242 containerd[1484]: time="2026-03-07T00:54:23.837997058Z" level=warning msg="cleaning up after shim disconnected" id=eb4447274e91adc00bcffae6e7426b94c010bc3d1c5203a99bd034fc57754c64 namespace=k8s.io Mar 7 00:54:23.838242 containerd[1484]: time="2026-03-07T00:54:23.838016620Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:24.075623 sshd[4336]: Accepted publickey for core from 20.161.92.111 port 54920 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:24.077664 sshd[4336]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:24.083364 systemd-logind[1455]: New session 22 of user core. Mar 7 00:54:24.087788 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 7 00:54:24.282693 containerd[1484]: time="2026-03-07T00:54:24.282520637Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 7 00:54:24.295808 containerd[1484]: time="2026-03-07T00:54:24.295756941Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1\"" Mar 7 00:54:24.296266 containerd[1484]: time="2026-03-07T00:54:24.296240491Z" level=info msg="StartContainer for \"11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1\"" Mar 7 00:54:24.324775 systemd[1]: Started cri-containerd-11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1.scope - libcontainer container 11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1. Mar 7 00:54:24.353819 containerd[1484]: time="2026-03-07T00:54:24.353675387Z" level=info msg="StartContainer for \"11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1\" returns successfully" Mar 7 00:54:24.362614 systemd[1]: cri-containerd-11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1.scope: Deactivated successfully. Mar 7 00:54:24.391493 containerd[1484]: time="2026-03-07T00:54:24.391209284Z" level=info msg="shim disconnected" id=11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1 namespace=k8s.io Mar 7 00:54:24.391493 containerd[1484]: time="2026-03-07T00:54:24.391273688Z" level=warning msg="cleaning up after shim disconnected" id=11148125a9c488fbdca42c208d560065d3e01223717617e1815837520189e0a1 namespace=k8s.io Mar 7 00:54:24.391493 containerd[1484]: time="2026-03-07T00:54:24.391285728Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:24.490644 sshd[4336]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:24.495652 systemd[1]: sshd@21-188.245.50.81:22-20.161.92.111:54920.service: Deactivated successfully. Mar 7 00:54:24.498817 systemd[1]: session-22.scope: Deactivated successfully. Mar 7 00:54:24.501602 systemd-logind[1455]: Session 22 logged out. Waiting for processes to exit. Mar 7 00:54:24.503028 systemd-logind[1455]: Removed session 22. Mar 7 00:54:24.614015 systemd[1]: Started sshd@22-188.245.50.81:22-20.161.92.111:54924.service - OpenSSH per-connection server daemon (20.161.92.111:54924). Mar 7 00:54:25.198609 sshd[4510]: Accepted publickey for core from 20.161.92.111 port 54924 ssh2: RSA SHA256:fFFMlaCBm9OkQatq7Cg+moKRVH6SG+EKtX7SFDagfEI Mar 7 00:54:25.203762 sshd[4510]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 7 00:54:25.210034 systemd-logind[1455]: New session 23 of user core. Mar 7 00:54:25.214815 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 7 00:54:25.286748 containerd[1484]: time="2026-03-07T00:54:25.286673194Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 7 00:54:25.309348 containerd[1484]: time="2026-03-07T00:54:25.309220269Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef\"" Mar 7 00:54:25.312605 containerd[1484]: time="2026-03-07T00:54:25.312002241Z" level=info msg="StartContainer for \"31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef\"" Mar 7 00:54:25.343784 systemd[1]: Started cri-containerd-31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef.scope - libcontainer container 31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef. Mar 7 00:54:25.381015 containerd[1484]: time="2026-03-07T00:54:25.380863483Z" level=info msg="StartContainer for \"31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef\" returns successfully" Mar 7 00:54:25.383118 systemd[1]: cri-containerd-31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef.scope: Deactivated successfully. Mar 7 00:54:25.410018 containerd[1484]: time="2026-03-07T00:54:25.409913748Z" level=info msg="shim disconnected" id=31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef namespace=k8s.io Mar 7 00:54:25.410018 containerd[1484]: time="2026-03-07T00:54:25.409969471Z" level=warning msg="cleaning up after shim disconnected" id=31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef namespace=k8s.io Mar 7 00:54:25.410018 containerd[1484]: time="2026-03-07T00:54:25.409978271Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:25.603945 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-31da274772b0ff09b3c54d75fd030fd764c4e16f608e94131161922f7bb2cfef-rootfs.mount: Deactivated successfully. Mar 7 00:54:26.300597 containerd[1484]: time="2026-03-07T00:54:26.298500325Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 7 00:54:26.321446 containerd[1484]: time="2026-03-07T00:54:26.319213337Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8\"" Mar 7 00:54:26.323049 containerd[1484]: time="2026-03-07T00:54:26.321759327Z" level=info msg="StartContainer for \"b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8\"" Mar 7 00:54:26.352833 systemd[1]: Started cri-containerd-b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8.scope - libcontainer container b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8. Mar 7 00:54:26.377469 systemd[1]: cri-containerd-b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8.scope: Deactivated successfully. Mar 7 00:54:26.381165 containerd[1484]: time="2026-03-07T00:54:26.381128589Z" level=info msg="StartContainer for \"b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8\" returns successfully" Mar 7 00:54:26.403614 containerd[1484]: time="2026-03-07T00:54:26.403543967Z" level=info msg="shim disconnected" id=b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8 namespace=k8s.io Mar 7 00:54:26.403614 containerd[1484]: time="2026-03-07T00:54:26.403608323Z" level=warning msg="cleaning up after shim disconnected" id=b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8 namespace=k8s.io Mar 7 00:54:26.403614 containerd[1484]: time="2026-03-07T00:54:26.403616882Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 7 00:54:26.417552 containerd[1484]: time="2026-03-07T00:54:26.417463355Z" level=warning msg="cleanup warnings time=\"2026-03-07T00:54:26Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 7 00:54:26.604724 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b4e4dc5855085a26fefcdccbe3d3b934d21d004ac5cfa112dd3df903ec0359d8-rootfs.mount: Deactivated successfully. Mar 7 00:54:26.810114 kubelet[2593]: E0307 00:54:26.810030 2593 kubelet.go:3012] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 7 00:54:27.301241 containerd[1484]: time="2026-03-07T00:54:27.301200737Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 7 00:54:27.322697 containerd[1484]: time="2026-03-07T00:54:27.321660124Z" level=info msg="CreateContainer within sandbox \"c8ce073512e3538d65b8777f4a0760a3b5e52e22abd1c37bb2216dc45a6aa39b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5\"" Mar 7 00:54:27.323251 containerd[1484]: time="2026-03-07T00:54:27.322660259Z" level=info msg="StartContainer for \"575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5\"" Mar 7 00:54:27.356793 systemd[1]: Started cri-containerd-575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5.scope - libcontainer container 575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5. Mar 7 00:54:27.397134 containerd[1484]: time="2026-03-07T00:54:27.397084450Z" level=info msg="StartContainer for \"575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5\" returns successfully" Mar 7 00:54:27.771596 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Mar 7 00:54:29.715968 systemd[1]: run-containerd-runc-k8s.io-575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5-runc.89gWQ2.mount: Deactivated successfully. Mar 7 00:54:30.730006 kubelet[2593]: I0307 00:54:30.729887 2593 setters.go:543] "Node became not ready" node="ci-4081-3-6-n-53d1697e27" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-07T00:54:30Z","lastTransitionTime":"2026-03-07T00:54:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Mar 7 00:54:30.738754 systemd-networkd[1380]: lxc_health: Link UP Mar 7 00:54:30.769909 systemd-networkd[1380]: lxc_health: Gained carrier Mar 7 00:54:31.658530 kubelet[2593]: I0307 00:54:31.658465 2593 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-kj5sn" podStartSLOduration=8.658448233 podStartE2EDuration="8.658448233s" podCreationTimestamp="2026-03-07 00:54:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-07 00:54:28.319883416 +0000 UTC m=+191.794427658" watchObservedRunningTime="2026-03-07 00:54:31.658448233 +0000 UTC m=+195.132992435" Mar 7 00:54:32.020713 systemd-networkd[1380]: lxc_health: Gained IPv6LL Mar 7 00:54:36.238651 systemd[1]: run-containerd-runc-k8s.io-575298d2e48bbe013b2f16b4a6ffb75c91bd43522b3ee1746a60d0be4d9cbcf5-runc.z4RxOk.mount: Deactivated successfully. Mar 7 00:54:38.517435 sshd[4510]: pam_unix(sshd:session): session closed for user core Mar 7 00:54:38.523020 systemd[1]: sshd@22-188.245.50.81:22-20.161.92.111:54924.service: Deactivated successfully. Mar 7 00:54:38.525955 systemd[1]: session-23.scope: Deactivated successfully. Mar 7 00:54:38.528518 systemd-logind[1455]: Session 23 logged out. Waiting for processes to exit. Mar 7 00:54:38.529480 systemd-logind[1455]: Removed session 23.