Apr 17 23:25:53.889673 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 17 23:25:53.889697 kernel: Linux version 6.6.127-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Apr 17 22:13:49 -00 2026 Apr 17 23:25:53.889708 kernel: KASLR enabled Apr 17 23:25:53.889714 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Apr 17 23:25:53.889720 kernel: efi: SMBIOS 3.0=0x139ed0000 MEMATTR=0x138595418 ACPI 2.0=0x136760018 RNG=0x13676e918 MEMRESERVE=0x136b43d18 Apr 17 23:25:53.889725 kernel: random: crng init done Apr 17 23:25:53.889733 kernel: ACPI: Early table checksum verification disabled Apr 17 23:25:53.889738 kernel: ACPI: RSDP 0x0000000136760018 000024 (v02 BOCHS ) Apr 17 23:25:53.889745 kernel: ACPI: XSDT 0x000000013676FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Apr 17 23:25:53.889752 kernel: ACPI: FACP 0x000000013676FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889759 kernel: ACPI: DSDT 0x0000000136767518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889765 kernel: ACPI: APIC 0x000000013676FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889771 kernel: ACPI: PPTT 0x000000013676FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889777 kernel: ACPI: GTDT 0x000000013676D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889784 kernel: ACPI: MCFG 0x000000013676FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889793 kernel: ACPI: SPCR 0x000000013676E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889799 kernel: ACPI: DBG2 0x000000013676E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889806 kernel: ACPI: IORT 0x000000013676E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 17 23:25:53.889812 kernel: ACPI: BGRT 0x000000013676E798 000038 (v01 INTEL EDK2 00000002 01000013) Apr 17 23:25:53.889819 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Apr 17 23:25:53.889825 kernel: NUMA: Failed to initialise from firmware Apr 17 23:25:53.889832 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Apr 17 23:25:53.889838 kernel: NUMA: NODE_DATA [mem 0x13966f800-0x139674fff] Apr 17 23:25:53.889844 kernel: Zone ranges: Apr 17 23:25:53.889851 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Apr 17 23:25:53.889858 kernel: DMA32 empty Apr 17 23:25:53.889865 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Apr 17 23:25:53.889871 kernel: Movable zone start for each node Apr 17 23:25:53.889877 kernel: Early memory node ranges Apr 17 23:25:53.889884 kernel: node 0: [mem 0x0000000040000000-0x000000013676ffff] Apr 17 23:25:53.889891 kernel: node 0: [mem 0x0000000136770000-0x0000000136b3ffff] Apr 17 23:25:53.889897 kernel: node 0: [mem 0x0000000136b40000-0x0000000139e1ffff] Apr 17 23:25:53.889903 kernel: node 0: [mem 0x0000000139e20000-0x0000000139eaffff] Apr 17 23:25:53.889910 kernel: node 0: [mem 0x0000000139eb0000-0x0000000139ebffff] Apr 17 23:25:53.889916 kernel: node 0: [mem 0x0000000139ec0000-0x0000000139fdffff] Apr 17 23:25:53.889923 kernel: node 0: [mem 0x0000000139fe0000-0x0000000139ffffff] Apr 17 23:25:53.889929 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Apr 17 23:25:53.889937 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Apr 17 23:25:53.889943 kernel: psci: probing for conduit method from ACPI. Apr 17 23:25:53.889950 kernel: psci: PSCIv1.1 detected in firmware. Apr 17 23:25:53.889959 kernel: psci: Using standard PSCI v0.2 function IDs Apr 17 23:25:53.889966 kernel: psci: Trusted OS migration not required Apr 17 23:25:53.889973 kernel: psci: SMC Calling Convention v1.1 Apr 17 23:25:53.889982 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Apr 17 23:25:53.889989 kernel: percpu: Embedded 30 pages/cpu s85736 r8192 d28952 u122880 Apr 17 23:25:53.889996 kernel: pcpu-alloc: s85736 r8192 d28952 u122880 alloc=30*4096 Apr 17 23:25:53.890003 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 17 23:25:53.890009 kernel: Detected PIPT I-cache on CPU0 Apr 17 23:25:53.890016 kernel: CPU features: detected: GIC system register CPU interface Apr 17 23:25:53.890023 kernel: CPU features: detected: Hardware dirty bit management Apr 17 23:25:53.890030 kernel: CPU features: detected: Spectre-v4 Apr 17 23:25:53.890036 kernel: CPU features: detected: Spectre-BHB Apr 17 23:25:53.890043 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 17 23:25:53.890052 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 17 23:25:53.890058 kernel: CPU features: detected: ARM erratum 1418040 Apr 17 23:25:53.890065 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 17 23:25:53.890072 kernel: alternatives: applying boot alternatives Apr 17 23:25:53.890080 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=f77c53ef012912081447488e689e924a7faa1d92b63ab5dfeba9709e9511e349 Apr 17 23:25:53.890087 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 17 23:25:53.890094 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 17 23:25:53.890101 kernel: Fallback order for Node 0: 0 Apr 17 23:25:53.890107 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Apr 17 23:25:53.890114 kernel: Policy zone: Normal Apr 17 23:25:53.890121 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 17 23:25:53.890129 kernel: software IO TLB: area num 2. Apr 17 23:25:53.890136 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Apr 17 23:25:53.890143 kernel: Memory: 3882816K/4096000K available (10304K kernel code, 2180K rwdata, 8116K rodata, 39424K init, 897K bss, 213184K reserved, 0K cma-reserved) Apr 17 23:25:53.890150 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 17 23:25:53.890157 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 17 23:25:53.890164 kernel: rcu: RCU event tracing is enabled. Apr 17 23:25:53.890171 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 17 23:25:53.892447 kernel: Trampoline variant of Tasks RCU enabled. Apr 17 23:25:53.892456 kernel: Tracing variant of Tasks RCU enabled. Apr 17 23:25:53.892463 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 17 23:25:53.892470 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 17 23:25:53.892477 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 17 23:25:53.892490 kernel: GICv3: 256 SPIs implemented Apr 17 23:25:53.892497 kernel: GICv3: 0 Extended SPIs implemented Apr 17 23:25:53.892504 kernel: Root IRQ handler: gic_handle_irq Apr 17 23:25:53.892511 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Apr 17 23:25:53.892518 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Apr 17 23:25:53.892524 kernel: ITS [mem 0x08080000-0x0809ffff] Apr 17 23:25:53.892532 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Apr 17 23:25:53.892539 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Apr 17 23:25:53.892546 kernel: GICv3: using LPI property table @0x00000001000e0000 Apr 17 23:25:53.892553 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Apr 17 23:25:53.892560 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 17 23:25:53.892568 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 17 23:25:53.892575 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 17 23:25:53.892582 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 17 23:25:53.892590 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 17 23:25:53.892597 kernel: Console: colour dummy device 80x25 Apr 17 23:25:53.892604 kernel: ACPI: Core revision 20230628 Apr 17 23:25:53.892611 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 17 23:25:53.892619 kernel: pid_max: default: 32768 minimum: 301 Apr 17 23:25:53.892626 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 17 23:25:53.892633 kernel: landlock: Up and running. Apr 17 23:25:53.892641 kernel: SELinux: Initializing. Apr 17 23:25:53.892649 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 17 23:25:53.892656 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 17 23:25:53.892664 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 17 23:25:53.892671 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 17 23:25:53.892678 kernel: rcu: Hierarchical SRCU implementation. Apr 17 23:25:53.892685 kernel: rcu: Max phase no-delay instances is 400. Apr 17 23:25:53.892692 kernel: Platform MSI: ITS@0x8080000 domain created Apr 17 23:25:53.892699 kernel: PCI/MSI: ITS@0x8080000 domain created Apr 17 23:25:53.892708 kernel: Remapping and enabling EFI services. Apr 17 23:25:53.892715 kernel: smp: Bringing up secondary CPUs ... Apr 17 23:25:53.892722 kernel: Detected PIPT I-cache on CPU1 Apr 17 23:25:53.892730 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Apr 17 23:25:53.892737 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Apr 17 23:25:53.892744 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 17 23:25:53.892751 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 17 23:25:53.892758 kernel: smp: Brought up 1 node, 2 CPUs Apr 17 23:25:53.892765 kernel: SMP: Total of 2 processors activated. Apr 17 23:25:53.892774 kernel: CPU features: detected: 32-bit EL0 Support Apr 17 23:25:53.892781 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 17 23:25:53.892788 kernel: CPU features: detected: Common not Private translations Apr 17 23:25:53.892800 kernel: CPU features: detected: CRC32 instructions Apr 17 23:25:53.892809 kernel: CPU features: detected: Enhanced Virtualization Traps Apr 17 23:25:53.892817 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 17 23:25:53.892824 kernel: CPU features: detected: LSE atomic instructions Apr 17 23:25:53.892832 kernel: CPU features: detected: Privileged Access Never Apr 17 23:25:53.892839 kernel: CPU features: detected: RAS Extension Support Apr 17 23:25:53.892848 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Apr 17 23:25:53.892856 kernel: CPU: All CPU(s) started at EL1 Apr 17 23:25:53.892863 kernel: alternatives: applying system-wide alternatives Apr 17 23:25:53.892871 kernel: devtmpfs: initialized Apr 17 23:25:53.892879 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 17 23:25:53.892886 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 17 23:25:53.892894 kernel: pinctrl core: initialized pinctrl subsystem Apr 17 23:25:53.892902 kernel: SMBIOS 3.0.0 present. Apr 17 23:25:53.892911 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Apr 17 23:25:53.892919 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 17 23:25:53.892926 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 17 23:25:53.892934 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 17 23:25:53.892941 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 17 23:25:53.892949 kernel: audit: initializing netlink subsys (disabled) Apr 17 23:25:53.892956 kernel: audit: type=2000 audit(0.012:1): state=initialized audit_enabled=0 res=1 Apr 17 23:25:53.892964 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 17 23:25:53.892971 kernel: cpuidle: using governor menu Apr 17 23:25:53.892980 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 17 23:25:53.892988 kernel: ASID allocator initialised with 32768 entries Apr 17 23:25:53.892995 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 17 23:25:53.893003 kernel: Serial: AMBA PL011 UART driver Apr 17 23:25:53.893010 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 17 23:25:53.893018 kernel: Modules: 0 pages in range for non-PLT usage Apr 17 23:25:53.893025 kernel: Modules: 509008 pages in range for PLT usage Apr 17 23:25:53.893033 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 17 23:25:53.893040 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 17 23:25:53.893050 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 17 23:25:53.893057 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 17 23:25:53.893064 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 17 23:25:53.893072 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 17 23:25:53.893079 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 17 23:25:53.893087 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 17 23:25:53.893094 kernel: ACPI: Added _OSI(Module Device) Apr 17 23:25:53.893101 kernel: ACPI: Added _OSI(Processor Device) Apr 17 23:25:53.893109 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 17 23:25:53.893117 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 17 23:25:53.893125 kernel: ACPI: Interpreter enabled Apr 17 23:25:53.893132 kernel: ACPI: Using GIC for interrupt routing Apr 17 23:25:53.893140 kernel: ACPI: MCFG table detected, 1 entries Apr 17 23:25:53.893147 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Apr 17 23:25:53.893155 kernel: printk: console [ttyAMA0] enabled Apr 17 23:25:53.893162 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 17 23:25:53.893557 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 17 23:25:53.893647 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Apr 17 23:25:53.893715 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Apr 17 23:25:53.893780 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Apr 17 23:25:53.893846 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Apr 17 23:25:53.893856 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Apr 17 23:25:53.893864 kernel: PCI host bridge to bus 0000:00 Apr 17 23:25:53.893942 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Apr 17 23:25:53.894008 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Apr 17 23:25:53.894068 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Apr 17 23:25:53.894127 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 17 23:25:53.894833 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Apr 17 23:25:53.894925 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Apr 17 23:25:53.894997 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Apr 17 23:25:53.895066 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Apr 17 23:25:53.895146 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.895241 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Apr 17 23:25:53.895319 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.895402 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Apr 17 23:25:53.895483 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.895556 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Apr 17 23:25:53.895638 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.895725 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Apr 17 23:25:53.895812 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.895896 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Apr 17 23:25:53.895978 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.896206 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Apr 17 23:25:53.896312 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.896424 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Apr 17 23:25:53.896515 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.896586 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Apr 17 23:25:53.896660 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Apr 17 23:25:53.896727 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Apr 17 23:25:53.896805 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Apr 17 23:25:53.896872 kernel: pci 0000:00:04.0: reg 0x10: [io 0x0000-0x0007] Apr 17 23:25:53.896957 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Apr 17 23:25:53.897033 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Apr 17 23:25:53.897104 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Apr 17 23:25:53.897192 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 17 23:25:53.897279 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Apr 17 23:25:53.897356 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Apr 17 23:25:53.897455 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Apr 17 23:25:53.897530 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Apr 17 23:25:53.897601 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Apr 17 23:25:53.897683 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Apr 17 23:25:53.897754 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Apr 17 23:25:53.897834 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Apr 17 23:25:53.897905 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x10800000-0x10800fff] Apr 17 23:25:53.897975 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Apr 17 23:25:53.898053 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Apr 17 23:25:53.898125 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Apr 17 23:25:53.902294 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Apr 17 23:25:53.902431 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Apr 17 23:25:53.902507 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Apr 17 23:25:53.902579 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Apr 17 23:25:53.902647 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Apr 17 23:25:53.902719 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 17 23:25:53.902790 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Apr 17 23:25:53.902857 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Apr 17 23:25:53.902933 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 17 23:25:53.903000 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 17 23:25:53.903066 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Apr 17 23:25:53.903137 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 17 23:25:53.903243 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Apr 17 23:25:53.903314 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Apr 17 23:25:53.903419 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 17 23:25:53.903498 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Apr 17 23:25:53.903570 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 17 23:25:53.903641 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Apr 17 23:25:53.903709 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Apr 17 23:25:53.903775 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Apr 17 23:25:53.903846 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Apr 17 23:25:53.903914 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Apr 17 23:25:53.903982 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Apr 17 23:25:53.904056 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Apr 17 23:25:53.904124 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Apr 17 23:25:53.904375 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Apr 17 23:25:53.904541 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Apr 17 23:25:53.904616 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Apr 17 23:25:53.904687 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Apr 17 23:25:53.904760 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Apr 17 23:25:53.904829 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Apr 17 23:25:53.904902 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Apr 17 23:25:53.904972 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Apr 17 23:25:53.905040 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Apr 17 23:25:53.905109 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Apr 17 23:25:53.905198 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Apr 17 23:25:53.905273 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Apr 17 23:25:53.905347 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Apr 17 23:25:53.905430 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Apr 17 23:25:53.905502 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Apr 17 23:25:53.905571 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Apr 17 23:25:53.905639 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Apr 17 23:25:53.905709 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Apr 17 23:25:53.905776 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 17 23:25:53.905850 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Apr 17 23:25:53.905930 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 17 23:25:53.906000 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Apr 17 23:25:53.906070 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 17 23:25:53.906139 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Apr 17 23:25:53.906229 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Apr 17 23:25:53.906306 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Apr 17 23:25:53.906378 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Apr 17 23:25:53.906500 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Apr 17 23:25:53.906570 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Apr 17 23:25:53.906640 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Apr 17 23:25:53.906708 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Apr 17 23:25:53.906775 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Apr 17 23:25:53.906843 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Apr 17 23:25:53.906910 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Apr 17 23:25:53.906983 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Apr 17 23:25:53.907054 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Apr 17 23:25:53.907122 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Apr 17 23:25:53.907219 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Apr 17 23:25:53.907293 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Apr 17 23:25:53.907363 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Apr 17 23:25:53.907449 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Apr 17 23:25:53.907520 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Apr 17 23:25:53.907593 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Apr 17 23:25:53.907661 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Apr 17 23:25:53.907738 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Apr 17 23:25:53.907811 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Apr 17 23:25:53.907889 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Apr 17 23:25:53.907962 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Apr 17 23:25:53.908032 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Apr 17 23:25:53.908100 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 17 23:25:53.908171 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Apr 17 23:25:53.908303 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Apr 17 23:25:53.908371 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Apr 17 23:25:53.908487 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Apr 17 23:25:53.908565 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 17 23:25:53.908632 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Apr 17 23:25:53.908699 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Apr 17 23:25:53.908765 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Apr 17 23:25:53.908840 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Apr 17 23:25:53.908911 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Apr 17 23:25:53.908978 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 17 23:25:53.909044 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Apr 17 23:25:53.909113 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Apr 17 23:25:53.909321 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Apr 17 23:25:53.909428 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Apr 17 23:25:53.909500 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 17 23:25:53.909566 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Apr 17 23:25:53.909631 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Apr 17 23:25:53.909697 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Apr 17 23:25:53.909770 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Apr 17 23:25:53.909847 kernel: pci 0000:05:00.0: BAR 1: assigned [mem 0x10800000-0x10800fff] Apr 17 23:25:53.909914 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 17 23:25:53.909981 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Apr 17 23:25:53.910046 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Apr 17 23:25:53.910111 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Apr 17 23:25:53.910197 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Apr 17 23:25:53.910269 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Apr 17 23:25:53.910336 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 17 23:25:53.910418 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Apr 17 23:25:53.910485 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Apr 17 23:25:53.910551 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 17 23:25:53.910624 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Apr 17 23:25:53.910693 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Apr 17 23:25:53.910762 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Apr 17 23:25:53.910829 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 17 23:25:53.910895 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Apr 17 23:25:53.910963 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Apr 17 23:25:53.911030 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 17 23:25:53.911098 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 17 23:25:53.911164 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Apr 17 23:25:53.911285 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Apr 17 23:25:53.911353 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 17 23:25:53.911465 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 17 23:25:53.911534 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Apr 17 23:25:53.911607 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Apr 17 23:25:53.911672 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Apr 17 23:25:53.911740 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Apr 17 23:25:53.911800 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Apr 17 23:25:53.911858 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Apr 17 23:25:53.911929 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Apr 17 23:25:53.911991 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Apr 17 23:25:53.912054 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Apr 17 23:25:53.912122 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Apr 17 23:25:53.912202 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Apr 17 23:25:53.912267 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Apr 17 23:25:53.912340 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Apr 17 23:25:53.912421 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Apr 17 23:25:53.912491 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Apr 17 23:25:53.912562 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Apr 17 23:25:53.912624 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Apr 17 23:25:53.912701 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Apr 17 23:25:53.912775 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Apr 17 23:25:53.912839 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Apr 17 23:25:53.912910 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Apr 17 23:25:53.912986 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Apr 17 23:25:53.913052 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Apr 17 23:25:53.913118 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Apr 17 23:25:53.913203 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Apr 17 23:25:53.913271 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Apr 17 23:25:53.913333 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Apr 17 23:25:53.913448 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Apr 17 23:25:53.913520 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Apr 17 23:25:53.913583 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Apr 17 23:25:53.913654 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Apr 17 23:25:53.913724 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Apr 17 23:25:53.913835 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Apr 17 23:25:53.913850 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Apr 17 23:25:53.913858 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Apr 17 23:25:53.913866 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Apr 17 23:25:53.913875 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Apr 17 23:25:53.913883 kernel: iommu: Default domain type: Translated Apr 17 23:25:53.913891 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 17 23:25:53.913899 kernel: efivars: Registered efivars operations Apr 17 23:25:53.913907 kernel: vgaarb: loaded Apr 17 23:25:53.913918 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 17 23:25:53.913926 kernel: VFS: Disk quotas dquot_6.6.0 Apr 17 23:25:53.913934 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 17 23:25:53.913942 kernel: pnp: PnP ACPI init Apr 17 23:25:53.914032 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Apr 17 23:25:53.914045 kernel: pnp: PnP ACPI: found 1 devices Apr 17 23:25:53.914053 kernel: NET: Registered PF_INET protocol family Apr 17 23:25:53.914061 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 17 23:25:53.914072 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 17 23:25:53.914080 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 17 23:25:53.914088 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 17 23:25:53.914096 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 17 23:25:53.914104 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 17 23:25:53.914112 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 17 23:25:53.914120 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 17 23:25:53.914128 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 17 23:25:53.914221 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Apr 17 23:25:53.914236 kernel: PCI: CLS 0 bytes, default 64 Apr 17 23:25:53.914245 kernel: kvm [1]: HYP mode not available Apr 17 23:25:53.914253 kernel: Initialise system trusted keyrings Apr 17 23:25:53.914261 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 17 23:25:53.914269 kernel: Key type asymmetric registered Apr 17 23:25:53.914277 kernel: Asymmetric key parser 'x509' registered Apr 17 23:25:53.914285 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 17 23:25:53.914292 kernel: io scheduler mq-deadline registered Apr 17 23:25:53.914300 kernel: io scheduler kyber registered Apr 17 23:25:53.914310 kernel: io scheduler bfq registered Apr 17 23:25:53.914319 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Apr 17 23:25:53.914406 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Apr 17 23:25:53.914486 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Apr 17 23:25:53.914556 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.914627 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Apr 17 23:25:53.914696 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Apr 17 23:25:53.914769 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.914839 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Apr 17 23:25:53.914908 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Apr 17 23:25:53.914975 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.915047 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Apr 17 23:25:53.915126 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Apr 17 23:25:53.915240 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.915315 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Apr 17 23:25:53.915394 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Apr 17 23:25:53.915466 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.915535 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Apr 17 23:25:53.915602 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Apr 17 23:25:53.915677 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.915748 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Apr 17 23:25:53.915817 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Apr 17 23:25:53.915885 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.915955 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Apr 17 23:25:53.916024 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Apr 17 23:25:53.916098 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.916109 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Apr 17 23:25:53.918263 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Apr 17 23:25:53.918443 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Apr 17 23:25:53.918522 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Apr 17 23:25:53.918535 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Apr 17 23:25:53.918550 kernel: ACPI: button: Power Button [PWRB] Apr 17 23:25:53.918559 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Apr 17 23:25:53.918635 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Apr 17 23:25:53.918711 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Apr 17 23:25:53.918723 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 17 23:25:53.918731 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Apr 17 23:25:53.918800 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Apr 17 23:25:53.918812 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Apr 17 23:25:53.918820 kernel: thunder_xcv, ver 1.0 Apr 17 23:25:53.918830 kernel: thunder_bgx, ver 1.0 Apr 17 23:25:53.918839 kernel: nicpf, ver 1.0 Apr 17 23:25:53.918846 kernel: nicvf, ver 1.0 Apr 17 23:25:53.918934 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 17 23:25:53.919001 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-04-17T23:25:53 UTC (1776468353) Apr 17 23:25:53.919011 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 17 23:25:53.919020 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Apr 17 23:25:53.919028 kernel: watchdog: Delayed init of the lockup detector failed: -19 Apr 17 23:25:53.919038 kernel: watchdog: Hard watchdog permanently disabled Apr 17 23:25:53.919046 kernel: NET: Registered PF_INET6 protocol family Apr 17 23:25:53.919054 kernel: Segment Routing with IPv6 Apr 17 23:25:53.919062 kernel: In-situ OAM (IOAM) with IPv6 Apr 17 23:25:53.919070 kernel: NET: Registered PF_PACKET protocol family Apr 17 23:25:53.919078 kernel: Key type dns_resolver registered Apr 17 23:25:53.919086 kernel: registered taskstats version 1 Apr 17 23:25:53.919094 kernel: Loading compiled-in X.509 certificates Apr 17 23:25:53.919102 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.127-flatcar: 1161289bfc8d953baa9f687fefeecf0e077bc535' Apr 17 23:25:53.919114 kernel: Key type .fscrypt registered Apr 17 23:25:53.919123 kernel: Key type fscrypt-provisioning registered Apr 17 23:25:53.919131 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 17 23:25:53.919139 kernel: ima: Allocated hash algorithm: sha1 Apr 17 23:25:53.919147 kernel: ima: No architecture policies found Apr 17 23:25:53.919155 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 17 23:25:53.919163 kernel: clk: Disabling unused clocks Apr 17 23:25:53.919171 kernel: Freeing unused kernel memory: 39424K Apr 17 23:25:53.919195 kernel: Run /init as init process Apr 17 23:25:53.919205 kernel: with arguments: Apr 17 23:25:53.919213 kernel: /init Apr 17 23:25:53.919221 kernel: with environment: Apr 17 23:25:53.919229 kernel: HOME=/ Apr 17 23:25:53.919237 kernel: TERM=linux Apr 17 23:25:53.919247 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 17 23:25:53.919257 systemd[1]: Detected virtualization kvm. Apr 17 23:25:53.919266 systemd[1]: Detected architecture arm64. Apr 17 23:25:53.919275 systemd[1]: Running in initrd. Apr 17 23:25:53.919284 systemd[1]: No hostname configured, using default hostname. Apr 17 23:25:53.919292 systemd[1]: Hostname set to . Apr 17 23:25:53.919300 systemd[1]: Initializing machine ID from VM UUID. Apr 17 23:25:53.919308 systemd[1]: Queued start job for default target initrd.target. Apr 17 23:25:53.919317 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 17 23:25:53.919326 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 17 23:25:53.919335 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 17 23:25:53.919345 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 17 23:25:53.919354 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 17 23:25:53.919363 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 17 23:25:53.919373 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 17 23:25:53.919392 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 17 23:25:53.919401 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 17 23:25:53.919409 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 17 23:25:53.919420 systemd[1]: Reached target paths.target - Path Units. Apr 17 23:25:53.919430 systemd[1]: Reached target slices.target - Slice Units. Apr 17 23:25:53.919439 systemd[1]: Reached target swap.target - Swaps. Apr 17 23:25:53.919447 systemd[1]: Reached target timers.target - Timer Units. Apr 17 23:25:53.919456 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 17 23:25:53.919464 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 17 23:25:53.919473 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 17 23:25:53.919482 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 17 23:25:53.919492 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 17 23:25:53.919500 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 17 23:25:53.919509 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 17 23:25:53.919517 systemd[1]: Reached target sockets.target - Socket Units. Apr 17 23:25:53.919526 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 17 23:25:53.919534 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 17 23:25:53.919543 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 17 23:25:53.919551 systemd[1]: Starting systemd-fsck-usr.service... Apr 17 23:25:53.919560 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 17 23:25:53.919570 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 17 23:25:53.919578 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 17 23:25:53.919587 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 17 23:25:53.919595 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 17 23:25:53.919603 systemd[1]: Finished systemd-fsck-usr.service. Apr 17 23:25:53.919636 systemd-journald[238]: Collecting audit messages is disabled. Apr 17 23:25:53.919659 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 17 23:25:53.919668 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 17 23:25:53.919679 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:25:53.919687 kernel: Bridge firewalling registered Apr 17 23:25:53.919695 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 17 23:25:53.919704 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 17 23:25:53.919713 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 17 23:25:53.919722 systemd-journald[238]: Journal started Apr 17 23:25:53.919743 systemd-journald[238]: Runtime Journal (/run/log/journal/a295e9f57e274b21a3592359c019d2b7) is 8.0M, max 76.6M, 68.6M free. Apr 17 23:25:53.922363 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 17 23:25:53.882041 systemd-modules-load[239]: Inserted module 'overlay' Apr 17 23:25:53.901726 systemd-modules-load[239]: Inserted module 'br_netfilter' Apr 17 23:25:53.929217 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 17 23:25:53.929270 systemd[1]: Started systemd-journald.service - Journal Service. Apr 17 23:25:53.942460 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 17 23:25:53.946479 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 17 23:25:53.950261 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 17 23:25:53.959676 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 17 23:25:53.961113 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 17 23:25:53.969357 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 17 23:25:53.974466 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 17 23:25:53.978621 dracut-cmdline[270]: dracut-dracut-053 Apr 17 23:25:53.981962 dracut-cmdline[270]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=f77c53ef012912081447488e689e924a7faa1d92b63ab5dfeba9709e9511e349 Apr 17 23:25:54.014997 systemd-resolved[281]: Positive Trust Anchors: Apr 17 23:25:54.015019 systemd-resolved[281]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 17 23:25:54.015051 systemd-resolved[281]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 17 23:25:54.021338 systemd-resolved[281]: Defaulting to hostname 'linux'. Apr 17 23:25:54.023376 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 17 23:25:54.025310 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 17 23:25:54.084205 kernel: SCSI subsystem initialized Apr 17 23:25:54.088197 kernel: Loading iSCSI transport class v2.0-870. Apr 17 23:25:54.096206 kernel: iscsi: registered transport (tcp) Apr 17 23:25:54.109280 kernel: iscsi: registered transport (qla4xxx) Apr 17 23:25:54.109364 kernel: QLogic iSCSI HBA Driver Apr 17 23:25:54.158465 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 17 23:25:54.170661 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 17 23:25:54.194796 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 17 23:25:54.194859 kernel: device-mapper: uevent: version 1.0.3 Apr 17 23:25:54.194872 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 17 23:25:54.245234 kernel: raid6: neonx8 gen() 15429 MB/s Apr 17 23:25:54.262265 kernel: raid6: neonx4 gen() 15424 MB/s Apr 17 23:25:54.279227 kernel: raid6: neonx2 gen() 13051 MB/s Apr 17 23:25:54.296240 kernel: raid6: neonx1 gen() 10401 MB/s Apr 17 23:25:54.313233 kernel: raid6: int64x8 gen() 6884 MB/s Apr 17 23:25:54.330233 kernel: raid6: int64x4 gen() 7300 MB/s Apr 17 23:25:54.347257 kernel: raid6: int64x2 gen() 6083 MB/s Apr 17 23:25:54.364238 kernel: raid6: int64x1 gen() 5018 MB/s Apr 17 23:25:54.364287 kernel: raid6: using algorithm neonx8 gen() 15429 MB/s Apr 17 23:25:54.381239 kernel: raid6: .... xor() 11905 MB/s, rmw enabled Apr 17 23:25:54.381317 kernel: raid6: using neon recovery algorithm Apr 17 23:25:54.386492 kernel: xor: measuring software checksum speed Apr 17 23:25:54.386567 kernel: 8regs : 19807 MB/sec Apr 17 23:25:54.386597 kernel: 32regs : 19669 MB/sec Apr 17 23:25:54.386625 kernel: arm64_neon : 22834 MB/sec Apr 17 23:25:54.387216 kernel: xor: using function: arm64_neon (22834 MB/sec) Apr 17 23:25:54.439219 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 17 23:25:54.454860 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 17 23:25:54.469517 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 17 23:25:54.483579 systemd-udevd[458]: Using default interface naming scheme 'v255'. Apr 17 23:25:54.487050 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 17 23:25:54.496569 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 17 23:25:54.516103 dracut-pre-trigger[470]: rd.md=0: removing MD RAID activation Apr 17 23:25:54.552679 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 17 23:25:54.558371 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 17 23:25:54.609996 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 17 23:25:54.620359 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 17 23:25:54.637879 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 17 23:25:54.639945 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 17 23:25:54.640695 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 17 23:25:54.642479 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 17 23:25:54.651718 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 17 23:25:54.663600 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 17 23:25:54.713533 kernel: ACPI: bus type USB registered Apr 17 23:25:54.713595 kernel: usbcore: registered new interface driver usbfs Apr 17 23:25:54.713606 kernel: usbcore: registered new interface driver hub Apr 17 23:25:54.713616 kernel: usbcore: registered new device driver usb Apr 17 23:25:54.733713 kernel: scsi host0: Virtio SCSI HBA Apr 17 23:25:54.735148 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 17 23:25:54.735279 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 17 23:25:54.740827 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 17 23:25:54.740909 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Apr 17 23:25:54.738286 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 17 23:25:54.743481 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 17 23:25:54.743779 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:25:54.747216 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 17 23:25:54.757423 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 17 23:25:54.757784 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Apr 17 23:25:54.757813 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 17 23:25:54.763644 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Apr 17 23:25:54.763831 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 17 23:25:54.763927 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Apr 17 23:25:54.765219 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Apr 17 23:25:54.767497 kernel: hub 1-0:1.0: USB hub found Apr 17 23:25:54.767845 kernel: hub 1-0:1.0: 4 ports detected Apr 17 23:25:54.770019 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Apr 17 23:25:54.772454 kernel: hub 2-0:1.0: USB hub found Apr 17 23:25:54.772644 kernel: hub 2-0:1.0: 4 ports detected Apr 17 23:25:54.783437 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:25:54.793411 kernel: sr 0:0:0:0: Power-on or device reset occurred Apr 17 23:25:54.793613 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Apr 17 23:25:54.794352 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 17 23:25:54.795210 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Apr 17 23:25:54.795466 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 17 23:25:54.805441 kernel: sd 0:0:0:1: Power-on or device reset occurred Apr 17 23:25:54.809068 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Apr 17 23:25:54.809300 kernel: sd 0:0:0:1: [sda] Write Protect is off Apr 17 23:25:54.809410 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Apr 17 23:25:54.809502 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Apr 17 23:25:54.812258 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 17 23:25:54.812313 kernel: GPT:17805311 != 80003071 Apr 17 23:25:54.812325 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 17 23:25:54.813279 kernel: GPT:17805311 != 80003071 Apr 17 23:25:54.813316 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 17 23:25:54.813327 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 17 23:25:54.815099 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Apr 17 23:25:54.827577 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 17 23:25:54.861208 kernel: BTRFS: device fsid 6218981f-ef91-4196-be05-d5f6a224b350 devid 1 transid 32 /dev/sda3 scanned by (udev-worker) (507) Apr 17 23:25:54.870199 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (517) Apr 17 23:25:54.875599 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Apr 17 23:25:54.881090 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Apr 17 23:25:54.891696 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Apr 17 23:25:54.892514 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Apr 17 23:25:54.897437 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 17 23:25:54.903415 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 17 23:25:54.909632 disk-uuid[578]: Primary Header is updated. Apr 17 23:25:54.909632 disk-uuid[578]: Secondary Entries is updated. Apr 17 23:25:54.909632 disk-uuid[578]: Secondary Header is updated. Apr 17 23:25:54.925222 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 17 23:25:55.014231 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Apr 17 23:25:55.149234 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Apr 17 23:25:55.149341 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Apr 17 23:25:55.151219 kernel: usbcore: registered new interface driver usbhid Apr 17 23:25:55.151271 kernel: usbhid: USB HID core driver Apr 17 23:25:55.257231 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Apr 17 23:25:55.386220 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Apr 17 23:25:55.440233 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Apr 17 23:25:55.930209 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 17 23:25:55.931071 disk-uuid[579]: The operation has completed successfully. Apr 17 23:25:55.984608 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 17 23:25:55.984714 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 17 23:25:55.997516 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 17 23:25:56.002497 sh[591]: Success Apr 17 23:25:56.018289 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 17 23:25:56.078578 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 17 23:25:56.081328 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 17 23:25:56.083144 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 17 23:25:56.105210 kernel: BTRFS info (device dm-0): first mount of filesystem 6218981f-ef91-4196-be05-d5f6a224b350 Apr 17 23:25:56.105275 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 17 23:25:56.105297 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 17 23:25:56.106489 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 17 23:25:56.106530 kernel: BTRFS info (device dm-0): using free space tree Apr 17 23:25:56.114241 kernel: BTRFS info (device dm-0): enabling ssd optimizations Apr 17 23:25:56.116649 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 17 23:25:56.117612 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 17 23:25:56.128479 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 17 23:25:56.133520 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 17 23:25:56.143615 kernel: BTRFS info (device sda6): first mount of filesystem 511634b8-962b-4ed3-9161-3f02d13492ea Apr 17 23:25:56.143670 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 17 23:25:56.143682 kernel: BTRFS info (device sda6): using free space tree Apr 17 23:25:56.149223 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 17 23:25:56.149280 kernel: BTRFS info (device sda6): auto enabling async discard Apr 17 23:25:56.164249 kernel: BTRFS info (device sda6): last unmount of filesystem 511634b8-962b-4ed3-9161-3f02d13492ea Apr 17 23:25:56.164890 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 17 23:25:56.172139 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 17 23:25:56.181359 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 17 23:25:56.252943 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 17 23:25:56.262869 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 17 23:25:56.280203 ignition[685]: Ignition 2.19.0 Apr 17 23:25:56.280214 ignition[685]: Stage: fetch-offline Apr 17 23:25:56.280252 ignition[685]: no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:56.280260 ignition[685]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:56.283042 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 17 23:25:56.280428 ignition[685]: parsed url from cmdline: "" Apr 17 23:25:56.280431 ignition[685]: no config URL provided Apr 17 23:25:56.280436 ignition[685]: reading system config file "/usr/lib/ignition/user.ign" Apr 17 23:25:56.280443 ignition[685]: no config at "/usr/lib/ignition/user.ign" Apr 17 23:25:56.280448 ignition[685]: failed to fetch config: resource requires networking Apr 17 23:25:56.280625 ignition[685]: Ignition finished successfully Apr 17 23:25:56.290552 systemd-networkd[777]: lo: Link UP Apr 17 23:25:56.290556 systemd-networkd[777]: lo: Gained carrier Apr 17 23:25:56.292130 systemd-networkd[777]: Enumeration completed Apr 17 23:25:56.292662 systemd-networkd[777]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:25:56.292665 systemd-networkd[777]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 17 23:25:56.293023 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 17 23:25:56.293854 systemd[1]: Reached target network.target - Network. Apr 17 23:25:56.294386 systemd-networkd[777]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:25:56.294389 systemd-networkd[777]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 17 23:25:56.294903 systemd-networkd[777]: eth0: Link UP Apr 17 23:25:56.294907 systemd-networkd[777]: eth0: Gained carrier Apr 17 23:25:56.294914 systemd-networkd[777]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:25:56.302484 systemd-networkd[777]: eth1: Link UP Apr 17 23:25:56.302491 systemd-networkd[777]: eth1: Gained carrier Apr 17 23:25:56.302500 systemd-networkd[777]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:25:56.303303 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 17 23:25:56.316783 ignition[780]: Ignition 2.19.0 Apr 17 23:25:56.316793 ignition[780]: Stage: fetch Apr 17 23:25:56.316988 ignition[780]: no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:56.316998 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:56.317095 ignition[780]: parsed url from cmdline: "" Apr 17 23:25:56.317099 ignition[780]: no config URL provided Apr 17 23:25:56.317104 ignition[780]: reading system config file "/usr/lib/ignition/user.ign" Apr 17 23:25:56.317112 ignition[780]: no config at "/usr/lib/ignition/user.ign" Apr 17 23:25:56.317132 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Apr 17 23:25:56.317836 ignition[780]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Apr 17 23:25:56.343297 systemd-networkd[777]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 17 23:25:56.353336 systemd-networkd[777]: eth0: DHCPv4 address 46.224.37.37/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 17 23:25:56.518354 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Apr 17 23:25:56.524931 ignition[780]: GET result: OK Apr 17 23:25:56.525136 ignition[780]: parsing config with SHA512: 3c63973222529d26d06bc44a3b6db93ba886db0bffc030cae8cb610f20dba6dd1c417820858d99b43a60e9b93428c8031bc10811582ffc7e8e18bc23199be216 Apr 17 23:25:56.532701 unknown[780]: fetched base config from "system" Apr 17 23:25:56.532717 unknown[780]: fetched base config from "system" Apr 17 23:25:56.533213 ignition[780]: fetch: fetch complete Apr 17 23:25:56.532727 unknown[780]: fetched user config from "hetzner" Apr 17 23:25:56.533218 ignition[780]: fetch: fetch passed Apr 17 23:25:56.535101 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 17 23:25:56.533268 ignition[780]: Ignition finished successfully Apr 17 23:25:56.545525 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 17 23:25:56.559780 ignition[787]: Ignition 2.19.0 Apr 17 23:25:56.559789 ignition[787]: Stage: kargs Apr 17 23:25:56.559966 ignition[787]: no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:56.559976 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:56.561033 ignition[787]: kargs: kargs passed Apr 17 23:25:56.561087 ignition[787]: Ignition finished successfully Apr 17 23:25:56.564523 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 17 23:25:56.570418 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 17 23:25:56.582839 ignition[793]: Ignition 2.19.0 Apr 17 23:25:56.582848 ignition[793]: Stage: disks Apr 17 23:25:56.583023 ignition[793]: no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:56.583034 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:56.586199 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 17 23:25:56.584017 ignition[793]: disks: disks passed Apr 17 23:25:56.587595 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 17 23:25:56.584064 ignition[793]: Ignition finished successfully Apr 17 23:25:56.588423 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 17 23:25:56.589171 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 17 23:25:56.590196 systemd[1]: Reached target sysinit.target - System Initialization. Apr 17 23:25:56.591027 systemd[1]: Reached target basic.target - Basic System. Apr 17 23:25:56.604016 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 17 23:25:56.622524 systemd-fsck[801]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Apr 17 23:25:56.627397 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 17 23:25:56.634320 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 17 23:25:56.687227 kernel: EXT4-fs (sda9): mounted filesystem 2a4b2d55-130a-4cda-bef1-b1e6ed7bcf6b r/w with ordered data mode. Quota mode: none. Apr 17 23:25:56.689259 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 17 23:25:56.690995 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 17 23:25:56.698340 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 17 23:25:56.701343 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 17 23:25:56.704449 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 17 23:25:56.706336 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 17 23:25:56.706402 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 17 23:25:56.715301 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (809) Apr 17 23:25:56.716665 kernel: BTRFS info (device sda6): first mount of filesystem 511634b8-962b-4ed3-9161-3f02d13492ea Apr 17 23:25:56.716694 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 17 23:25:56.717187 kernel: BTRFS info (device sda6): using free space tree Apr 17 23:25:56.721310 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 17 23:25:56.723243 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 17 23:25:56.726692 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 17 23:25:56.726724 kernel: BTRFS info (device sda6): auto enabling async discard Apr 17 23:25:56.733286 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 17 23:25:56.773399 initrd-setup-root[838]: cut: /sysroot/etc/passwd: No such file or directory Apr 17 23:25:56.780919 coreos-metadata[811]: Apr 17 23:25:56.780 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Apr 17 23:25:56.782441 initrd-setup-root[845]: cut: /sysroot/etc/group: No such file or directory Apr 17 23:25:56.783461 coreos-metadata[811]: Apr 17 23:25:56.783 INFO Fetch successful Apr 17 23:25:56.784081 coreos-metadata[811]: Apr 17 23:25:56.783 INFO wrote hostname ci-4081-3-6-n-8954fb58c4 to /sysroot/etc/hostname Apr 17 23:25:56.786509 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 17 23:25:56.792108 initrd-setup-root[853]: cut: /sysroot/etc/shadow: No such file or directory Apr 17 23:25:56.796335 initrd-setup-root[860]: cut: /sysroot/etc/gshadow: No such file or directory Apr 17 23:25:56.898294 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 17 23:25:56.905314 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 17 23:25:56.910382 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 17 23:25:56.919234 kernel: BTRFS info (device sda6): last unmount of filesystem 511634b8-962b-4ed3-9161-3f02d13492ea Apr 17 23:25:56.946085 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 17 23:25:56.949682 ignition[928]: INFO : Ignition 2.19.0 Apr 17 23:25:56.949682 ignition[928]: INFO : Stage: mount Apr 17 23:25:56.950628 ignition[928]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:56.950628 ignition[928]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:56.953060 ignition[928]: INFO : mount: mount passed Apr 17 23:25:56.953060 ignition[928]: INFO : Ignition finished successfully Apr 17 23:25:56.954488 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 17 23:25:56.960349 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 17 23:25:57.106607 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 17 23:25:57.112715 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 17 23:25:57.122223 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (940) Apr 17 23:25:57.123532 kernel: BTRFS info (device sda6): first mount of filesystem 511634b8-962b-4ed3-9161-3f02d13492ea Apr 17 23:25:57.123579 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 17 23:25:57.123603 kernel: BTRFS info (device sda6): using free space tree Apr 17 23:25:57.127242 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 17 23:25:57.127299 kernel: BTRFS info (device sda6): auto enabling async discard Apr 17 23:25:57.129870 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 17 23:25:57.150839 ignition[957]: INFO : Ignition 2.19.0 Apr 17 23:25:57.151599 ignition[957]: INFO : Stage: files Apr 17 23:25:57.153239 ignition[957]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:57.153239 ignition[957]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:57.153239 ignition[957]: DEBUG : files: compiled without relabeling support, skipping Apr 17 23:25:57.156352 ignition[957]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 17 23:25:57.156352 ignition[957]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 17 23:25:57.160089 ignition[957]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 17 23:25:57.161397 ignition[957]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 17 23:25:57.162826 ignition[957]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 17 23:25:57.162686 unknown[957]: wrote ssh authorized keys file for user: core Apr 17 23:25:57.166817 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 17 23:25:57.166817 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 17 23:25:57.166817 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 17 23:25:57.166817 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Apr 17 23:25:57.236880 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 17 23:25:57.417442 systemd-networkd[777]: eth1: Gained IPv6LL Apr 17 23:25:57.692322 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Apr 17 23:25:57.692322 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 17 23:25:57.696537 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Apr 17 23:25:57.737719 systemd-networkd[777]: eth0: Gained IPv6LL Apr 17 23:25:58.065071 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Apr 17 23:25:58.384503 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 17 23:25:58.386015 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 17 23:25:58.394490 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 17 23:25:58.394490 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 17 23:25:58.394490 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 17 23:25:58.394490 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 17 23:25:58.394490 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-arm64.raw: attempt #1 Apr 17 23:25:58.655207 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Apr 17 23:25:59.477468 ignition[957]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Apr 17 23:25:59.477468 ignition[957]: INFO : files: op(d): [started] processing unit "containerd.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(d): [finished] processing unit "containerd.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(11): [started] processing unit "coreos-metadata.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(11): op(12): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(11): op(12): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(11): [finished] processing unit "coreos-metadata.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(13): [started] setting preset to enabled for "prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: op(13): [finished] setting preset to enabled for "prepare-helm.service" Apr 17 23:25:59.482070 ignition[957]: INFO : files: createResultFile: createFiles: op(14): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 17 23:25:59.482070 ignition[957]: INFO : files: createResultFile: createFiles: op(14): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 17 23:25:59.482070 ignition[957]: INFO : files: files passed Apr 17 23:25:59.482070 ignition[957]: INFO : Ignition finished successfully Apr 17 23:25:59.482666 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 17 23:25:59.492479 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 17 23:25:59.496342 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 17 23:25:59.504120 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 17 23:25:59.504226 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 17 23:25:59.514757 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 17 23:25:59.514757 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 17 23:25:59.517874 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 17 23:25:59.522216 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 17 23:25:59.523120 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 17 23:25:59.532506 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 17 23:25:59.563745 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 17 23:25:59.564529 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 17 23:25:59.566456 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 17 23:25:59.568054 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 17 23:25:59.569841 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 17 23:25:59.577502 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 17 23:25:59.590966 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 17 23:25:59.599547 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 17 23:25:59.610197 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 17 23:25:59.611630 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 17 23:25:59.613241 systemd[1]: Stopped target timers.target - Timer Units. Apr 17 23:25:59.613932 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 17 23:25:59.614075 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 17 23:25:59.615440 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 17 23:25:59.616139 systemd[1]: Stopped target basic.target - Basic System. Apr 17 23:25:59.617497 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 17 23:25:59.619128 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 17 23:25:59.619852 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 17 23:25:59.620908 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 17 23:25:59.621959 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 17 23:25:59.623095 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 17 23:25:59.624086 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 17 23:25:59.625165 systemd[1]: Stopped target swap.target - Swaps. Apr 17 23:25:59.626071 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 17 23:25:59.626208 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 17 23:25:59.627528 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 17 23:25:59.628125 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 17 23:25:59.629149 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 17 23:25:59.629233 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 17 23:25:59.630282 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 17 23:25:59.630401 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 17 23:25:59.631888 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 17 23:25:59.631999 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 17 23:25:59.633096 systemd[1]: ignition-files.service: Deactivated successfully. Apr 17 23:25:59.633196 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 17 23:25:59.634304 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 17 23:25:59.634430 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 17 23:25:59.645609 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 17 23:25:59.650547 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 17 23:25:59.651844 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 17 23:25:59.651982 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 17 23:25:59.655861 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 17 23:25:59.656309 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 17 23:25:59.665450 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 17 23:25:59.665820 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 17 23:25:59.670618 ignition[1010]: INFO : Ignition 2.19.0 Apr 17 23:25:59.670618 ignition[1010]: INFO : Stage: umount Apr 17 23:25:59.670618 ignition[1010]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 17 23:25:59.670618 ignition[1010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 17 23:25:59.670618 ignition[1010]: INFO : umount: umount passed Apr 17 23:25:59.670618 ignition[1010]: INFO : Ignition finished successfully Apr 17 23:25:59.671681 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 17 23:25:59.671797 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 17 23:25:59.672858 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 17 23:25:59.672902 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 17 23:25:59.675737 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 17 23:25:59.675787 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 17 23:25:59.676885 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 17 23:25:59.676922 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 17 23:25:59.677819 systemd[1]: Stopped target network.target - Network. Apr 17 23:25:59.678667 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 17 23:25:59.678719 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 17 23:25:59.679676 systemd[1]: Stopped target paths.target - Path Units. Apr 17 23:25:59.681634 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 17 23:25:59.685365 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 17 23:25:59.686681 systemd[1]: Stopped target slices.target - Slice Units. Apr 17 23:25:59.688415 systemd[1]: Stopped target sockets.target - Socket Units. Apr 17 23:25:59.689558 systemd[1]: iscsid.socket: Deactivated successfully. Apr 17 23:25:59.689622 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 17 23:25:59.692480 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 17 23:25:59.692546 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 17 23:25:59.694092 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 17 23:25:59.694151 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 17 23:25:59.695412 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 17 23:25:59.696426 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 17 23:25:59.697828 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 17 23:25:59.700486 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 17 23:25:59.702665 systemd-networkd[777]: eth0: DHCPv6 lease lost Apr 17 23:25:59.706254 systemd-networkd[777]: eth1: DHCPv6 lease lost Apr 17 23:25:59.707889 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 17 23:25:59.712227 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 17 23:25:59.712442 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 17 23:25:59.716680 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 17 23:25:59.716817 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 17 23:25:59.719611 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 17 23:25:59.719661 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 17 23:25:59.728416 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 17 23:25:59.729238 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 17 23:25:59.729304 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 17 23:25:59.732636 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 17 23:25:59.732700 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 17 23:25:59.733929 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 17 23:25:59.733978 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 17 23:25:59.735434 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 17 23:25:59.735508 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 17 23:25:59.738549 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 17 23:25:59.766963 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 17 23:25:59.767170 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 17 23:25:59.769424 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 17 23:25:59.769539 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 17 23:25:59.774533 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 17 23:25:59.775272 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 17 23:25:59.777080 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 17 23:25:59.777166 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 17 23:25:59.779040 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 17 23:25:59.779098 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 17 23:25:59.780658 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 17 23:25:59.780695 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 17 23:25:59.782002 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 17 23:25:59.782048 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 17 23:25:59.783486 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 17 23:25:59.783531 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 17 23:25:59.784922 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 17 23:25:59.784967 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 17 23:25:59.791416 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 17 23:25:59.791945 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 17 23:25:59.792001 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 17 23:25:59.793447 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Apr 17 23:25:59.793487 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 17 23:25:59.794980 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 17 23:25:59.795022 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 17 23:25:59.796908 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 17 23:25:59.796953 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:25:59.807426 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 17 23:25:59.807600 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 17 23:25:59.809693 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 17 23:25:59.823499 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 17 23:25:59.838371 systemd[1]: Switching root. Apr 17 23:25:59.869416 systemd-journald[238]: Journal stopped Apr 17 23:26:00.823546 systemd-journald[238]: Received SIGTERM from PID 1 (systemd). Apr 17 23:26:00.823627 kernel: SELinux: policy capability network_peer_controls=1 Apr 17 23:26:00.823645 kernel: SELinux: policy capability open_perms=1 Apr 17 23:26:00.823658 kernel: SELinux: policy capability extended_socket_class=1 Apr 17 23:26:00.823668 kernel: SELinux: policy capability always_check_network=0 Apr 17 23:26:00.823677 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 17 23:26:00.823692 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 17 23:26:00.823706 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 17 23:26:00.823718 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 17 23:26:00.823729 kernel: audit: type=1403 audit(1776468360.057:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 17 23:26:00.823740 systemd[1]: Successfully loaded SELinux policy in 37.331ms. Apr 17 23:26:00.823764 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 11.124ms. Apr 17 23:26:00.823776 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 17 23:26:00.823786 systemd[1]: Detected virtualization kvm. Apr 17 23:26:00.823797 systemd[1]: Detected architecture arm64. Apr 17 23:26:00.823813 systemd[1]: Detected first boot. Apr 17 23:26:00.823823 systemd[1]: Hostname set to . Apr 17 23:26:00.823834 systemd[1]: Initializing machine ID from VM UUID. Apr 17 23:26:00.823844 zram_generator::config[1070]: No configuration found. Apr 17 23:26:00.823855 systemd[1]: Populated /etc with preset unit settings. Apr 17 23:26:00.823866 systemd[1]: Queued start job for default target multi-user.target. Apr 17 23:26:00.823877 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 17 23:26:00.823888 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 17 23:26:00.823900 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 17 23:26:00.823911 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 17 23:26:00.823921 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 17 23:26:00.823932 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 17 23:26:00.823942 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 17 23:26:00.823957 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 17 23:26:00.823969 systemd[1]: Created slice user.slice - User and Session Slice. Apr 17 23:26:00.823980 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 17 23:26:00.823990 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 17 23:26:00.824002 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 17 23:26:00.824012 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 17 23:26:00.824024 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 17 23:26:00.824034 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 17 23:26:00.824046 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Apr 17 23:26:00.824057 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 17 23:26:00.824067 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 17 23:26:00.824079 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 17 23:26:00.824091 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 17 23:26:00.824109 systemd[1]: Reached target slices.target - Slice Units. Apr 17 23:26:00.824121 systemd[1]: Reached target swap.target - Swaps. Apr 17 23:26:00.824132 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 17 23:26:00.824143 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 17 23:26:00.824153 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 17 23:26:00.824164 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 17 23:26:00.824197 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 17 23:26:00.824213 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 17 23:26:00.824224 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 17 23:26:00.824236 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 17 23:26:00.824246 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 17 23:26:00.824257 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 17 23:26:00.824267 systemd[1]: Mounting media.mount - External Media Directory... Apr 17 23:26:00.824278 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 17 23:26:00.824292 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 17 23:26:00.824304 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 17 23:26:00.824315 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 17 23:26:00.824327 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 17 23:26:00.824346 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 17 23:26:00.824359 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 17 23:26:00.824370 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 17 23:26:00.824383 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 17 23:26:00.824394 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 17 23:26:00.824405 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 17 23:26:00.824415 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 17 23:26:00.824428 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 17 23:26:00.824439 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Apr 17 23:26:00.824450 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Apr 17 23:26:00.824460 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 17 23:26:00.824474 kernel: ACPI: bus type drm_connector registered Apr 17 23:26:00.824485 kernel: loop: module loaded Apr 17 23:26:00.824496 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 17 23:26:00.824506 kernel: fuse: init (API version 7.39) Apr 17 23:26:00.824516 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 17 23:26:00.824527 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 17 23:26:00.824538 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 17 23:26:00.824548 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 17 23:26:00.824580 systemd-journald[1155]: Collecting audit messages is disabled. Apr 17 23:26:00.824608 systemd-journald[1155]: Journal started Apr 17 23:26:00.824630 systemd-journald[1155]: Runtime Journal (/run/log/journal/a295e9f57e274b21a3592359c019d2b7) is 8.0M, max 76.6M, 68.6M free. Apr 17 23:26:00.828221 systemd[1]: Started systemd-journald.service - Journal Service. Apr 17 23:26:00.829745 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 17 23:26:00.831980 systemd[1]: Mounted media.mount - External Media Directory. Apr 17 23:26:00.832730 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 17 23:26:00.833698 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 17 23:26:00.836515 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 17 23:26:00.838141 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 17 23:26:00.841772 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 17 23:26:00.841958 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 17 23:26:00.843019 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 17 23:26:00.843249 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 17 23:26:00.845590 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 17 23:26:00.845788 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 17 23:26:00.846786 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 17 23:26:00.846945 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 17 23:26:00.848068 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 17 23:26:00.848256 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 17 23:26:00.849616 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 17 23:26:00.849827 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 17 23:26:00.852950 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 17 23:26:00.853915 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 17 23:26:00.858658 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 17 23:26:00.864452 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 17 23:26:00.871420 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 17 23:26:00.879310 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 17 23:26:00.885496 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 17 23:26:00.886324 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 17 23:26:00.893569 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 17 23:26:00.902427 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 17 23:26:00.903539 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 17 23:26:00.909258 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 17 23:26:00.909868 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 17 23:26:00.915451 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 17 23:26:00.922626 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 17 23:26:00.925955 systemd-journald[1155]: Time spent on flushing to /var/log/journal/a295e9f57e274b21a3592359c019d2b7 is 31.336ms for 1114 entries. Apr 17 23:26:00.925955 systemd-journald[1155]: System Journal (/var/log/journal/a295e9f57e274b21a3592359c019d2b7) is 8.0M, max 584.8M, 576.8M free. Apr 17 23:26:00.972126 systemd-journald[1155]: Received client request to flush runtime journal. Apr 17 23:26:00.928492 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 17 23:26:00.930775 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 17 23:26:00.939634 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 17 23:26:00.947467 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 17 23:26:00.962596 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 17 23:26:00.963374 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 17 23:26:00.982304 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 17 23:26:00.988695 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Apr 17 23:26:00.989022 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Apr 17 23:26:00.992658 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 17 23:26:00.993805 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 17 23:26:00.997436 udevadm[1213]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 17 23:26:01.016745 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 17 23:26:01.055046 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 17 23:26:01.062405 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 17 23:26:01.076315 systemd-tmpfiles[1229]: ACLs are not supported, ignoring. Apr 17 23:26:01.076382 systemd-tmpfiles[1229]: ACLs are not supported, ignoring. Apr 17 23:26:01.082712 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 17 23:26:01.435920 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 17 23:26:01.444887 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 17 23:26:01.466953 systemd-udevd[1235]: Using default interface naming scheme 'v255'. Apr 17 23:26:01.489070 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 17 23:26:01.502518 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 17 23:26:01.526599 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 17 23:26:01.557894 systemd[1]: Found device dev-ttyAMA0.device - /dev/ttyAMA0. Apr 17 23:26:01.596974 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 17 23:26:01.677369 systemd-networkd[1242]: lo: Link UP Apr 17 23:26:01.677708 systemd-networkd[1242]: lo: Gained carrier Apr 17 23:26:01.680518 systemd-networkd[1242]: Enumeration completed Apr 17 23:26:01.681568 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 17 23:26:01.682115 systemd-networkd[1242]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.682196 systemd-networkd[1242]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 17 23:26:01.683541 systemd-networkd[1242]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.683612 systemd-networkd[1242]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 17 23:26:01.684130 systemd-networkd[1242]: eth0: Link UP Apr 17 23:26:01.684134 systemd-networkd[1242]: eth0: Gained carrier Apr 17 23:26:01.684147 systemd-networkd[1242]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.692656 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 17 23:26:01.695496 systemd-networkd[1242]: eth1: Link UP Apr 17 23:26:01.695505 systemd-networkd[1242]: eth1: Gained carrier Apr 17 23:26:01.695544 systemd-networkd[1242]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.699615 systemd-networkd[1242]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.707204 kernel: mousedev: PS/2 mouse device common for all mice Apr 17 23:26:01.722570 systemd-networkd[1242]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 17 23:26:01.746739 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (1241) Apr 17 23:26:01.749937 systemd-networkd[1242]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 17 23:26:01.757497 systemd-networkd[1242]: eth0: DHCPv4 address 46.224.37.37/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 17 23:26:01.795535 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 17 23:26:01.803425 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 17 23:26:01.816425 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 17 23:26:01.820393 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 17 23:26:01.820960 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 17 23:26:01.820998 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 17 23:26:01.825995 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 17 23:26:01.826592 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 17 23:26:01.842714 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 17 23:26:01.842945 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 17 23:26:01.844040 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 17 23:26:01.844427 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 17 23:26:01.845198 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 17 23:26:01.845257 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 17 23:26:01.856602 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 17 23:26:01.862213 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Apr 17 23:26:01.862288 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Apr 17 23:26:01.862300 kernel: [drm] features: -context_init Apr 17 23:26:01.863243 kernel: [drm] number of scanouts: 1 Apr 17 23:26:01.863311 kernel: [drm] number of cap sets: 0 Apr 17 23:26:01.866212 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Apr 17 23:26:01.866796 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 17 23:26:01.872203 kernel: Console: switching to colour frame buffer device 160x50 Apr 17 23:26:01.881201 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Apr 17 23:26:01.884746 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 17 23:26:01.885013 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:26:01.892548 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 17 23:26:01.957692 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 17 23:26:02.018116 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 17 23:26:02.036662 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 17 23:26:02.053224 lvm[1308]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 17 23:26:02.079925 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 17 23:26:02.082014 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 17 23:26:02.090502 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 17 23:26:02.097115 lvm[1311]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 17 23:26:02.128728 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 17 23:26:02.130237 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 17 23:26:02.131553 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 17 23:26:02.131695 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 17 23:26:02.132636 systemd[1]: Reached target machines.target - Containers. Apr 17 23:26:02.134874 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 17 23:26:02.142533 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 17 23:26:02.146170 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 17 23:26:02.147285 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 17 23:26:02.154721 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 17 23:26:02.161727 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 17 23:26:02.172503 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 17 23:26:02.181691 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 17 23:26:02.193086 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 17 23:26:02.200356 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 17 23:26:02.204476 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 17 23:26:02.209995 kernel: loop0: detected capacity change from 0 to 209336 Apr 17 23:26:02.233201 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 17 23:26:02.265269 kernel: loop1: detected capacity change from 0 to 114328 Apr 17 23:26:02.294253 kernel: loop2: detected capacity change from 0 to 8 Apr 17 23:26:02.324228 kernel: loop3: detected capacity change from 0 to 114432 Apr 17 23:26:02.360230 kernel: loop4: detected capacity change from 0 to 209336 Apr 17 23:26:02.383196 kernel: loop5: detected capacity change from 0 to 114328 Apr 17 23:26:02.395205 kernel: loop6: detected capacity change from 0 to 8 Apr 17 23:26:02.397310 kernel: loop7: detected capacity change from 0 to 114432 Apr 17 23:26:02.408937 (sd-merge)[1333]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Apr 17 23:26:02.409509 (sd-merge)[1333]: Merged extensions into '/usr'. Apr 17 23:26:02.416898 systemd[1]: Reloading requested from client PID 1319 ('systemd-sysext') (unit systemd-sysext.service)... Apr 17 23:26:02.416922 systemd[1]: Reloading... Apr 17 23:26:02.519210 zram_generator::config[1361]: No configuration found. Apr 17 23:26:02.615974 ldconfig[1315]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 17 23:26:02.647899 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 17 23:26:02.707950 systemd[1]: Reloading finished in 290 ms. Apr 17 23:26:02.727294 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 17 23:26:02.728658 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 17 23:26:02.740346 systemd[1]: Starting ensure-sysext.service... Apr 17 23:26:02.743883 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 17 23:26:02.751104 systemd[1]: Reloading requested from client PID 1405 ('systemctl') (unit ensure-sysext.service)... Apr 17 23:26:02.751284 systemd[1]: Reloading... Apr 17 23:26:02.780037 systemd-tmpfiles[1406]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 17 23:26:02.780769 systemd-tmpfiles[1406]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 17 23:26:02.781669 systemd-tmpfiles[1406]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 17 23:26:02.782086 systemd-tmpfiles[1406]: ACLs are not supported, ignoring. Apr 17 23:26:02.782253 systemd-tmpfiles[1406]: ACLs are not supported, ignoring. Apr 17 23:26:02.785497 systemd-tmpfiles[1406]: Detected autofs mount point /boot during canonicalization of boot. Apr 17 23:26:02.785644 systemd-tmpfiles[1406]: Skipping /boot Apr 17 23:26:02.794265 systemd-tmpfiles[1406]: Detected autofs mount point /boot during canonicalization of boot. Apr 17 23:26:02.794414 systemd-tmpfiles[1406]: Skipping /boot Apr 17 23:26:02.818218 zram_generator::config[1434]: No configuration found. Apr 17 23:26:02.948727 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 17 23:26:03.011021 systemd[1]: Reloading finished in 259 ms. Apr 17 23:26:03.029294 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 17 23:26:03.043464 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 17 23:26:03.054497 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 17 23:26:03.058445 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 17 23:26:03.069435 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 17 23:26:03.075387 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 17 23:26:03.089147 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 17 23:26:03.091588 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 17 23:26:03.098486 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 17 23:26:03.110644 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 17 23:26:03.111302 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 17 23:26:03.120475 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 17 23:26:03.120697 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 17 23:26:03.122559 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 17 23:26:03.122824 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 17 23:26:03.128785 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 17 23:26:03.133334 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 17 23:26:03.151586 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 17 23:26:03.152289 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 17 23:26:03.153637 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 17 23:26:03.156257 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 17 23:26:03.156447 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 17 23:26:03.162276 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 17 23:26:03.162944 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 17 23:26:03.164817 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 17 23:26:03.175359 systemd[1]: Finished ensure-sysext.service. Apr 17 23:26:03.183271 systemd-networkd[1242]: eth1: Gained IPv6LL Apr 17 23:26:03.185825 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 17 23:26:03.186885 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 17 23:26:03.190692 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 17 23:26:03.192493 augenrules[1513]: No rules Apr 17 23:26:03.194514 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 17 23:26:03.194737 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 17 23:26:03.197483 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 17 23:26:03.200856 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 17 23:26:03.204689 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 17 23:26:03.204790 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 17 23:26:03.212384 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 17 23:26:03.220372 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 17 23:26:03.226265 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 17 23:26:03.238357 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 17 23:26:03.243912 systemd-resolved[1484]: Positive Trust Anchors: Apr 17 23:26:03.243934 systemd-resolved[1484]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 17 23:26:03.243966 systemd-resolved[1484]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 17 23:26:03.249568 systemd-resolved[1484]: Using system hostname 'ci-4081-3-6-n-8954fb58c4'. Apr 17 23:26:03.251971 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 17 23:26:03.252960 systemd[1]: Reached target network.target - Network. Apr 17 23:26:03.253942 systemd[1]: Reached target network-online.target - Network is Online. Apr 17 23:26:03.254739 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 17 23:26:03.280217 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 17 23:26:03.282026 systemd[1]: Reached target sysinit.target - System Initialization. Apr 17 23:26:03.283887 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 17 23:26:03.285032 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 17 23:26:03.285788 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 17 23:26:03.286629 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 17 23:26:03.286706 systemd[1]: Reached target paths.target - Path Units. Apr 17 23:26:03.287168 systemd[1]: Reached target time-set.target - System Time Set. Apr 17 23:26:03.287921 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 17 23:26:03.288654 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 17 23:26:03.289283 systemd[1]: Reached target timers.target - Timer Units. Apr 17 23:26:03.290525 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 17 23:26:03.292876 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 17 23:26:03.295386 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 17 23:26:03.302826 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 17 23:26:03.303914 systemd[1]: Reached target sockets.target - Socket Units. Apr 17 23:26:03.304783 systemd[1]: Reached target basic.target - Basic System. Apr 17 23:26:03.305780 systemd[1]: System is tainted: cgroupsv1 Apr 17 23:26:03.305833 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 17 23:26:03.305865 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 17 23:26:03.318478 systemd[1]: Starting containerd.service - containerd container runtime... Apr 17 23:26:03.324566 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 17 23:26:03.330985 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 17 23:26:03.335344 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 17 23:26:03.337958 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 17 23:26:03.340257 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 17 23:26:03.344251 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:03.354748 jq[1546]: false Apr 17 23:26:03.359398 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 17 23:26:03.365162 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 17 23:26:03.369483 systemd-networkd[1242]: eth0: Gained IPv6LL Apr 17 23:26:03.371676 coreos-metadata[1542]: Apr 17 23:26:03.371 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Apr 17 23:26:03.376144 coreos-metadata[1542]: Apr 17 23:26:03.375 INFO Fetch successful Apr 17 23:26:03.376144 coreos-metadata[1542]: Apr 17 23:26:03.375 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Apr 17 23:26:03.375901 dbus-daemon[1543]: [system] SELinux support is enabled Apr 17 23:26:03.375306 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 17 23:26:03.382525 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Apr 17 23:26:03.389415 coreos-metadata[1542]: Apr 17 23:26:03.383 INFO Fetch successful Apr 17 23:26:03.383399 systemd-timesyncd[1533]: Contacted time server 84.46.245.73:123 (0.flatcar.pool.ntp.org). Apr 17 23:26:03.383472 systemd-timesyncd[1533]: Initial clock synchronization to Fri 2026-04-17 23:26:03.458319 UTC. Apr 17 23:26:03.398497 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 17 23:26:03.409966 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 17 23:26:03.417915 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 17 23:26:03.421088 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 17 23:26:03.428955 systemd[1]: Starting update-engine.service - Update Engine... Apr 17 23:26:03.433104 extend-filesystems[1547]: Found loop4 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found loop5 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found loop6 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found loop7 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda1 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda2 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda3 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found usr Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda4 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda6 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda7 Apr 17 23:26:03.437056 extend-filesystems[1547]: Found sda9 Apr 17 23:26:03.437056 extend-filesystems[1547]: Checking size of /dev/sda9 Apr 17 23:26:03.445103 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 17 23:26:03.453820 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 17 23:26:03.463004 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 17 23:26:03.472599 jq[1573]: true Apr 17 23:26:03.463271 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 17 23:26:03.468506 systemd[1]: motdgen.service: Deactivated successfully. Apr 17 23:26:03.468753 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 17 23:26:03.482723 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 17 23:26:03.482961 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 17 23:26:03.506910 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 17 23:26:03.518145 extend-filesystems[1547]: Resized partition /dev/sda9 Apr 17 23:26:03.536459 extend-filesystems[1602]: resize2fs 1.47.1 (20-May-2024) Apr 17 23:26:03.536405 (ntainerd)[1594]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 17 23:26:03.543520 jq[1589]: true Apr 17 23:26:03.543664 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 17 23:26:03.556335 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Apr 17 23:26:03.556559 tar[1585]: linux-arm64/LICENSE Apr 17 23:26:03.556559 tar[1585]: linux-arm64/helm Apr 17 23:26:03.543710 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 17 23:26:03.545303 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 17 23:26:03.545360 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 17 23:26:03.589545 update_engine[1567]: I20260417 23:26:03.588342 1567 main.cc:92] Flatcar Update Engine starting Apr 17 23:26:03.603058 systemd-logind[1565]: New seat seat0. Apr 17 23:26:03.608455 systemd-logind[1565]: Watching system buttons on /dev/input/event0 (Power Button) Apr 17 23:26:03.618073 update_engine[1567]: I20260417 23:26:03.610024 1567 update_check_scheduler.cc:74] Next update check in 9m56s Apr 17 23:26:03.608471 systemd-logind[1565]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Apr 17 23:26:03.616668 systemd[1]: Started systemd-logind.service - User Login Management. Apr 17 23:26:03.617346 systemd[1]: Started update-engine.service - Update Engine. Apr 17 23:26:03.620500 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 17 23:26:03.623480 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 17 23:26:03.651133 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 17 23:26:03.653706 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 17 23:26:03.703252 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (1253) Apr 17 23:26:03.750295 bash[1635]: Updated "/home/core/.ssh/authorized_keys" Apr 17 23:26:03.761365 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 17 23:26:03.775276 systemd[1]: Starting sshkeys.service... Apr 17 23:26:03.789203 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Apr 17 23:26:03.802508 locksmithd[1620]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 17 23:26:03.809950 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 17 23:26:03.820975 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 17 23:26:03.823717 extend-filesystems[1602]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Apr 17 23:26:03.823717 extend-filesystems[1602]: old_desc_blocks = 1, new_desc_blocks = 5 Apr 17 23:26:03.823717 extend-filesystems[1602]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Apr 17 23:26:03.857283 extend-filesystems[1547]: Resized filesystem in /dev/sda9 Apr 17 23:26:03.857283 extend-filesystems[1547]: Found sr0 Apr 17 23:26:03.834730 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 17 23:26:03.836627 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 17 23:26:03.928411 coreos-metadata[1649]: Apr 17 23:26:03.928 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Apr 17 23:26:03.929743 coreos-metadata[1649]: Apr 17 23:26:03.929 INFO Fetch successful Apr 17 23:26:03.935283 unknown[1649]: wrote ssh authorized keys file for user: core Apr 17 23:26:03.969478 update-ssh-keys[1656]: Updated "/home/core/.ssh/authorized_keys" Apr 17 23:26:03.972566 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 17 23:26:03.979209 systemd[1]: Finished sshkeys.service. Apr 17 23:26:04.019227 containerd[1594]: time="2026-04-17T23:26:04.018948463Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 17 23:26:04.094861 containerd[1594]: time="2026-04-17T23:26:04.093494404Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.099403 containerd[1594]: time="2026-04-17T23:26:04.099351631Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.127-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 17 23:26:04.099726 containerd[1594]: time="2026-04-17T23:26:04.099705791Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 17 23:26:04.099827 containerd[1594]: time="2026-04-17T23:26:04.099813037Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 17 23:26:04.100073 containerd[1594]: time="2026-04-17T23:26:04.100056035Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 17 23:26:04.100159 containerd[1594]: time="2026-04-17T23:26:04.100145514Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.100389 containerd[1594]: time="2026-04-17T23:26:04.100357137Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 17 23:26:04.101063 containerd[1594]: time="2026-04-17T23:26:04.100437894Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.102709 containerd[1594]: time="2026-04-17T23:26:04.102509151Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 17 23:26:04.102709 containerd[1594]: time="2026-04-17T23:26:04.102533661Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.102709 containerd[1594]: time="2026-04-17T23:26:04.102559544Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 17 23:26:04.102709 containerd[1594]: time="2026-04-17T23:26:04.102570446Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.102709 containerd[1594]: time="2026-04-17T23:26:04.102672725Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.103482 containerd[1594]: time="2026-04-17T23:26:04.103461600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 17 23:26:04.104309 containerd[1594]: time="2026-04-17T23:26:04.104275227Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 17 23:26:04.104425 containerd[1594]: time="2026-04-17T23:26:04.104408961Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 17 23:26:04.105117 containerd[1594]: time="2026-04-17T23:26:04.104963640Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 17 23:26:04.105117 containerd[1594]: time="2026-04-17T23:26:04.105053604Z" level=info msg="metadata content store policy set" policy=shared Apr 17 23:26:04.111579 containerd[1594]: time="2026-04-17T23:26:04.111531651Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 17 23:26:04.111733 containerd[1594]: time="2026-04-17T23:26:04.111719452Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 17 23:26:04.112670 containerd[1594]: time="2026-04-17T23:26:04.111827505Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 17 23:26:04.112670 containerd[1594]: time="2026-04-17T23:26:04.111851126Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 17 23:26:04.112670 containerd[1594]: time="2026-04-17T23:26:04.111870225Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 17 23:26:04.112670 containerd[1594]: time="2026-04-17T23:26:04.112036706Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 17 23:26:04.113237 containerd[1594]: time="2026-04-17T23:26:04.113213780Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 17 23:26:04.113444 containerd[1594]: time="2026-04-17T23:26:04.113424314Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 17 23:26:04.113516 containerd[1594]: time="2026-04-17T23:26:04.113495582Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 17 23:26:04.113727 containerd[1594]: time="2026-04-17T23:26:04.113697394Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 17 23:26:04.113786 containerd[1594]: time="2026-04-17T23:26:04.113736723Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113786 containerd[1594]: time="2026-04-17T23:26:04.113752834Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113786 containerd[1594]: time="2026-04-17T23:26:04.113769227Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113786 containerd[1594]: time="2026-04-17T23:26:04.113784329Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113864 containerd[1594]: time="2026-04-17T23:26:04.113802055Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113864 containerd[1594]: time="2026-04-17T23:26:04.113815541Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113864 containerd[1594]: time="2026-04-17T23:26:04.113827938Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113864 containerd[1594]: time="2026-04-17T23:26:04.113840334Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 17 23:26:04.113864 containerd[1594]: time="2026-04-17T23:26:04.113861250Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113876513Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113890686Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113904334Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113916609Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113933326Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.113957 containerd[1594]: time="2026-04-17T23:26:04.113946368Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.113960218Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.113974108Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.113989008Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114000596Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114012023Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114025066Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114042025Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114064960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114077558Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114109 containerd[1594]: time="2026-04-17T23:26:04.114090317Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114220175Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114242988Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114255183Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114266933Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114277149Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114288697Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114298549Z" level=info msg="NRI interface is disabled by configuration." Apr 17 23:26:04.114499 containerd[1594]: time="2026-04-17T23:26:04.114308886Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 17 23:26:04.114763 containerd[1594]: time="2026-04-17T23:26:04.114676613Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 17 23:26:04.114763 containerd[1594]: time="2026-04-17T23:26:04.114749295Z" level=info msg="Connect containerd service" Apr 17 23:26:04.114939 containerd[1594]: time="2026-04-17T23:26:04.114784222Z" level=info msg="using legacy CRI server" Apr 17 23:26:04.114939 containerd[1594]: time="2026-04-17T23:26:04.114792096Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 17 23:26:04.114939 containerd[1594]: time="2026-04-17T23:26:04.114889570Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 17 23:26:04.119843 containerd[1594]: time="2026-04-17T23:26:04.119782397Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.119965958Z" level=info msg="Start subscribing containerd event" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.120034318Z" level=info msg="Start recovering state" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.120110270Z" level=info msg="Start event monitor" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.120122344Z" level=info msg="Start snapshots syncer" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.120131308Z" level=info msg="Start cni network conf syncer for default" Apr 17 23:26:04.121253 containerd[1594]: time="2026-04-17T23:26:04.120138858Z" level=info msg="Start streaming server" Apr 17 23:26:04.123012 containerd[1594]: time="2026-04-17T23:26:04.122977913Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 17 23:26:04.123070 containerd[1594]: time="2026-04-17T23:26:04.123055399Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 17 23:26:04.126601 containerd[1594]: time="2026-04-17T23:26:04.123792146Z" level=info msg="containerd successfully booted in 0.108058s" Apr 17 23:26:04.123924 systemd[1]: Started containerd.service - containerd container runtime. Apr 17 23:26:04.377696 tar[1585]: linux-arm64/README.md Apr 17 23:26:04.399683 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 17 23:26:04.569699 sshd_keygen[1586]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 17 23:26:04.597554 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 17 23:26:04.609639 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 17 23:26:04.621227 systemd[1]: issuegen.service: Deactivated successfully. Apr 17 23:26:04.622489 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 17 23:26:04.631678 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 17 23:26:04.643707 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 17 23:26:04.656433 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 17 23:26:04.663570 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Apr 17 23:26:04.664492 systemd[1]: Reached target getty.target - Login Prompts. Apr 17 23:26:04.711429 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:04.713637 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 17 23:26:04.715210 systemd[1]: Startup finished in 7.113s (kernel) + 4.695s (userspace) = 11.809s. Apr 17 23:26:04.728707 (kubelet)[1700]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 17 23:26:05.276611 kubelet[1700]: E0417 23:26:05.276567 1700 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 17 23:26:05.282443 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 17 23:26:05.282690 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 17 23:26:06.667783 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 17 23:26:06.674639 systemd[1]: Started sshd@0-46.224.37.37:22-50.85.169.122:59218.service - OpenSSH per-connection server daemon (50.85.169.122:59218). Apr 17 23:26:06.799765 sshd[1712]: Accepted publickey for core from 50.85.169.122 port 59218 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:06.801977 sshd[1712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:06.812505 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 17 23:26:06.824606 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 17 23:26:06.830090 systemd-logind[1565]: New session 1 of user core. Apr 17 23:26:06.842110 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 17 23:26:06.849632 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 17 23:26:06.855633 (systemd)[1718]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 17 23:26:06.990697 systemd[1718]: Queued start job for default target default.target. Apr 17 23:26:06.991475 systemd[1718]: Created slice app.slice - User Application Slice. Apr 17 23:26:06.991603 systemd[1718]: Reached target paths.target - Paths. Apr 17 23:26:06.991683 systemd[1718]: Reached target timers.target - Timers. Apr 17 23:26:07.000873 systemd[1718]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 17 23:26:07.011981 systemd[1718]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 17 23:26:07.012051 systemd[1718]: Reached target sockets.target - Sockets. Apr 17 23:26:07.012064 systemd[1718]: Reached target basic.target - Basic System. Apr 17 23:26:07.012111 systemd[1718]: Reached target default.target - Main User Target. Apr 17 23:26:07.012136 systemd[1718]: Startup finished in 149ms. Apr 17 23:26:07.012573 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 17 23:26:07.028141 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 17 23:26:07.151719 systemd[1]: Started sshd@1-46.224.37.37:22-50.85.169.122:59228.service - OpenSSH per-connection server daemon (50.85.169.122:59228). Apr 17 23:26:07.276057 sshd[1730]: Accepted publickey for core from 50.85.169.122 port 59228 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:07.277422 sshd[1730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:07.283909 systemd-logind[1565]: New session 2 of user core. Apr 17 23:26:07.290734 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 17 23:26:07.396364 sshd[1730]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:07.403489 systemd-logind[1565]: Session 2 logged out. Waiting for processes to exit. Apr 17 23:26:07.404229 systemd[1]: sshd@1-46.224.37.37:22-50.85.169.122:59228.service: Deactivated successfully. Apr 17 23:26:07.407527 systemd[1]: session-2.scope: Deactivated successfully. Apr 17 23:26:07.426797 systemd[1]: Started sshd@2-46.224.37.37:22-50.85.169.122:59240.service - OpenSSH per-connection server daemon (50.85.169.122:59240). Apr 17 23:26:07.429255 systemd-logind[1565]: Removed session 2. Apr 17 23:26:07.544382 sshd[1738]: Accepted publickey for core from 50.85.169.122 port 59240 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:07.546401 sshd[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:07.552261 systemd-logind[1565]: New session 3 of user core. Apr 17 23:26:07.562583 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 17 23:26:07.656496 sshd[1738]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:07.663333 systemd-logind[1565]: Session 3 logged out. Waiting for processes to exit. Apr 17 23:26:07.664482 systemd[1]: sshd@2-46.224.37.37:22-50.85.169.122:59240.service: Deactivated successfully. Apr 17 23:26:07.666125 systemd[1]: session-3.scope: Deactivated successfully. Apr 17 23:26:07.667489 systemd-logind[1565]: Removed session 3. Apr 17 23:26:07.680715 systemd[1]: Started sshd@3-46.224.37.37:22-50.85.169.122:59254.service - OpenSSH per-connection server daemon (50.85.169.122:59254). Apr 17 23:26:07.805335 sshd[1746]: Accepted publickey for core from 50.85.169.122 port 59254 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:07.807345 sshd[1746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:07.813773 systemd-logind[1565]: New session 4 of user core. Apr 17 23:26:07.819612 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 17 23:26:07.924510 sshd[1746]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:07.929107 systemd[1]: sshd@3-46.224.37.37:22-50.85.169.122:59254.service: Deactivated successfully. Apr 17 23:26:07.933113 systemd[1]: session-4.scope: Deactivated successfully. Apr 17 23:26:07.934384 systemd-logind[1565]: Session 4 logged out. Waiting for processes to exit. Apr 17 23:26:07.952798 systemd[1]: Started sshd@4-46.224.37.37:22-50.85.169.122:59258.service - OpenSSH per-connection server daemon (50.85.169.122:59258). Apr 17 23:26:07.955017 systemd-logind[1565]: Removed session 4. Apr 17 23:26:08.069317 sshd[1754]: Accepted publickey for core from 50.85.169.122 port 59258 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:08.070962 sshd[1754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:08.076239 systemd-logind[1565]: New session 5 of user core. Apr 17 23:26:08.080430 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 17 23:26:08.176570 sudo[1758]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 17 23:26:08.176864 sudo[1758]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 17 23:26:08.198795 sudo[1758]: pam_unix(sudo:session): session closed for user root Apr 17 23:26:08.214648 sshd[1754]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:08.220910 systemd[1]: sshd@4-46.224.37.37:22-50.85.169.122:59258.service: Deactivated successfully. Apr 17 23:26:08.224347 systemd[1]: session-5.scope: Deactivated successfully. Apr 17 23:26:08.225172 systemd-logind[1565]: Session 5 logged out. Waiting for processes to exit. Apr 17 23:26:08.226304 systemd-logind[1565]: Removed session 5. Apr 17 23:26:08.244716 systemd[1]: Started sshd@5-46.224.37.37:22-50.85.169.122:59270.service - OpenSSH per-connection server daemon (50.85.169.122:59270). Apr 17 23:26:08.370571 sshd[1763]: Accepted publickey for core from 50.85.169.122 port 59270 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:08.373785 sshd[1763]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:08.379716 systemd-logind[1565]: New session 6 of user core. Apr 17 23:26:08.390870 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 17 23:26:08.480798 sudo[1768]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 17 23:26:08.481706 sudo[1768]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 17 23:26:08.486354 sudo[1768]: pam_unix(sudo:session): session closed for user root Apr 17 23:26:08.493863 sudo[1767]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 17 23:26:08.494166 sudo[1767]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 17 23:26:08.513987 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 17 23:26:08.516218 auditctl[1771]: No rules Apr 17 23:26:08.515854 systemd[1]: audit-rules.service: Deactivated successfully. Apr 17 23:26:08.517553 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 17 23:26:08.522678 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 17 23:26:08.567891 augenrules[1790]: No rules Apr 17 23:26:08.570659 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 17 23:26:08.573413 sudo[1767]: pam_unix(sudo:session): session closed for user root Apr 17 23:26:08.591266 sshd[1763]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:08.595488 systemd-logind[1565]: Session 6 logged out. Waiting for processes to exit. Apr 17 23:26:08.596891 systemd[1]: sshd@5-46.224.37.37:22-50.85.169.122:59270.service: Deactivated successfully. Apr 17 23:26:08.602023 systemd[1]: session-6.scope: Deactivated successfully. Apr 17 23:26:08.603113 systemd-logind[1565]: Removed session 6. Apr 17 23:26:08.614593 systemd[1]: Started sshd@6-46.224.37.37:22-50.85.169.122:59282.service - OpenSSH per-connection server daemon (50.85.169.122:59282). Apr 17 23:26:08.728616 sshd[1799]: Accepted publickey for core from 50.85.169.122 port 59282 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:26:08.731536 sshd[1799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:26:08.736284 systemd-logind[1565]: New session 7 of user core. Apr 17 23:26:08.743680 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 17 23:26:08.830626 sudo[1803]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 17 23:26:08.830895 sudo[1803]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 17 23:26:09.133585 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 17 23:26:09.134885 (dockerd)[1819]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 17 23:26:09.384325 dockerd[1819]: time="2026-04-17T23:26:09.383869938Z" level=info msg="Starting up" Apr 17 23:26:09.583767 dockerd[1819]: time="2026-04-17T23:26:09.583709493Z" level=info msg="Loading containers: start." Apr 17 23:26:09.700233 kernel: Initializing XFRM netlink socket Apr 17 23:26:09.793872 systemd-networkd[1242]: docker0: Link UP Apr 17 23:26:09.818375 dockerd[1819]: time="2026-04-17T23:26:09.818262442Z" level=info msg="Loading containers: done." Apr 17 23:26:09.847046 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4114009858-merged.mount: Deactivated successfully. Apr 17 23:26:09.850384 dockerd[1819]: time="2026-04-17T23:26:09.847134300Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 17 23:26:09.850384 dockerd[1819]: time="2026-04-17T23:26:09.847335794Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 17 23:26:09.850384 dockerd[1819]: time="2026-04-17T23:26:09.847582586Z" level=info msg="Daemon has completed initialization" Apr 17 23:26:09.891621 dockerd[1819]: time="2026-04-17T23:26:09.891443485Z" level=info msg="API listen on /run/docker.sock" Apr 17 23:26:09.894336 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 17 23:26:10.398693 containerd[1594]: time="2026-04-17T23:26:10.398650252Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\"" Apr 17 23:26:10.963130 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1085758779.mount: Deactivated successfully. Apr 17 23:26:11.877143 containerd[1594]: time="2026-04-17T23:26:11.875865442Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:11.878687 containerd[1594]: time="2026-04-17T23:26:11.878656375Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.11: active requests=0, bytes read=27008885" Apr 17 23:26:11.880355 containerd[1594]: time="2026-04-17T23:26:11.880315598Z" level=info msg="ImageCreate event name:\"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:11.883733 containerd[1594]: time="2026-04-17T23:26:11.883702939Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:11.885632 containerd[1594]: time="2026-04-17T23:26:11.885598919Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.11\" with image id \"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:18e9f2b6e4d67c24941e14b2d41ec0aa6e5f628e39f2ef2163e176de85bbe39e\", size \"27005386\" in 1.486906051s" Apr 17 23:26:11.886289 containerd[1594]: time="2026-04-17T23:26:11.886261090Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.11\" returns image reference \"sha256:51b83c5cb2f791f72696c040be904535bad3c81a6ffc19a55013ac150a24d9b0\"" Apr 17 23:26:11.887105 containerd[1594]: time="2026-04-17T23:26:11.887074783Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\"" Apr 17 23:26:13.001035 containerd[1594]: time="2026-04-17T23:26:13.000957783Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.003326 containerd[1594]: time="2026-04-17T23:26:13.003277615Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.11: active requests=0, bytes read=23297794" Apr 17 23:26:13.004159 containerd[1594]: time="2026-04-17T23:26:13.003650272Z" level=info msg="ImageCreate event name:\"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.007203 containerd[1594]: time="2026-04-17T23:26:13.007123646Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.008378 containerd[1594]: time="2026-04-17T23:26:13.008334602Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.11\" with image id \"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7579451c5b3c2715da4a263c5d80a3367a24fdc12e86fde6851674d567d1dfb2\", size \"24804413\" in 1.121106862s" Apr 17 23:26:13.008378 containerd[1594]: time="2026-04-17T23:26:13.008378567Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.11\" returns image reference \"sha256:df8bcecad66863646fb4016494163838761da38376bae5a7592e04041db8489a\"" Apr 17 23:26:13.009206 containerd[1594]: time="2026-04-17T23:26:13.008833216Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\"" Apr 17 23:26:13.931434 containerd[1594]: time="2026-04-17T23:26:13.931389136Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.933081 containerd[1594]: time="2026-04-17T23:26:13.933040541Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.11: active requests=0, bytes read=18141378" Apr 17 23:26:13.934897 containerd[1594]: time="2026-04-17T23:26:13.934841972Z" level=info msg="ImageCreate event name:\"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.941422 containerd[1594]: time="2026-04-17T23:26:13.941077102Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:13.942572 containerd[1594]: time="2026-04-17T23:26:13.942286092Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.11\" with image id \"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:5506f0f94c4d9aeb071664893aabc12166bcb7f775008a6fff02d004e6091d28\", size \"19648015\" in 933.41926ms" Apr 17 23:26:13.942572 containerd[1594]: time="2026-04-17T23:26:13.942323799Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.11\" returns image reference \"sha256:8c8e25fd00e5c108fb9ab5490c25bfaeb0231b1c59f749dab4f5300f1c49995b\"" Apr 17 23:26:13.943236 containerd[1594]: time="2026-04-17T23:26:13.942957396Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\"" Apr 17 23:26:14.816915 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount422151215.mount: Deactivated successfully. Apr 17 23:26:15.154729 containerd[1594]: time="2026-04-17T23:26:15.153239290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:15.154729 containerd[1594]: time="2026-04-17T23:26:15.154580846Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.11: active requests=0, bytes read=28040534" Apr 17 23:26:15.155472 containerd[1594]: time="2026-04-17T23:26:15.155427767Z" level=info msg="ImageCreate event name:\"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:15.159773 containerd[1594]: time="2026-04-17T23:26:15.159728555Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:15.160462 containerd[1594]: time="2026-04-17T23:26:15.160424829Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.11\" with image id \"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\", repo tag \"registry.k8s.io/kube-proxy:v1.33.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:8d18637b5c5f58a4ca0163d3cf184e53d4c522963c242860562be7cb25e9303e\", size \"28039527\" in 1.217425199s" Apr 17 23:26:15.160462 containerd[1594]: time="2026-04-17T23:26:15.160460586Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.11\" returns image reference \"sha256:7ce14d6fb1e5134a578d2aaa327fd701273e3d222b9b8d88054dd86b87a7dc36\"" Apr 17 23:26:15.160955 containerd[1594]: time="2026-04-17T23:26:15.160930287Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Apr 17 23:26:15.411848 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 17 23:26:15.418481 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:15.638395 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:15.644246 (kubelet)[2042]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 17 23:26:15.690429 kubelet[2042]: E0417 23:26:15.688561 2042 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 17 23:26:15.692821 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 17 23:26:15.693079 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 17 23:26:15.784046 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3859133439.mount: Deactivated successfully. Apr 17 23:26:16.754934 containerd[1594]: time="2026-04-17T23:26:16.754863898Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:16.757322 containerd[1594]: time="2026-04-17T23:26:16.757277008Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152209" Apr 17 23:26:16.757426 containerd[1594]: time="2026-04-17T23:26:16.757391747Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:16.762559 containerd[1594]: time="2026-04-17T23:26:16.762491769Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:16.763926 containerd[1594]: time="2026-04-17T23:26:16.763746596Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 1.602702985s" Apr 17 23:26:16.763926 containerd[1594]: time="2026-04-17T23:26:16.763789959Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Apr 17 23:26:16.764294 containerd[1594]: time="2026-04-17T23:26:16.764257728Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 17 23:26:17.209775 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2196060529.mount: Deactivated successfully. Apr 17 23:26:17.216237 containerd[1594]: time="2026-04-17T23:26:17.215925643Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:17.217001 containerd[1594]: time="2026-04-17T23:26:17.216962770Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268723" Apr 17 23:26:17.218106 containerd[1594]: time="2026-04-17T23:26:17.217806574Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:17.220714 containerd[1594]: time="2026-04-17T23:26:17.220370764Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:17.221495 containerd[1594]: time="2026-04-17T23:26:17.221337093Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 456.94968ms" Apr 17 23:26:17.221495 containerd[1594]: time="2026-04-17T23:26:17.221376318Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Apr 17 23:26:17.222047 containerd[1594]: time="2026-04-17T23:26:17.221805753Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Apr 17 23:26:17.707987 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount773034557.mount: Deactivated successfully. Apr 17 23:26:18.603208 containerd[1594]: time="2026-04-17T23:26:18.603126391Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:18.629233 containerd[1594]: time="2026-04-17T23:26:18.628518071Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=21886470" Apr 17 23:26:18.630971 containerd[1594]: time="2026-04-17T23:26:18.630857920Z" level=info msg="ImageCreate event name:\"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:18.636222 containerd[1594]: time="2026-04-17T23:26:18.636020283Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:18.637600 containerd[1594]: time="2026-04-17T23:26:18.637380104Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"21882972\" in 1.415536531s" Apr 17 23:26:18.637600 containerd[1594]: time="2026-04-17T23:26:18.637426211Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\"" Apr 17 23:26:23.283592 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:23.290585 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:23.338915 systemd[1]: Reloading requested from client PID 2198 ('systemctl') (unit session-7.scope)... Apr 17 23:26:23.338941 systemd[1]: Reloading... Apr 17 23:26:23.450219 zram_generator::config[2242]: No configuration found. Apr 17 23:26:23.558154 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 17 23:26:23.628147 systemd[1]: Reloading finished in 288 ms. Apr 17 23:26:23.679427 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 17 23:26:23.679662 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 17 23:26:23.680168 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:23.693944 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:23.827503 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:23.842613 (kubelet)[2298]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 17 23:26:23.886652 kubelet[2298]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 23:26:23.886652 kubelet[2298]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 17 23:26:23.886652 kubelet[2298]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 23:26:23.886652 kubelet[2298]: I0417 23:26:23.886383 2298 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 17 23:26:25.163931 kubelet[2298]: I0417 23:26:25.163778 2298 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 17 23:26:25.163931 kubelet[2298]: I0417 23:26:25.163847 2298 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 17 23:26:25.165137 kubelet[2298]: I0417 23:26:25.165114 2298 server.go:956] "Client rotation is on, will bootstrap in background" Apr 17 23:26:25.193000 kubelet[2298]: E0417 23:26:25.192919 2298 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://46.224.37.37:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 17 23:26:25.193189 kubelet[2298]: I0417 23:26:25.193114 2298 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 17 23:26:25.201541 kubelet[2298]: E0417 23:26:25.201414 2298 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 17 23:26:25.201541 kubelet[2298]: I0417 23:26:25.201443 2298 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 17 23:26:25.204656 kubelet[2298]: I0417 23:26:25.204630 2298 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 17 23:26:25.205929 kubelet[2298]: I0417 23:26:25.205856 2298 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 17 23:26:25.206088 kubelet[2298]: I0417 23:26:25.205904 2298 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-8954fb58c4","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Apr 17 23:26:25.206088 kubelet[2298]: I0417 23:26:25.206090 2298 topology_manager.go:138] "Creating topology manager with none policy" Apr 17 23:26:25.206272 kubelet[2298]: I0417 23:26:25.206102 2298 container_manager_linux.go:303] "Creating device plugin manager" Apr 17 23:26:25.206334 kubelet[2298]: I0417 23:26:25.206316 2298 state_mem.go:36] "Initialized new in-memory state store" Apr 17 23:26:25.210009 kubelet[2298]: I0417 23:26:25.209961 2298 kubelet.go:480] "Attempting to sync node with API server" Apr 17 23:26:25.210152 kubelet[2298]: I0417 23:26:25.210117 2298 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 17 23:26:25.210244 kubelet[2298]: I0417 23:26:25.210162 2298 kubelet.go:386] "Adding apiserver pod source" Apr 17 23:26:25.210244 kubelet[2298]: I0417 23:26:25.210188 2298 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 17 23:26:25.218854 kubelet[2298]: E0417 23:26:25.218141 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://46.224.37.37:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-8954fb58c4&limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 17 23:26:25.220240 kubelet[2298]: E0417 23:26:25.220201 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://46.224.37.37:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 17 23:26:25.220797 kubelet[2298]: I0417 23:26:25.220775 2298 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 17 23:26:25.221770 kubelet[2298]: I0417 23:26:25.221748 2298 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 17 23:26:25.221998 kubelet[2298]: W0417 23:26:25.221986 2298 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 17 23:26:25.228755 kubelet[2298]: I0417 23:26:25.228647 2298 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 17 23:26:25.228755 kubelet[2298]: I0417 23:26:25.228705 2298 server.go:1289] "Started kubelet" Apr 17 23:26:25.232542 kubelet[2298]: I0417 23:26:25.232481 2298 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 17 23:26:25.236141 kubelet[2298]: E0417 23:26:25.234793 2298 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://46.224.37.37:6443/api/v1/namespaces/default/events\": dial tcp 46.224.37.37:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-6-n-8954fb58c4.18a748959044522e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-6-n-8954fb58c4,UID:ci-4081-3-6-n-8954fb58c4,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-8954fb58c4,},FirstTimestamp:2026-04-17 23:26:25.228673582 +0000 UTC m=+1.381793984,LastTimestamp:2026-04-17 23:26:25.228673582 +0000 UTC m=+1.381793984,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-8954fb58c4,}" Apr 17 23:26:25.238114 kubelet[2298]: I0417 23:26:25.238034 2298 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 17 23:26:25.239269 kubelet[2298]: I0417 23:26:25.239136 2298 server.go:317] "Adding debug handlers to kubelet server" Apr 17 23:26:25.242781 kubelet[2298]: I0417 23:26:25.242701 2298 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 17 23:26:25.243334 kubelet[2298]: I0417 23:26:25.242968 2298 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 17 23:26:25.243334 kubelet[2298]: I0417 23:26:25.243267 2298 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 17 23:26:25.245825 kubelet[2298]: I0417 23:26:25.245792 2298 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 17 23:26:25.246222 kubelet[2298]: E0417 23:26:25.246190 2298 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" Apr 17 23:26:25.246858 kubelet[2298]: I0417 23:26:25.246818 2298 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 17 23:26:25.247486 kubelet[2298]: I0417 23:26:25.246954 2298 reconciler.go:26] "Reconciler: start to sync state" Apr 17 23:26:25.247486 kubelet[2298]: E0417 23:26:25.247334 2298 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://46.224.37.37:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-8954fb58c4?timeout=10s\": dial tcp 46.224.37.37:6443: connect: connection refused" interval="200ms" Apr 17 23:26:25.247922 kubelet[2298]: E0417 23:26:25.247884 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://46.224.37.37:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 17 23:26:25.248864 kubelet[2298]: I0417 23:26:25.248818 2298 factory.go:223] Registration of the systemd container factory successfully Apr 17 23:26:25.249536 kubelet[2298]: I0417 23:26:25.249497 2298 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 17 23:26:25.251303 kubelet[2298]: E0417 23:26:25.251253 2298 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 17 23:26:25.252525 kubelet[2298]: I0417 23:26:25.251549 2298 factory.go:223] Registration of the containerd container factory successfully Apr 17 23:26:25.272602 kubelet[2298]: I0417 23:26:25.272564 2298 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 17 23:26:25.274099 kubelet[2298]: I0417 23:26:25.274036 2298 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 17 23:26:25.274099 kubelet[2298]: I0417 23:26:25.274079 2298 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 17 23:26:25.274099 kubelet[2298]: I0417 23:26:25.274099 2298 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 17 23:26:25.274306 kubelet[2298]: I0417 23:26:25.274142 2298 kubelet.go:2436] "Starting kubelet main sync loop" Apr 17 23:26:25.274306 kubelet[2298]: E0417 23:26:25.274260 2298 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 17 23:26:25.278658 kubelet[2298]: I0417 23:26:25.278605 2298 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 17 23:26:25.278658 kubelet[2298]: I0417 23:26:25.278636 2298 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 17 23:26:25.278658 kubelet[2298]: I0417 23:26:25.278659 2298 state_mem.go:36] "Initialized new in-memory state store" Apr 17 23:26:25.278993 kubelet[2298]: E0417 23:26:25.278962 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://46.224.37.37:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 17 23:26:25.281329 kubelet[2298]: I0417 23:26:25.281286 2298 policy_none.go:49] "None policy: Start" Apr 17 23:26:25.281329 kubelet[2298]: I0417 23:26:25.281324 2298 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 17 23:26:25.281329 kubelet[2298]: I0417 23:26:25.281337 2298 state_mem.go:35] "Initializing new in-memory state store" Apr 17 23:26:25.286653 kubelet[2298]: E0417 23:26:25.286600 2298 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 17 23:26:25.286886 kubelet[2298]: I0417 23:26:25.286868 2298 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 17 23:26:25.286955 kubelet[2298]: I0417 23:26:25.286885 2298 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 17 23:26:25.289224 kubelet[2298]: I0417 23:26:25.289170 2298 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 17 23:26:25.296359 kubelet[2298]: E0417 23:26:25.294561 2298 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 17 23:26:25.296359 kubelet[2298]: E0417 23:26:25.294636 2298 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-6-n-8954fb58c4\" not found" Apr 17 23:26:25.383429 kubelet[2298]: E0417 23:26:25.383376 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.391839 kubelet[2298]: E0417 23:26:25.391801 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.392836 kubelet[2298]: I0417 23:26:25.392814 2298 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.395265 kubelet[2298]: E0417 23:26:25.395243 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.395766 kubelet[2298]: E0417 23:26:25.395742 2298 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://46.224.37.37:6443/api/v1/nodes\": dial tcp 46.224.37.37:6443: connect: connection refused" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.448319 kubelet[2298]: I0417 23:26:25.448109 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.450161 kubelet[2298]: E0417 23:26:25.450105 2298 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://46.224.37.37:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-8954fb58c4?timeout=10s\": dial tcp 46.224.37.37:6443: connect: connection refused" interval="400ms" Apr 17 23:26:25.549046 kubelet[2298]: I0417 23:26:25.548946 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549046 kubelet[2298]: I0417 23:26:25.549005 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549046 kubelet[2298]: I0417 23:26:25.549035 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f58eb785934bb7933bad54cca6033cc4-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-8954fb58c4\" (UID: \"f58eb785934bb7933bad54cca6033cc4\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549356 kubelet[2298]: I0417 23:26:25.549097 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549356 kubelet[2298]: I0417 23:26:25.549119 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549356 kubelet[2298]: I0417 23:26:25.549139 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549356 kubelet[2298]: I0417 23:26:25.549160 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.549356 kubelet[2298]: I0417 23:26:25.549194 2298 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.599156 kubelet[2298]: I0417 23:26:25.599103 2298 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.599690 kubelet[2298]: E0417 23:26:25.599645 2298 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://46.224.37.37:6443/api/v1/nodes\": dial tcp 46.224.37.37:6443: connect: connection refused" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:25.686491 containerd[1594]: time="2026-04-17T23:26:25.686024442Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-8954fb58c4,Uid:c6eb327a67cbb76d10da0426ea735d46,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:25.693219 containerd[1594]: time="2026-04-17T23:26:25.693131193Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-8954fb58c4,Uid:2d52f0823cfa50fca4ab14c7ab5edabf,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:25.697268 containerd[1594]: time="2026-04-17T23:26:25.696901112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-8954fb58c4,Uid:f58eb785934bb7933bad54cca6033cc4,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:25.851366 kubelet[2298]: E0417 23:26:25.851208 2298 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://46.224.37.37:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-6-n-8954fb58c4?timeout=10s\": dial tcp 46.224.37.37:6443: connect: connection refused" interval="800ms" Apr 17 23:26:26.002643 kubelet[2298]: I0417 23:26:26.002599 2298 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:26.003148 kubelet[2298]: E0417 23:26:26.003076 2298 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://46.224.37.37:6443/api/v1/nodes\": dial tcp 46.224.37.37:6443: connect: connection refused" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:26.124325 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1311549103.mount: Deactivated successfully. Apr 17 23:26:26.131746 containerd[1594]: time="2026-04-17T23:26:26.131674544Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 17 23:26:26.133221 containerd[1594]: time="2026-04-17T23:26:26.133182660Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Apr 17 23:26:26.137360 containerd[1594]: time="2026-04-17T23:26:26.137314611Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 17 23:26:26.139364 containerd[1594]: time="2026-04-17T23:26:26.139266789Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 17 23:26:26.139930 containerd[1594]: time="2026-04-17T23:26:26.139744749Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 17 23:26:26.142779 containerd[1594]: time="2026-04-17T23:26:26.142224952Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 17 23:26:26.142779 containerd[1594]: time="2026-04-17T23:26:26.142323761Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 17 23:26:26.144934 containerd[1594]: time="2026-04-17T23:26:26.144413128Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 17 23:26:26.145583 containerd[1594]: time="2026-04-17T23:26:26.145547497Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 459.411271ms" Apr 17 23:26:26.149105 containerd[1594]: time="2026-04-17T23:26:26.149034044Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 452.025351ms" Apr 17 23:26:26.152667 containerd[1594]: time="2026-04-17T23:26:26.152616560Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 459.377746ms" Apr 17 23:26:26.243923 kubelet[2298]: E0417 23:26:26.243865 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://46.224.37.37:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-6-n-8954fb58c4&limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 17 23:26:26.264021 kubelet[2298]: E0417 23:26:26.263885 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://46.224.37.37:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Apr 17 23:26:26.271390 containerd[1594]: time="2026-04-17T23:26:26.271263063Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:26.271520 containerd[1594]: time="2026-04-17T23:26:26.271363113Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:26.271520 containerd[1594]: time="2026-04-17T23:26:26.271377240Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.273372 containerd[1594]: time="2026-04-17T23:26:26.273312770Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.275533 containerd[1594]: time="2026-04-17T23:26:26.275208280Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:26.275533 containerd[1594]: time="2026-04-17T23:26:26.275297084Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:26.275533 containerd[1594]: time="2026-04-17T23:26:26.275314293Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.275533 containerd[1594]: time="2026-04-17T23:26:26.275416704Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.277419 containerd[1594]: time="2026-04-17T23:26:26.277301329Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:26.277419 containerd[1594]: time="2026-04-17T23:26:26.277375886Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:26.277642 containerd[1594]: time="2026-04-17T23:26:26.277392054Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.279379 containerd[1594]: time="2026-04-17T23:26:26.279222171Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:26.324976 kubelet[2298]: E0417 23:26:26.324440 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://46.224.37.37:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 17 23:26:26.353432 containerd[1594]: time="2026-04-17T23:26:26.353388182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-6-n-8954fb58c4,Uid:2d52f0823cfa50fca4ab14c7ab5edabf,Namespace:kube-system,Attempt:0,} returns sandbox id \"dca3a8ba705d3cf59da649c4b608edf73517c9030022b715a26c8a6d086a2fed\"" Apr 17 23:26:26.359868 containerd[1594]: time="2026-04-17T23:26:26.359689700Z" level=info msg="CreateContainer within sandbox \"dca3a8ba705d3cf59da649c4b608edf73517c9030022b715a26c8a6d086a2fed\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 17 23:26:26.365530 containerd[1594]: time="2026-04-17T23:26:26.365431818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-6-n-8954fb58c4,Uid:f58eb785934bb7933bad54cca6033cc4,Namespace:kube-system,Attempt:0,} returns sandbox id \"ce55e4a2842f0c78d57831dfe00541366bea6bcd37e6108fd7f3fd4103c7a975\"" Apr 17 23:26:26.366774 kubelet[2298]: E0417 23:26:26.366167 2298 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://46.224.37.37:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 46.224.37.37:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 17 23:26:26.370869 containerd[1594]: time="2026-04-17T23:26:26.370761649Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-6-n-8954fb58c4,Uid:c6eb327a67cbb76d10da0426ea735d46,Namespace:kube-system,Attempt:0,} returns sandbox id \"281786496d3b4f60ab4597e312b1690f0a67e64c7c158bd0d09e921d0a977158\"" Apr 17 23:26:26.373259 containerd[1594]: time="2026-04-17T23:26:26.372294577Z" level=info msg="CreateContainer within sandbox \"ce55e4a2842f0c78d57831dfe00541366bea6bcd37e6108fd7f3fd4103c7a975\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 17 23:26:26.375713 containerd[1594]: time="2026-04-17T23:26:26.375600994Z" level=info msg="CreateContainer within sandbox \"281786496d3b4f60ab4597e312b1690f0a67e64c7c158bd0d09e921d0a977158\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 17 23:26:26.388167 containerd[1594]: time="2026-04-17T23:26:26.388110424Z" level=info msg="CreateContainer within sandbox \"dca3a8ba705d3cf59da649c4b608edf73517c9030022b715a26c8a6d086a2fed\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75\"" Apr 17 23:26:26.389278 containerd[1594]: time="2026-04-17T23:26:26.389124492Z" level=info msg="StartContainer for \"40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75\"" Apr 17 23:26:26.393204 containerd[1594]: time="2026-04-17T23:26:26.393118454Z" level=info msg="CreateContainer within sandbox \"ce55e4a2842f0c78d57831dfe00541366bea6bcd37e6108fd7f3fd4103c7a975\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351\"" Apr 17 23:26:26.394281 containerd[1594]: time="2026-04-17T23:26:26.393772862Z" level=info msg="StartContainer for \"29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351\"" Apr 17 23:26:26.396967 containerd[1594]: time="2026-04-17T23:26:26.396927723Z" level=info msg="CreateContainer within sandbox \"281786496d3b4f60ab4597e312b1690f0a67e64c7c158bd0d09e921d0a977158\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"92115b5e4b80b9622f708c9460b4a4b662104cffd96c7243067ae03ae3c41c70\"" Apr 17 23:26:26.397551 containerd[1594]: time="2026-04-17T23:26:26.397520220Z" level=info msg="StartContainer for \"92115b5e4b80b9622f708c9460b4a4b662104cffd96c7243067ae03ae3c41c70\"" Apr 17 23:26:26.488012 containerd[1594]: time="2026-04-17T23:26:26.487439165Z" level=info msg="StartContainer for \"40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75\" returns successfully" Apr 17 23:26:26.496338 containerd[1594]: time="2026-04-17T23:26:26.495729440Z" level=info msg="StartContainer for \"92115b5e4b80b9622f708c9460b4a4b662104cffd96c7243067ae03ae3c41c70\" returns successfully" Apr 17 23:26:26.517363 containerd[1594]: time="2026-04-17T23:26:26.517305814Z" level=info msg="StartContainer for \"29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351\" returns successfully" Apr 17 23:26:26.805647 kubelet[2298]: I0417 23:26:26.805599 2298 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:27.293628 kubelet[2298]: E0417 23:26:27.293238 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:27.298186 kubelet[2298]: E0417 23:26:27.296790 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:27.301199 kubelet[2298]: E0417 23:26:27.300265 2298 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:27.977196 kubelet[2298]: E0417 23:26:27.977135 2298 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-3-6-n-8954fb58c4\" not found" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.052203 kubelet[2298]: I0417 23:26:28.050087 2298 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.052203 kubelet[2298]: E0417 23:26:28.050126 2298 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4081-3-6-n-8954fb58c4\": node \"ci-4081-3-6-n-8954fb58c4\" not found" Apr 17 23:26:28.093999 kubelet[2298]: E0417 23:26:28.093771 2298 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" Apr 17 23:26:28.147045 kubelet[2298]: I0417 23:26:28.146830 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.162230 kubelet[2298]: E0417 23:26:28.161078 2298 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.162230 kubelet[2298]: I0417 23:26:28.161112 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.165228 kubelet[2298]: E0417 23:26:28.165037 2298 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.165228 kubelet[2298]: I0417 23:26:28.165071 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.168396 kubelet[2298]: E0417 23:26:28.168354 2298 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-8954fb58c4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.219275 kubelet[2298]: I0417 23:26:28.219212 2298 apiserver.go:52] "Watching apiserver" Apr 17 23:26:28.247847 kubelet[2298]: I0417 23:26:28.247661 2298 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 17 23:26:28.303260 kubelet[2298]: I0417 23:26:28.302717 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.305999 kubelet[2298]: I0417 23:26:28.305425 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.306941 kubelet[2298]: E0417 23:26:28.306758 2298 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:28.308031 kubelet[2298]: E0417 23:26:28.308007 2298 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-8954fb58c4\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:30.239675 kubelet[2298]: I0417 23:26:30.239593 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:30.301674 systemd[1]: Reloading requested from client PID 2585 ('systemctl') (unit session-7.scope)... Apr 17 23:26:30.301964 systemd[1]: Reloading... Apr 17 23:26:30.356613 kubelet[2298]: I0417 23:26:30.356572 2298 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:30.394300 zram_generator::config[2625]: No configuration found. Apr 17 23:26:30.511002 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 17 23:26:30.587500 systemd[1]: Reloading finished in 285 ms. Apr 17 23:26:30.622480 kubelet[2298]: I0417 23:26:30.622402 2298 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 17 23:26:30.622792 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:30.637423 systemd[1]: kubelet.service: Deactivated successfully. Apr 17 23:26:30.637964 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:30.649754 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 17 23:26:30.780363 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 17 23:26:30.783317 (kubelet)[2680]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 17 23:26:30.836209 kubelet[2680]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 23:26:30.836209 kubelet[2680]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 17 23:26:30.836209 kubelet[2680]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 23:26:30.836209 kubelet[2680]: I0417 23:26:30.835442 2680 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 17 23:26:30.850264 kubelet[2680]: I0417 23:26:30.849093 2680 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Apr 17 23:26:30.850750 kubelet[2680]: I0417 23:26:30.850728 2680 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 17 23:26:30.851003 kubelet[2680]: I0417 23:26:30.850989 2680 server.go:956] "Client rotation is on, will bootstrap in background" Apr 17 23:26:30.852376 kubelet[2680]: I0417 23:26:30.852349 2680 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 17 23:26:30.855092 kubelet[2680]: I0417 23:26:30.855060 2680 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 17 23:26:30.859919 kubelet[2680]: E0417 23:26:30.859703 2680 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 17 23:26:30.859919 kubelet[2680]: I0417 23:26:30.859731 2680 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 17 23:26:30.862489 kubelet[2680]: I0417 23:26:30.862445 2680 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 17 23:26:30.864236 kubelet[2680]: I0417 23:26:30.863277 2680 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 17 23:26:30.864236 kubelet[2680]: I0417 23:26:30.863308 2680 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-6-n-8954fb58c4","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Apr 17 23:26:30.864236 kubelet[2680]: I0417 23:26:30.863543 2680 topology_manager.go:138] "Creating topology manager with none policy" Apr 17 23:26:30.864236 kubelet[2680]: I0417 23:26:30.863552 2680 container_manager_linux.go:303] "Creating device plugin manager" Apr 17 23:26:30.864236 kubelet[2680]: I0417 23:26:30.863605 2680 state_mem.go:36] "Initialized new in-memory state store" Apr 17 23:26:30.864608 kubelet[2680]: I0417 23:26:30.863743 2680 kubelet.go:480] "Attempting to sync node with API server" Apr 17 23:26:30.864608 kubelet[2680]: I0417 23:26:30.863753 2680 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 17 23:26:30.864608 kubelet[2680]: I0417 23:26:30.863776 2680 kubelet.go:386] "Adding apiserver pod source" Apr 17 23:26:30.864608 kubelet[2680]: I0417 23:26:30.863785 2680 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 17 23:26:30.872465 kubelet[2680]: I0417 23:26:30.872039 2680 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 17 23:26:30.872720 kubelet[2680]: I0417 23:26:30.872678 2680 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 17 23:26:30.877041 kubelet[2680]: I0417 23:26:30.876957 2680 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 17 23:26:30.877041 kubelet[2680]: I0417 23:26:30.877001 2680 server.go:1289] "Started kubelet" Apr 17 23:26:30.884428 kubelet[2680]: I0417 23:26:30.882988 2680 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 17 23:26:30.884428 kubelet[2680]: I0417 23:26:30.883836 2680 server.go:317] "Adding debug handlers to kubelet server" Apr 17 23:26:30.887748 kubelet[2680]: I0417 23:26:30.885161 2680 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 17 23:26:30.887748 kubelet[2680]: I0417 23:26:30.887574 2680 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 17 23:26:30.887883 kubelet[2680]: I0417 23:26:30.887770 2680 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 17 23:26:30.889259 kubelet[2680]: I0417 23:26:30.888552 2680 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 17 23:26:30.892208 kubelet[2680]: I0417 23:26:30.890707 2680 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 17 23:26:30.892208 kubelet[2680]: I0417 23:26:30.890829 2680 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 17 23:26:30.892208 kubelet[2680]: I0417 23:26:30.890952 2680 reconciler.go:26] "Reconciler: start to sync state" Apr 17 23:26:30.892208 kubelet[2680]: E0417 23:26:30.891597 2680 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081-3-6-n-8954fb58c4\" not found" Apr 17 23:26:30.924192 kubelet[2680]: I0417 23:26:30.922057 2680 factory.go:223] Registration of the systemd container factory successfully Apr 17 23:26:30.924192 kubelet[2680]: I0417 23:26:30.922213 2680 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 17 23:26:30.930978 kubelet[2680]: I0417 23:26:30.930952 2680 factory.go:223] Registration of the containerd container factory successfully Apr 17 23:26:30.937009 kubelet[2680]: I0417 23:26:30.936947 2680 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 17 23:26:30.938052 kubelet[2680]: I0417 23:26:30.938017 2680 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 17 23:26:30.938052 kubelet[2680]: I0417 23:26:30.938058 2680 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 17 23:26:30.938164 kubelet[2680]: I0417 23:26:30.938081 2680 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 17 23:26:30.938164 kubelet[2680]: I0417 23:26:30.938089 2680 kubelet.go:2436] "Starting kubelet main sync loop" Apr 17 23:26:30.938164 kubelet[2680]: E0417 23:26:30.938134 2680 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 17 23:26:30.997695 kubelet[2680]: I0417 23:26:30.997668 2680 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 17 23:26:30.997983 kubelet[2680]: I0417 23:26:30.997942 2680 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 17 23:26:30.998053 kubelet[2680]: I0417 23:26:30.998045 2680 state_mem.go:36] "Initialized new in-memory state store" Apr 17 23:26:30.998263 kubelet[2680]: I0417 23:26:30.998248 2680 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 17 23:26:30.998356 kubelet[2680]: I0417 23:26:30.998333 2680 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 17 23:26:30.998405 kubelet[2680]: I0417 23:26:30.998398 2680 policy_none.go:49] "None policy: Start" Apr 17 23:26:30.998505 kubelet[2680]: I0417 23:26:30.998495 2680 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 17 23:26:30.998558 kubelet[2680]: I0417 23:26:30.998551 2680 state_mem.go:35] "Initializing new in-memory state store" Apr 17 23:26:30.998707 kubelet[2680]: I0417 23:26:30.998695 2680 state_mem.go:75] "Updated machine memory state" Apr 17 23:26:31.000013 kubelet[2680]: E0417 23:26:30.999991 2680 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 17 23:26:31.000306 kubelet[2680]: I0417 23:26:31.000290 2680 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 17 23:26:31.000407 kubelet[2680]: I0417 23:26:31.000379 2680 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 17 23:26:31.002321 kubelet[2680]: I0417 23:26:31.002289 2680 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 17 23:26:31.004553 kubelet[2680]: E0417 23:26:31.004518 2680 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 17 23:26:31.040680 kubelet[2680]: I0417 23:26:31.040104 2680 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.040680 kubelet[2680]: I0417 23:26:31.040158 2680 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.041308 kubelet[2680]: I0417 23:26:31.041285 2680 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.055573 kubelet[2680]: E0417 23:26:31.055531 2680 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.056528 kubelet[2680]: E0417 23:26:31.056385 2680 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" already exists" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092106 kubelet[2680]: I0417 23:26:31.091963 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-k8s-certs\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092106 kubelet[2680]: I0417 23:26:31.092056 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-ca-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092399 kubelet[2680]: I0417 23:26:31.092163 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092399 kubelet[2680]: I0417 23:26:31.092321 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092399 kubelet[2680]: I0417 23:26:31.092388 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f58eb785934bb7933bad54cca6033cc4-kubeconfig\") pod \"kube-scheduler-ci-4081-3-6-n-8954fb58c4\" (UID: \"f58eb785934bb7933bad54cca6033cc4\") " pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092905 kubelet[2680]: I0417 23:26:31.092431 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-ca-certs\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092905 kubelet[2680]: I0417 23:26:31.092517 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c6eb327a67cbb76d10da0426ea735d46-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" (UID: \"c6eb327a67cbb76d10da0426ea735d46\") " pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092905 kubelet[2680]: I0417 23:26:31.092590 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.092905 kubelet[2680]: I0417 23:26:31.092662 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2d52f0823cfa50fca4ab14c7ab5edabf-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-6-n-8954fb58c4\" (UID: \"2d52f0823cfa50fca4ab14c7ab5edabf\") " pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.112369 kubelet[2680]: I0417 23:26:31.112328 2680 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.125750 kubelet[2680]: I0417 23:26:31.125685 2680 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.125885 kubelet[2680]: I0417 23:26:31.125778 2680 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.299546 sudo[2715]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Apr 17 23:26:31.299836 sudo[2715]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Apr 17 23:26:31.810341 sudo[2715]: pam_unix(sudo:session): session closed for user root Apr 17 23:26:31.874388 kubelet[2680]: I0417 23:26:31.874337 2680 apiserver.go:52] "Watching apiserver" Apr 17 23:26:31.893757 kubelet[2680]: I0417 23:26:31.892828 2680 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 17 23:26:31.969459 kubelet[2680]: I0417 23:26:31.969418 2680 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.969781 kubelet[2680]: I0417 23:26:31.969758 2680 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.979090 kubelet[2680]: E0417 23:26:31.978788 2680 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-6-n-8954fb58c4\" already exists" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.979090 kubelet[2680]: E0417 23:26:31.978986 2680 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-6-n-8954fb58c4\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" Apr 17 23:26:31.998218 kubelet[2680]: I0417 23:26:31.998130 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-6-n-8954fb58c4" podStartSLOduration=1.996631485 podStartE2EDuration="1.996631485s" podCreationTimestamp="2026-04-17 23:26:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:26:31.996270832 +0000 UTC m=+1.207265205" watchObservedRunningTime="2026-04-17 23:26:31.996631485 +0000 UTC m=+1.207625858" Apr 17 23:26:32.019440 kubelet[2680]: I0417 23:26:32.018770 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-6-n-8954fb58c4" podStartSLOduration=1.018744766 podStartE2EDuration="1.018744766s" podCreationTimestamp="2026-04-17 23:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:26:32.006813722 +0000 UTC m=+1.217808055" watchObservedRunningTime="2026-04-17 23:26:32.018744766 +0000 UTC m=+1.229739099" Apr 17 23:26:33.499468 sudo[1803]: pam_unix(sudo:session): session closed for user root Apr 17 23:26:33.516387 sshd[1799]: pam_unix(sshd:session): session closed for user core Apr 17 23:26:33.523591 systemd[1]: sshd@6-46.224.37.37:22-50.85.169.122:59282.service: Deactivated successfully. Apr 17 23:26:33.526775 systemd[1]: session-7.scope: Deactivated successfully. Apr 17 23:26:33.526843 systemd-logind[1565]: Session 7 logged out. Waiting for processes to exit. Apr 17 23:26:33.531145 systemd-logind[1565]: Removed session 7. Apr 17 23:26:36.355298 kubelet[2680]: I0417 23:26:36.355229 2680 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 17 23:26:36.355933 containerd[1594]: time="2026-04-17T23:26:36.355825271Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 17 23:26:36.356374 kubelet[2680]: I0417 23:26:36.356135 2680 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 17 23:26:36.792471 kubelet[2680]: I0417 23:26:36.791591 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-6-n-8954fb58c4" podStartSLOduration=6.791574031 podStartE2EDuration="6.791574031s" podCreationTimestamp="2026-04-17 23:26:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:26:32.020216978 +0000 UTC m=+1.231211351" watchObservedRunningTime="2026-04-17 23:26:36.791574031 +0000 UTC m=+6.002568364" Apr 17 23:26:36.833986 kubelet[2680]: I0417 23:26:36.833944 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/5a57875e-512f-48a6-ba4b-89da0436f9bd-kube-proxy\") pod \"kube-proxy-6g7l2\" (UID: \"5a57875e-512f-48a6-ba4b-89da0436f9bd\") " pod="kube-system/kube-proxy-6g7l2" Apr 17 23:26:36.833986 kubelet[2680]: I0417 23:26:36.833985 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a57875e-512f-48a6-ba4b-89da0436f9bd-lib-modules\") pod \"kube-proxy-6g7l2\" (UID: \"5a57875e-512f-48a6-ba4b-89da0436f9bd\") " pod="kube-system/kube-proxy-6g7l2" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834008 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x755\" (UniqueName: \"kubernetes.io/projected/5a57875e-512f-48a6-ba4b-89da0436f9bd-kube-api-access-6x755\") pod \"kube-proxy-6g7l2\" (UID: \"5a57875e-512f-48a6-ba4b-89da0436f9bd\") " pod="kube-system/kube-proxy-6g7l2" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834028 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-etc-cni-netd\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834043 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-xtables-lock\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834057 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hubble-tls\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834071 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-run\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834150 kubelet[2680]: I0417 23:26:36.834085 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-bpf-maps\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834099 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hostproc\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834113 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-config-path\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834126 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-kernel\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834140 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59zzq\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834160 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5a57875e-512f-48a6-ba4b-89da0436f9bd-xtables-lock\") pod \"kube-proxy-6g7l2\" (UID: \"5a57875e-512f-48a6-ba4b-89da0436f9bd\") " pod="kube-system/kube-proxy-6g7l2" Apr 17 23:26:36.834330 kubelet[2680]: I0417 23:26:36.834193 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cni-path\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834452 kubelet[2680]: I0417 23:26:36.834208 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-lib-modules\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834452 kubelet[2680]: I0417 23:26:36.834263 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-clustermesh-secrets\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834452 kubelet[2680]: I0417 23:26:36.834285 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-cgroup\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.834452 kubelet[2680]: I0417 23:26:36.834301 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-net\") pod \"cilium-c7nff\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " pod="kube-system/cilium-c7nff" Apr 17 23:26:36.953516 kubelet[2680]: E0417 23:26:36.953469 2680 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Apr 17 23:26:36.953516 kubelet[2680]: E0417 23:26:36.953502 2680 projected.go:194] Error preparing data for projected volume kube-api-access-59zzq for pod kube-system/cilium-c7nff: configmap "kube-root-ca.crt" not found Apr 17 23:26:36.953764 kubelet[2680]: E0417 23:26:36.953556 2680 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq podName:e561372a-e5ae-4fc2-ab53-44127aa5ca4a nodeName:}" failed. No retries permitted until 2026-04-17 23:26:37.453535774 +0000 UTC m=+6.664530107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-59zzq" (UniqueName: "kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq") pod "cilium-c7nff" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a") : configmap "kube-root-ca.crt" not found Apr 17 23:26:36.953828 kubelet[2680]: E0417 23:26:36.953776 2680 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Apr 17 23:26:36.953828 kubelet[2680]: E0417 23:26:36.953790 2680 projected.go:194] Error preparing data for projected volume kube-api-access-6x755 for pod kube-system/kube-proxy-6g7l2: configmap "kube-root-ca.crt" not found Apr 17 23:26:36.953828 kubelet[2680]: E0417 23:26:36.953819 2680 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5a57875e-512f-48a6-ba4b-89da0436f9bd-kube-api-access-6x755 podName:5a57875e-512f-48a6-ba4b-89da0436f9bd nodeName:}" failed. No retries permitted until 2026-04-17 23:26:37.453808724 +0000 UTC m=+6.664803057 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6x755" (UniqueName: "kubernetes.io/projected/5a57875e-512f-48a6-ba4b-89da0436f9bd-kube-api-access-6x755") pod "kube-proxy-6g7l2" (UID: "5a57875e-512f-48a6-ba4b-89da0436f9bd") : configmap "kube-root-ca.crt" not found Apr 17 23:26:37.541557 kubelet[2680]: I0417 23:26:37.540716 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j4f9\" (UniqueName: \"kubernetes.io/projected/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-kube-api-access-8j4f9\") pod \"cilium-operator-6c4d7847fc-bpq9j\" (UID: \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\") " pod="kube-system/cilium-operator-6c4d7847fc-bpq9j" Apr 17 23:26:37.541557 kubelet[2680]: I0417 23:26:37.540758 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-bpq9j\" (UID: \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\") " pod="kube-system/cilium-operator-6c4d7847fc-bpq9j" Apr 17 23:26:37.703893 containerd[1594]: time="2026-04-17T23:26:37.703171079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6g7l2,Uid:5a57875e-512f-48a6-ba4b-89da0436f9bd,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:37.728375 containerd[1594]: time="2026-04-17T23:26:37.728337373Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c7nff,Uid:e561372a-e5ae-4fc2-ab53-44127aa5ca4a,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:37.731736 containerd[1594]: time="2026-04-17T23:26:37.731446396Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:37.731736 containerd[1594]: time="2026-04-17T23:26:37.731531421Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:37.731736 containerd[1594]: time="2026-04-17T23:26:37.731549218Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.731736 containerd[1594]: time="2026-04-17T23:26:37.731657400Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.759441 containerd[1594]: time="2026-04-17T23:26:37.759261632Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:37.759441 containerd[1594]: time="2026-04-17T23:26:37.759339979Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:37.759441 containerd[1594]: time="2026-04-17T23:26:37.759374213Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.760199 containerd[1594]: time="2026-04-17T23:26:37.759682640Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.776164 containerd[1594]: time="2026-04-17T23:26:37.776115322Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6g7l2,Uid:5a57875e-512f-48a6-ba4b-89da0436f9bd,Namespace:kube-system,Attempt:0,} returns sandbox id \"7f5cca4fcb5b57791d302d9c9519da6681b9fb22206fb9b3d975a13a0652f4bb\"" Apr 17 23:26:37.785422 containerd[1594]: time="2026-04-17T23:26:37.785227668Z" level=info msg="CreateContainer within sandbox \"7f5cca4fcb5b57791d302d9c9519da6681b9fb22206fb9b3d975a13a0652f4bb\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 17 23:26:37.804585 containerd[1594]: time="2026-04-17T23:26:37.804366243Z" level=info msg="CreateContainer within sandbox \"7f5cca4fcb5b57791d302d9c9519da6681b9fb22206fb9b3d975a13a0652f4bb\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d66d3306e674b59cb38cf85c27eacb827c5ad30e6bc1c714aff5b7ced0d4dd10\"" Apr 17 23:26:37.806039 containerd[1594]: time="2026-04-17T23:26:37.805914175Z" level=info msg="StartContainer for \"d66d3306e674b59cb38cf85c27eacb827c5ad30e6bc1c714aff5b7ced0d4dd10\"" Apr 17 23:26:37.810245 containerd[1594]: time="2026-04-17T23:26:37.810143805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c7nff,Uid:e561372a-e5ae-4fc2-ab53-44127aa5ca4a,Namespace:kube-system,Attempt:0,} returns sandbox id \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\"" Apr 17 23:26:37.815195 containerd[1594]: time="2026-04-17T23:26:37.814687540Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Apr 17 23:26:37.826327 containerd[1594]: time="2026-04-17T23:26:37.826271580Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-bpq9j,Uid:1aa6e5bd-1e8e-4a18-b263-e64831e7b699,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:37.851967 containerd[1594]: time="2026-04-17T23:26:37.851702668Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:26:37.851967 containerd[1594]: time="2026-04-17T23:26:37.851758498Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:26:37.851967 containerd[1594]: time="2026-04-17T23:26:37.851770096Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.851967 containerd[1594]: time="2026-04-17T23:26:37.851849842Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:26:37.880637 containerd[1594]: time="2026-04-17T23:26:37.880598038Z" level=info msg="StartContainer for \"d66d3306e674b59cb38cf85c27eacb827c5ad30e6bc1c714aff5b7ced0d4dd10\" returns successfully" Apr 17 23:26:37.920641 containerd[1594]: time="2026-04-17T23:26:37.920456114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-bpq9j,Uid:1aa6e5bd-1e8e-4a18-b263-e64831e7b699,Namespace:kube-system,Attempt:0,} returns sandbox id \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\"" Apr 17 23:26:40.224046 kubelet[2680]: I0417 23:26:40.223948 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-6g7l2" podStartSLOduration=4.223922894 podStartE2EDuration="4.223922894s" podCreationTimestamp="2026-04-17 23:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:26:38.004355535 +0000 UTC m=+7.215349868" watchObservedRunningTime="2026-04-17 23:26:40.223922894 +0000 UTC m=+9.434917227" Apr 17 23:26:48.918914 update_engine[1567]: I20260417 23:26:48.918004 1567 update_attempter.cc:509] Updating boot flags... Apr 17 23:26:48.964258 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (3061) Apr 17 23:26:49.062465 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (3065) Apr 17 23:26:49.103273 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 32 scanned by (udev-worker) (3065) Apr 17 23:26:52.046262 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1639101789.mount: Deactivated successfully. Apr 17 23:26:53.523751 containerd[1594]: time="2026-04-17T23:26:53.522641436Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:53.524302 containerd[1594]: time="2026-04-17T23:26:53.523801349Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Apr 17 23:26:53.525625 containerd[1594]: time="2026-04-17T23:26:53.525166048Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:53.527440 containerd[1594]: time="2026-04-17T23:26:53.527289729Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 15.712498925s" Apr 17 23:26:53.527440 containerd[1594]: time="2026-04-17T23:26:53.527338446Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Apr 17 23:26:53.529309 containerd[1594]: time="2026-04-17T23:26:53.529261542Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Apr 17 23:26:53.533156 containerd[1594]: time="2026-04-17T23:26:53.532959626Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 17 23:26:53.546002 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount244174629.mount: Deactivated successfully. Apr 17 23:26:53.555736 containerd[1594]: time="2026-04-17T23:26:53.555639375Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\"" Apr 17 23:26:53.557006 containerd[1594]: time="2026-04-17T23:26:53.556273608Z" level=info msg="StartContainer for \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\"" Apr 17 23:26:53.614669 containerd[1594]: time="2026-04-17T23:26:53.614633376Z" level=info msg="StartContainer for \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\" returns successfully" Apr 17 23:26:53.803591 containerd[1594]: time="2026-04-17T23:26:53.803118321Z" level=info msg="shim disconnected" id=a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78 namespace=k8s.io Apr 17 23:26:53.803591 containerd[1594]: time="2026-04-17T23:26:53.803488173Z" level=warning msg="cleaning up after shim disconnected" id=a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78 namespace=k8s.io Apr 17 23:26:53.803591 containerd[1594]: time="2026-04-17T23:26:53.803509932Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:26:54.029774 containerd[1594]: time="2026-04-17T23:26:54.029725683Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 17 23:26:54.042312 containerd[1594]: time="2026-04-17T23:26:54.042262311Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\"" Apr 17 23:26:54.044543 containerd[1594]: time="2026-04-17T23:26:54.043409989Z" level=info msg="StartContainer for \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\"" Apr 17 23:26:54.095307 containerd[1594]: time="2026-04-17T23:26:54.094910923Z" level=info msg="StartContainer for \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\" returns successfully" Apr 17 23:26:54.108772 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 17 23:26:54.109539 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 17 23:26:54.109792 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Apr 17 23:26:54.116532 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 17 23:26:54.148488 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 17 23:26:54.158712 containerd[1594]: time="2026-04-17T23:26:54.158634827Z" level=info msg="shim disconnected" id=04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e namespace=k8s.io Apr 17 23:26:54.158712 containerd[1594]: time="2026-04-17T23:26:54.158696943Z" level=warning msg="cleaning up after shim disconnected" id=04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e namespace=k8s.io Apr 17 23:26:54.158712 containerd[1594]: time="2026-04-17T23:26:54.158707262Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:26:54.544311 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78-rootfs.mount: Deactivated successfully. Apr 17 23:26:55.041234 containerd[1594]: time="2026-04-17T23:26:55.040627857Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 17 23:26:55.081854 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4096327375.mount: Deactivated successfully. Apr 17 23:26:55.133439 containerd[1594]: time="2026-04-17T23:26:55.133373391Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\"" Apr 17 23:26:55.134434 containerd[1594]: time="2026-04-17T23:26:55.134400681Z" level=info msg="StartContainer for \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\"" Apr 17 23:26:55.214626 containerd[1594]: time="2026-04-17T23:26:55.213599896Z" level=info msg="StartContainer for \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\" returns successfully" Apr 17 23:26:55.251436 containerd[1594]: time="2026-04-17T23:26:55.251299732Z" level=info msg="shim disconnected" id=8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1 namespace=k8s.io Apr 17 23:26:55.251436 containerd[1594]: time="2026-04-17T23:26:55.251369607Z" level=warning msg="cleaning up after shim disconnected" id=8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1 namespace=k8s.io Apr 17 23:26:55.251436 containerd[1594]: time="2026-04-17T23:26:55.251378647Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:26:55.267441 containerd[1594]: time="2026-04-17T23:26:55.267026503Z" level=warning msg="cleanup warnings time=\"2026-04-17T23:26:55Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 17 23:26:55.533213 containerd[1594]: time="2026-04-17T23:26:55.533115249Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:55.534766 containerd[1594]: time="2026-04-17T23:26:55.534720220Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Apr 17 23:26:55.536714 containerd[1594]: time="2026-04-17T23:26:55.536609851Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 17 23:26:55.538433 containerd[1594]: time="2026-04-17T23:26:55.538231221Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.008925642s" Apr 17 23:26:55.538433 containerd[1594]: time="2026-04-17T23:26:55.538272138Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Apr 17 23:26:55.550561 containerd[1594]: time="2026-04-17T23:26:55.550396074Z" level=info msg="CreateContainer within sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Apr 17 23:26:55.561834 containerd[1594]: time="2026-04-17T23:26:55.561691826Z" level=info msg="CreateContainer within sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\"" Apr 17 23:26:55.564100 containerd[1594]: time="2026-04-17T23:26:55.563389390Z" level=info msg="StartContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\"" Apr 17 23:26:55.619609 containerd[1594]: time="2026-04-17T23:26:55.619402621Z" level=info msg="StartContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" returns successfully" Apr 17 23:26:56.047415 containerd[1594]: time="2026-04-17T23:26:56.047370698Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 17 23:26:56.065638 containerd[1594]: time="2026-04-17T23:26:56.063913942Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\"" Apr 17 23:26:56.068200 containerd[1594]: time="2026-04-17T23:26:56.065955529Z" level=info msg="StartContainer for \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\"" Apr 17 23:26:56.094719 kubelet[2680]: I0417 23:26:56.092495 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-bpq9j" podStartSLOduration=1.476343976 podStartE2EDuration="19.092475365s" podCreationTimestamp="2026-04-17 23:26:37 +0000 UTC" firstStartedPulling="2026-04-17 23:26:37.923232115 +0000 UTC m=+7.134226448" lastFinishedPulling="2026-04-17 23:26:55.539363504 +0000 UTC m=+24.750357837" observedRunningTime="2026-04-17 23:26:56.090141397 +0000 UTC m=+25.301135770" watchObservedRunningTime="2026-04-17 23:26:56.092475365 +0000 UTC m=+25.303469698" Apr 17 23:26:56.211320 containerd[1594]: time="2026-04-17T23:26:56.210529929Z" level=info msg="StartContainer for \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\" returns successfully" Apr 17 23:26:56.294572 containerd[1594]: time="2026-04-17T23:26:56.294335080Z" level=info msg="shim disconnected" id=e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b namespace=k8s.io Apr 17 23:26:56.294572 containerd[1594]: time="2026-04-17T23:26:56.294411355Z" level=warning msg="cleaning up after shim disconnected" id=e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b namespace=k8s.io Apr 17 23:26:56.294572 containerd[1594]: time="2026-04-17T23:26:56.294419635Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:26:57.054497 containerd[1594]: time="2026-04-17T23:26:57.054455007Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 17 23:26:57.074126 containerd[1594]: time="2026-04-17T23:26:57.074040228Z" level=info msg="CreateContainer within sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\"" Apr 17 23:26:57.076888 containerd[1594]: time="2026-04-17T23:26:57.075397304Z" level=info msg="StartContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\"" Apr 17 23:26:57.151071 containerd[1594]: time="2026-04-17T23:26:57.151014159Z" level=info msg="StartContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" returns successfully" Apr 17 23:26:57.321853 kubelet[2680]: I0417 23:26:57.320593 2680 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Apr 17 23:26:57.369613 kubelet[2680]: I0417 23:26:57.369571 2680 status_manager.go:895] "Failed to get status for pod" podUID="0f85c524-feb4-4cd3-8152-9f76ba7da00f" pod="kube-system/coredns-674b8bbfcf-8xb5l" err="pods \"coredns-674b8bbfcf-8xb5l\" is forbidden: User \"system:node:ci-4081-3-6-n-8954fb58c4\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081-3-6-n-8954fb58c4' and this object" Apr 17 23:26:57.396469 kubelet[2680]: I0417 23:26:57.396309 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ppx\" (UniqueName: \"kubernetes.io/projected/0f85c524-feb4-4cd3-8152-9f76ba7da00f-kube-api-access-p2ppx\") pod \"coredns-674b8bbfcf-8xb5l\" (UID: \"0f85c524-feb4-4cd3-8152-9f76ba7da00f\") " pod="kube-system/coredns-674b8bbfcf-8xb5l" Apr 17 23:26:57.396469 kubelet[2680]: I0417 23:26:57.396389 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f85c524-feb4-4cd3-8152-9f76ba7da00f-config-volume\") pod \"coredns-674b8bbfcf-8xb5l\" (UID: \"0f85c524-feb4-4cd3-8152-9f76ba7da00f\") " pod="kube-system/coredns-674b8bbfcf-8xb5l" Apr 17 23:26:57.497458 kubelet[2680]: I0417 23:26:57.497418 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gvpk\" (UniqueName: \"kubernetes.io/projected/62da7207-6dff-45bd-b9ca-93fe5d21900c-kube-api-access-6gvpk\") pod \"coredns-674b8bbfcf-28f4s\" (UID: \"62da7207-6dff-45bd-b9ca-93fe5d21900c\") " pod="kube-system/coredns-674b8bbfcf-28f4s" Apr 17 23:26:57.497604 kubelet[2680]: I0417 23:26:57.497488 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62da7207-6dff-45bd-b9ca-93fe5d21900c-config-volume\") pod \"coredns-674b8bbfcf-28f4s\" (UID: \"62da7207-6dff-45bd-b9ca-93fe5d21900c\") " pod="kube-system/coredns-674b8bbfcf-28f4s" Apr 17 23:26:57.545641 systemd[1]: run-containerd-runc-k8s.io-48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7-runc.r7rwZ5.mount: Deactivated successfully. Apr 17 23:26:57.676772 containerd[1594]: time="2026-04-17T23:26:57.676717687Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-8xb5l,Uid:0f85c524-feb4-4cd3-8152-9f76ba7da00f,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:57.678401 containerd[1594]: time="2026-04-17T23:26:57.677373166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-28f4s,Uid:62da7207-6dff-45bd-b9ca-93fe5d21900c,Namespace:kube-system,Attempt:0,}" Apr 17 23:26:58.082593 kubelet[2680]: I0417 23:26:58.081654 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-c7nff" podStartSLOduration=6.36571512 podStartE2EDuration="22.081626263s" podCreationTimestamp="2026-04-17 23:26:36 +0000 UTC" firstStartedPulling="2026-04-17 23:26:37.812657051 +0000 UTC m=+7.023651384" lastFinishedPulling="2026-04-17 23:26:53.528568194 +0000 UTC m=+22.739562527" observedRunningTime="2026-04-17 23:26:58.080965862 +0000 UTC m=+27.291960275" watchObservedRunningTime="2026-04-17 23:26:58.081626263 +0000 UTC m=+27.292620636" Apr 17 23:26:59.358507 systemd-networkd[1242]: cilium_host: Link UP Apr 17 23:26:59.358630 systemd-networkd[1242]: cilium_net: Link UP Apr 17 23:26:59.358748 systemd-networkd[1242]: cilium_net: Gained carrier Apr 17 23:26:59.358909 systemd-networkd[1242]: cilium_host: Gained carrier Apr 17 23:26:59.475584 systemd-networkd[1242]: cilium_vxlan: Link UP Apr 17 23:26:59.476645 systemd-networkd[1242]: cilium_vxlan: Gained carrier Apr 17 23:26:59.705411 systemd-networkd[1242]: cilium_host: Gained IPv6LL Apr 17 23:26:59.770212 kernel: NET: Registered PF_ALG protocol family Apr 17 23:27:00.073348 systemd-networkd[1242]: cilium_net: Gained IPv6LL Apr 17 23:27:00.470805 systemd-networkd[1242]: lxc_health: Link UP Apr 17 23:27:00.480647 systemd-networkd[1242]: lxc_health: Gained carrier Apr 17 23:27:00.523265 systemd-networkd[1242]: cilium_vxlan: Gained IPv6LL Apr 17 23:27:00.765329 systemd-networkd[1242]: lxc3f3b650d23d3: Link UP Apr 17 23:27:00.771229 kernel: eth0: renamed from tmp2676a Apr 17 23:27:00.780006 systemd-networkd[1242]: lxcd52a3f3bc931: Link UP Apr 17 23:27:00.786216 kernel: eth0: renamed from tmp82352 Apr 17 23:27:00.791404 systemd-networkd[1242]: lxc3f3b650d23d3: Gained carrier Apr 17 23:27:00.798820 systemd-networkd[1242]: lxcd52a3f3bc931: Gained carrier Apr 17 23:27:02.057759 systemd-networkd[1242]: lxcd52a3f3bc931: Gained IPv6LL Apr 17 23:27:02.313919 systemd-networkd[1242]: lxc_health: Gained IPv6LL Apr 17 23:27:02.378031 systemd-networkd[1242]: lxc3f3b650d23d3: Gained IPv6LL Apr 17 23:27:04.871211 containerd[1594]: time="2026-04-17T23:27:04.870080662Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:27:04.871211 containerd[1594]: time="2026-04-17T23:27:04.870229335Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:27:04.871211 containerd[1594]: time="2026-04-17T23:27:04.870250454Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:04.871211 containerd[1594]: time="2026-04-17T23:27:04.870809667Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:04.896243 containerd[1594]: time="2026-04-17T23:27:04.886669322Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:27:04.896243 containerd[1594]: time="2026-04-17T23:27:04.886736439Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:27:04.896243 containerd[1594]: time="2026-04-17T23:27:04.886761838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:04.896243 containerd[1594]: time="2026-04-17T23:27:04.886858473Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:04.983901 containerd[1594]: time="2026-04-17T23:27:04.983758880Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-28f4s,Uid:62da7207-6dff-45bd-b9ca-93fe5d21900c,Namespace:kube-system,Attempt:0,} returns sandbox id \"823528228f883011428c3a44bd95f87a1ab9e4b6bf70c193ed5f34cf428c6e78\"" Apr 17 23:27:04.997206 containerd[1594]: time="2026-04-17T23:27:04.991517275Z" level=info msg="CreateContainer within sandbox \"823528228f883011428c3a44bd95f87a1ab9e4b6bf70c193ed5f34cf428c6e78\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 17 23:27:05.009267 containerd[1594]: time="2026-04-17T23:27:05.007671047Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-8xb5l,Uid:0f85c524-feb4-4cd3-8152-9f76ba7da00f,Namespace:kube-system,Attempt:0,} returns sandbox id \"2676a02226b70646779ea7dc1ed3413640cb34299a64122d64c7fa5a494ab347\"" Apr 17 23:27:05.016281 containerd[1594]: time="2026-04-17T23:27:05.016230299Z" level=info msg="CreateContainer within sandbox \"2676a02226b70646779ea7dc1ed3413640cb34299a64122d64c7fa5a494ab347\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 17 23:27:05.033788 containerd[1594]: time="2026-04-17T23:27:05.033647710Z" level=info msg="CreateContainer within sandbox \"823528228f883011428c3a44bd95f87a1ab9e4b6bf70c193ed5f34cf428c6e78\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"112678f7fa9dd44dda130370bc00b621db9e541df668209661d91cc4617c2990\"" Apr 17 23:27:05.036191 containerd[1594]: time="2026-04-17T23:27:05.036114998Z" level=info msg="StartContainer for \"112678f7fa9dd44dda130370bc00b621db9e541df668209661d91cc4617c2990\"" Apr 17 23:27:05.038731 containerd[1594]: time="2026-04-17T23:27:05.038626284Z" level=info msg="CreateContainer within sandbox \"2676a02226b70646779ea7dc1ed3413640cb34299a64122d64c7fa5a494ab347\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"2d1659d00fe323e6d5928a92209d41344250d09dcd89809a95b9481bcd2acd20\"" Apr 17 23:27:05.040392 containerd[1594]: time="2026-04-17T23:27:05.039230177Z" level=info msg="StartContainer for \"2d1659d00fe323e6d5928a92209d41344250d09dcd89809a95b9481bcd2acd20\"" Apr 17 23:27:05.115799 containerd[1594]: time="2026-04-17T23:27:05.115735229Z" level=info msg="StartContainer for \"2d1659d00fe323e6d5928a92209d41344250d09dcd89809a95b9481bcd2acd20\" returns successfully" Apr 17 23:27:05.118845 containerd[1594]: time="2026-04-17T23:27:05.118799490Z" level=info msg="StartContainer for \"112678f7fa9dd44dda130370bc00b621db9e541df668209661d91cc4617c2990\" returns successfully" Apr 17 23:27:05.886369 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2928216603.mount: Deactivated successfully. Apr 17 23:27:06.118548 kubelet[2680]: I0417 23:27:06.118397 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-28f4s" podStartSLOduration=29.118351531 podStartE2EDuration="29.118351531s" podCreationTimestamp="2026-04-17 23:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:27:06.115466938 +0000 UTC m=+35.326461311" watchObservedRunningTime="2026-04-17 23:27:06.118351531 +0000 UTC m=+35.329345864" Apr 17 23:27:06.155141 kubelet[2680]: I0417 23:27:06.154768 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-8xb5l" podStartSLOduration=29.154748859 podStartE2EDuration="29.154748859s" podCreationTimestamp="2026-04-17 23:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:27:06.136160352 +0000 UTC m=+35.347154685" watchObservedRunningTime="2026-04-17 23:27:06.154748859 +0000 UTC m=+35.365743272" Apr 17 23:27:13.579517 systemd[1]: Started sshd@7-46.224.37.37:22-50.85.169.122:55090.service - OpenSSH per-connection server daemon (50.85.169.122:55090). Apr 17 23:27:13.694238 sshd[4095]: Accepted publickey for core from 50.85.169.122 port 55090 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:13.696618 sshd[4095]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:13.703122 systemd-logind[1565]: New session 8 of user core. Apr 17 23:27:13.707471 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 17 23:27:13.893429 sshd[4095]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:13.900455 systemd-logind[1565]: Session 8 logged out. Waiting for processes to exit. Apr 17 23:27:13.901544 systemd[1]: sshd@7-46.224.37.37:22-50.85.169.122:55090.service: Deactivated successfully. Apr 17 23:27:13.909636 systemd[1]: session-8.scope: Deactivated successfully. Apr 17 23:27:13.912993 systemd-logind[1565]: Removed session 8. Apr 17 23:27:18.916502 systemd[1]: Started sshd@8-46.224.37.37:22-50.85.169.122:55096.service - OpenSSH per-connection server daemon (50.85.169.122:55096). Apr 17 23:27:19.031692 sshd[4111]: Accepted publickey for core from 50.85.169.122 port 55096 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:19.034089 sshd[4111]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:19.039965 systemd-logind[1565]: New session 9 of user core. Apr 17 23:27:19.048702 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 17 23:27:19.227425 sshd[4111]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:19.232255 systemd[1]: sshd@8-46.224.37.37:22-50.85.169.122:55096.service: Deactivated successfully. Apr 17 23:27:19.238394 systemd[1]: session-9.scope: Deactivated successfully. Apr 17 23:27:19.240653 systemd-logind[1565]: Session 9 logged out. Waiting for processes to exit. Apr 17 23:27:19.242158 systemd-logind[1565]: Removed session 9. Apr 17 23:27:24.253863 systemd[1]: Started sshd@9-46.224.37.37:22-50.85.169.122:40266.service - OpenSSH per-connection server daemon (50.85.169.122:40266). Apr 17 23:27:24.378082 sshd[4127]: Accepted publickey for core from 50.85.169.122 port 40266 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:24.380455 sshd[4127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:24.388207 systemd-logind[1565]: New session 10 of user core. Apr 17 23:27:24.393750 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 17 23:27:24.568206 sshd[4127]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:24.572998 systemd-logind[1565]: Session 10 logged out. Waiting for processes to exit. Apr 17 23:27:24.575268 systemd[1]: sshd@9-46.224.37.37:22-50.85.169.122:40266.service: Deactivated successfully. Apr 17 23:27:24.579937 systemd[1]: session-10.scope: Deactivated successfully. Apr 17 23:27:24.581918 systemd-logind[1565]: Removed session 10. Apr 17 23:27:29.589469 systemd[1]: Started sshd@10-46.224.37.37:22-50.85.169.122:37326.service - OpenSSH per-connection server daemon (50.85.169.122:37326). Apr 17 23:27:29.704401 sshd[4141]: Accepted publickey for core from 50.85.169.122 port 37326 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:29.706807 sshd[4141]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:29.712837 systemd-logind[1565]: New session 11 of user core. Apr 17 23:27:29.718856 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 17 23:27:29.889434 sshd[4141]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:29.894975 systemd-logind[1565]: Session 11 logged out. Waiting for processes to exit. Apr 17 23:27:29.896065 systemd[1]: sshd@10-46.224.37.37:22-50.85.169.122:37326.service: Deactivated successfully. Apr 17 23:27:29.901142 systemd[1]: session-11.scope: Deactivated successfully. Apr 17 23:27:29.902628 systemd-logind[1565]: Removed session 11. Apr 17 23:27:29.917748 systemd[1]: Started sshd@11-46.224.37.37:22-50.85.169.122:37338.service - OpenSSH per-connection server daemon (50.85.169.122:37338). Apr 17 23:27:30.033244 sshd[4156]: Accepted publickey for core from 50.85.169.122 port 37338 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:30.035640 sshd[4156]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:30.040676 systemd-logind[1565]: New session 12 of user core. Apr 17 23:27:30.049135 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 17 23:27:30.267706 sshd[4156]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:30.276539 systemd[1]: sshd@11-46.224.37.37:22-50.85.169.122:37338.service: Deactivated successfully. Apr 17 23:27:30.288613 systemd[1]: session-12.scope: Deactivated successfully. Apr 17 23:27:30.291631 systemd-logind[1565]: Session 12 logged out. Waiting for processes to exit. Apr 17 23:27:30.306273 systemd[1]: Started sshd@12-46.224.37.37:22-50.85.169.122:37354.service - OpenSSH per-connection server daemon (50.85.169.122:37354). Apr 17 23:27:30.307260 systemd-logind[1565]: Removed session 12. Apr 17 23:27:30.453262 sshd[4169]: Accepted publickey for core from 50.85.169.122 port 37354 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:30.455049 sshd[4169]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:30.460436 systemd-logind[1565]: New session 13 of user core. Apr 17 23:27:30.468848 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 17 23:27:30.650446 sshd[4169]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:30.657526 systemd[1]: sshd@12-46.224.37.37:22-50.85.169.122:37354.service: Deactivated successfully. Apr 17 23:27:30.663226 systemd[1]: session-13.scope: Deactivated successfully. Apr 17 23:27:30.665695 systemd-logind[1565]: Session 13 logged out. Waiting for processes to exit. Apr 17 23:27:30.667582 systemd-logind[1565]: Removed session 13. Apr 17 23:27:35.674483 systemd[1]: Started sshd@13-46.224.37.37:22-50.85.169.122:37364.service - OpenSSH per-connection server daemon (50.85.169.122:37364). Apr 17 23:27:35.795245 sshd[4186]: Accepted publickey for core from 50.85.169.122 port 37364 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:35.797713 sshd[4186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:35.802870 systemd-logind[1565]: New session 14 of user core. Apr 17 23:27:35.809459 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 17 23:27:35.978588 sshd[4186]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:35.984994 systemd[1]: sshd@13-46.224.37.37:22-50.85.169.122:37364.service: Deactivated successfully. Apr 17 23:27:35.988490 systemd-logind[1565]: Session 14 logged out. Waiting for processes to exit. Apr 17 23:27:35.989035 systemd[1]: session-14.scope: Deactivated successfully. Apr 17 23:27:35.990800 systemd-logind[1565]: Removed session 14. Apr 17 23:27:41.002647 systemd[1]: Started sshd@14-46.224.37.37:22-50.85.169.122:54512.service - OpenSSH per-connection server daemon (50.85.169.122:54512). Apr 17 23:27:41.143227 sshd[4201]: Accepted publickey for core from 50.85.169.122 port 54512 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:41.144759 sshd[4201]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:41.151848 systemd-logind[1565]: New session 15 of user core. Apr 17 23:27:41.154532 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 17 23:27:41.328465 sshd[4201]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:41.332480 systemd[1]: sshd@14-46.224.37.37:22-50.85.169.122:54512.service: Deactivated successfully. Apr 17 23:27:41.336105 systemd-logind[1565]: Session 15 logged out. Waiting for processes to exit. Apr 17 23:27:41.336821 systemd[1]: session-15.scope: Deactivated successfully. Apr 17 23:27:41.338992 systemd-logind[1565]: Removed session 15. Apr 17 23:27:41.346610 systemd[1]: Started sshd@15-46.224.37.37:22-50.85.169.122:54516.service - OpenSSH per-connection server daemon (50.85.169.122:54516). Apr 17 23:27:41.460956 sshd[4215]: Accepted publickey for core from 50.85.169.122 port 54516 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:41.463925 sshd[4215]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:41.471291 systemd-logind[1565]: New session 16 of user core. Apr 17 23:27:41.474537 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 17 23:27:41.722539 sshd[4215]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:41.727827 systemd[1]: sshd@15-46.224.37.37:22-50.85.169.122:54516.service: Deactivated successfully. Apr 17 23:27:41.733153 systemd-logind[1565]: Session 16 logged out. Waiting for processes to exit. Apr 17 23:27:41.733597 systemd[1]: session-16.scope: Deactivated successfully. Apr 17 23:27:41.736205 systemd-logind[1565]: Removed session 16. Apr 17 23:27:41.749543 systemd[1]: Started sshd@16-46.224.37.37:22-50.85.169.122:54528.service - OpenSSH per-connection server daemon (50.85.169.122:54528). Apr 17 23:27:41.875688 sshd[4227]: Accepted publickey for core from 50.85.169.122 port 54528 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:41.878070 sshd[4227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:41.883389 systemd-logind[1565]: New session 17 of user core. Apr 17 23:27:41.885524 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 17 23:27:42.658350 sshd[4227]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:42.667428 systemd[1]: sshd@16-46.224.37.37:22-50.85.169.122:54528.service: Deactivated successfully. Apr 17 23:27:42.678159 systemd[1]: session-17.scope: Deactivated successfully. Apr 17 23:27:42.680721 systemd-logind[1565]: Session 17 logged out. Waiting for processes to exit. Apr 17 23:27:42.690634 systemd[1]: Started sshd@17-46.224.37.37:22-50.85.169.122:54536.service - OpenSSH per-connection server daemon (50.85.169.122:54536). Apr 17 23:27:42.692854 systemd-logind[1565]: Removed session 17. Apr 17 23:27:42.809130 sshd[4246]: Accepted publickey for core from 50.85.169.122 port 54536 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:42.811912 sshd[4246]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:42.817266 systemd-logind[1565]: New session 18 of user core. Apr 17 23:27:42.820569 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 17 23:27:43.114559 sshd[4246]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:43.123544 systemd[1]: sshd@17-46.224.37.37:22-50.85.169.122:54536.service: Deactivated successfully. Apr 17 23:27:43.127854 systemd[1]: session-18.scope: Deactivated successfully. Apr 17 23:27:43.127975 systemd-logind[1565]: Session 18 logged out. Waiting for processes to exit. Apr 17 23:27:43.138532 systemd[1]: Started sshd@18-46.224.37.37:22-50.85.169.122:54546.service - OpenSSH per-connection server daemon (50.85.169.122:54546). Apr 17 23:27:43.140285 systemd-logind[1565]: Removed session 18. Apr 17 23:27:43.266263 sshd[4258]: Accepted publickey for core from 50.85.169.122 port 54546 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:43.268864 sshd[4258]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:43.275602 systemd-logind[1565]: New session 19 of user core. Apr 17 23:27:43.285807 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 17 23:27:43.463956 sshd[4258]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:43.469946 systemd[1]: sshd@18-46.224.37.37:22-50.85.169.122:54546.service: Deactivated successfully. Apr 17 23:27:43.474248 systemd[1]: session-19.scope: Deactivated successfully. Apr 17 23:27:43.476264 systemd-logind[1565]: Session 19 logged out. Waiting for processes to exit. Apr 17 23:27:43.477619 systemd-logind[1565]: Removed session 19. Apr 17 23:27:48.489485 systemd[1]: Started sshd@19-46.224.37.37:22-50.85.169.122:54562.service - OpenSSH per-connection server daemon (50.85.169.122:54562). Apr 17 23:27:48.604833 sshd[4274]: Accepted publickey for core from 50.85.169.122 port 54562 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:48.607339 sshd[4274]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:48.613255 systemd-logind[1565]: New session 20 of user core. Apr 17 23:27:48.623308 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 17 23:27:48.802556 sshd[4274]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:48.807131 systemd[1]: sshd@19-46.224.37.37:22-50.85.169.122:54562.service: Deactivated successfully. Apr 17 23:27:48.811239 systemd-logind[1565]: Session 20 logged out. Waiting for processes to exit. Apr 17 23:27:48.811673 systemd[1]: session-20.scope: Deactivated successfully. Apr 17 23:27:48.813832 systemd-logind[1565]: Removed session 20. Apr 17 23:27:53.827426 systemd[1]: Started sshd@20-46.224.37.37:22-50.85.169.122:49774.service - OpenSSH per-connection server daemon (50.85.169.122:49774). Apr 17 23:27:53.962239 sshd[4287]: Accepted publickey for core from 50.85.169.122 port 49774 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:53.963580 sshd[4287]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:53.968572 systemd-logind[1565]: New session 21 of user core. Apr 17 23:27:53.974759 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 17 23:27:54.141983 sshd[4287]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:54.148739 systemd[1]: sshd@20-46.224.37.37:22-50.85.169.122:49774.service: Deactivated successfully. Apr 17 23:27:54.152626 systemd-logind[1565]: Session 21 logged out. Waiting for processes to exit. Apr 17 23:27:54.154000 systemd[1]: session-21.scope: Deactivated successfully. Apr 17 23:27:54.161730 systemd[1]: Started sshd@21-46.224.37.37:22-50.85.169.122:49784.service - OpenSSH per-connection server daemon (50.85.169.122:49784). Apr 17 23:27:54.162695 systemd-logind[1565]: Removed session 21. Apr 17 23:27:54.275033 sshd[4301]: Accepted publickey for core from 50.85.169.122 port 49784 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:54.277462 sshd[4301]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:54.282890 systemd-logind[1565]: New session 22 of user core. Apr 17 23:27:54.290450 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 17 23:27:55.908819 containerd[1594]: time="2026-04-17T23:27:55.908319733Z" level=info msg="StopContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" with timeout 30 (s)" Apr 17 23:27:55.910420 containerd[1594]: time="2026-04-17T23:27:55.908970365Z" level=info msg="Stop container \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" with signal terminated" Apr 17 23:27:55.939363 systemd[1]: run-containerd-runc-k8s.io-48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7-runc.SNcyIi.mount: Deactivated successfully. Apr 17 23:27:55.953371 containerd[1594]: time="2026-04-17T23:27:55.952706331Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 17 23:27:55.961704 containerd[1594]: time="2026-04-17T23:27:55.961590908Z" level=info msg="StopContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" with timeout 2 (s)" Apr 17 23:27:55.963669 containerd[1594]: time="2026-04-17T23:27:55.963638853Z" level=info msg="Stop container \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" with signal terminated" Apr 17 23:27:55.964044 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075-rootfs.mount: Deactivated successfully. Apr 17 23:27:55.975780 systemd-networkd[1242]: lxc_health: Link DOWN Apr 17 23:27:55.975787 systemd-networkd[1242]: lxc_health: Lost carrier Apr 17 23:27:55.990439 containerd[1594]: time="2026-04-17T23:27:55.987810750Z" level=info msg="shim disconnected" id=39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075 namespace=k8s.io Apr 17 23:27:55.990439 containerd[1594]: time="2026-04-17T23:27:55.987869196Z" level=warning msg="cleaning up after shim disconnected" id=39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075 namespace=k8s.io Apr 17 23:27:55.990439 containerd[1594]: time="2026-04-17T23:27:55.987905000Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:27:56.001847 containerd[1594]: time="2026-04-17T23:27:56.001713236Z" level=warning msg="cleanup warnings time=\"2026-04-17T23:27:55Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Apr 17 23:27:56.005368 containerd[1594]: time="2026-04-17T23:27:56.005325540Z" level=info msg="StopContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" returns successfully" Apr 17 23:27:56.006381 containerd[1594]: time="2026-04-17T23:27:56.005990131Z" level=info msg="StopPodSandbox for \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\"" Apr 17 23:27:56.006381 containerd[1594]: time="2026-04-17T23:27:56.006036856Z" level=info msg="Container to stop \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.009372 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e-shm.mount: Deactivated successfully. Apr 17 23:27:56.023157 kubelet[2680]: E0417 23:27:56.023110 2680 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 17 23:27:56.026648 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7-rootfs.mount: Deactivated successfully. Apr 17 23:27:56.035859 containerd[1594]: time="2026-04-17T23:27:56.035635889Z" level=info msg="shim disconnected" id=48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7 namespace=k8s.io Apr 17 23:27:56.035859 containerd[1594]: time="2026-04-17T23:27:56.035688975Z" level=warning msg="cleaning up after shim disconnected" id=48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7 namespace=k8s.io Apr 17 23:27:56.035859 containerd[1594]: time="2026-04-17T23:27:56.035697256Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:27:56.053918 containerd[1594]: time="2026-04-17T23:27:56.053754739Z" level=info msg="shim disconnected" id=694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e namespace=k8s.io Apr 17 23:27:56.053918 containerd[1594]: time="2026-04-17T23:27:56.053805705Z" level=warning msg="cleaning up after shim disconnected" id=694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e namespace=k8s.io Apr 17 23:27:56.053918 containerd[1594]: time="2026-04-17T23:27:56.053814385Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:27:56.061345 containerd[1594]: time="2026-04-17T23:27:56.061217414Z" level=info msg="StopContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" returns successfully" Apr 17 23:27:56.062089 containerd[1594]: time="2026-04-17T23:27:56.062063144Z" level=info msg="StopPodSandbox for \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\"" Apr 17 23:27:56.062158 containerd[1594]: time="2026-04-17T23:27:56.062105829Z" level=info msg="Container to stop \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.062158 containerd[1594]: time="2026-04-17T23:27:56.062118110Z" level=info msg="Container to stop \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.062158 containerd[1594]: time="2026-04-17T23:27:56.062127831Z" level=info msg="Container to stop \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.062158 containerd[1594]: time="2026-04-17T23:27:56.062137752Z" level=info msg="Container to stop \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.062158 containerd[1594]: time="2026-04-17T23:27:56.062146873Z" level=info msg="Container to stop \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 17 23:27:56.077730 containerd[1594]: time="2026-04-17T23:27:56.077676327Z" level=info msg="TearDown network for sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" successfully" Apr 17 23:27:56.077730 containerd[1594]: time="2026-04-17T23:27:56.077715091Z" level=info msg="StopPodSandbox for \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" returns successfully" Apr 17 23:27:56.101554 containerd[1594]: time="2026-04-17T23:27:56.101270160Z" level=info msg="shim disconnected" id=7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13 namespace=k8s.io Apr 17 23:27:56.101554 containerd[1594]: time="2026-04-17T23:27:56.101399174Z" level=warning msg="cleaning up after shim disconnected" id=7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13 namespace=k8s.io Apr 17 23:27:56.101554 containerd[1594]: time="2026-04-17T23:27:56.101410855Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:27:56.115032 containerd[1594]: time="2026-04-17T23:27:56.114661347Z" level=info msg="TearDown network for sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" successfully" Apr 17 23:27:56.115032 containerd[1594]: time="2026-04-17T23:27:56.114698711Z" level=info msg="StopPodSandbox for \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" returns successfully" Apr 17 23:27:56.227267 kubelet[2680]: I0417 23:27:56.226871 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hubble-tls\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227484 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hostproc\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227519 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-config-path\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227539 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-kernel\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227555 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cni-path\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227571 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-etc-cni-netd\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.227996 kubelet[2680]: I0417 23:27:56.227587 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-xtables-lock\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227624 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-bpf-maps\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227643 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-cilium-config-path\") pod \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\" (UID: \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227657 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-lib-modules\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227678 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59zzq\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227692 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-cgroup\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228190 kubelet[2680]: I0417 23:27:56.227708 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j4f9\" (UniqueName: \"kubernetes.io/projected/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-kube-api-access-8j4f9\") pod \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\" (UID: \"1aa6e5bd-1e8e-4a18-b263-e64831e7b699\") " Apr 17 23:27:56.228384 kubelet[2680]: I0417 23:27:56.227723 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-net\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228384 kubelet[2680]: I0417 23:27:56.227741 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-run\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.228384 kubelet[2680]: I0417 23:27:56.227760 2680 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-clustermesh-secrets\") pod \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\" (UID: \"e561372a-e5ae-4fc2-ab53-44127aa5ca4a\") " Apr 17 23:27:56.231929 kubelet[2680]: I0417 23:27:56.231896 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 23:27:56.233786 kubelet[2680]: I0417 23:27:56.232376 2680 scope.go:117] "RemoveContainer" containerID="48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7" Apr 17 23:27:56.234270 kubelet[2680]: I0417 23:27:56.234237 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.234350 kubelet[2680]: I0417 23:27:56.234302 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hostproc" (OuterVolumeSpecName: "hostproc") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.236427 kubelet[2680]: I0417 23:27:56.236353 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 23:27:56.236427 kubelet[2680]: I0417 23:27:56.236418 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.236531 kubelet[2680]: I0417 23:27:56.236438 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cni-path" (OuterVolumeSpecName: "cni-path") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.236531 kubelet[2680]: I0417 23:27:56.236454 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.236531 kubelet[2680]: I0417 23:27:56.236469 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.236531 kubelet[2680]: I0417 23:27:56.236513 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 23:27:56.237123 kubelet[2680]: I0417 23:27:56.236851 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.237123 kubelet[2680]: I0417 23:27:56.236849 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1aa6e5bd-1e8e-4a18-b263-e64831e7b699" (UID: "1aa6e5bd-1e8e-4a18-b263-e64831e7b699"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 23:27:56.237123 kubelet[2680]: I0417 23:27:56.236892 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.240471 kubelet[2680]: I0417 23:27:56.239721 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.240471 kubelet[2680]: I0417 23:27:56.240102 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 17 23:27:56.243678 kubelet[2680]: I0417 23:27:56.243635 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq" (OuterVolumeSpecName: "kube-api-access-59zzq") pod "e561372a-e5ae-4fc2-ab53-44127aa5ca4a" (UID: "e561372a-e5ae-4fc2-ab53-44127aa5ca4a"). InnerVolumeSpecName "kube-api-access-59zzq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 23:27:56.243929 containerd[1594]: time="2026-04-17T23:27:56.243891752Z" level=info msg="RemoveContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\"" Apr 17 23:27:56.246168 kubelet[2680]: I0417 23:27:56.246113 2680 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-kube-api-access-8j4f9" (OuterVolumeSpecName: "kube-api-access-8j4f9") pod "1aa6e5bd-1e8e-4a18-b263-e64831e7b699" (UID: "1aa6e5bd-1e8e-4a18-b263-e64831e7b699"). InnerVolumeSpecName "kube-api-access-8j4f9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 23:27:56.249722 containerd[1594]: time="2026-04-17T23:27:56.249683609Z" level=info msg="RemoveContainer for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" returns successfully" Apr 17 23:27:56.249988 kubelet[2680]: I0417 23:27:56.249927 2680 scope.go:117] "RemoveContainer" containerID="e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b" Apr 17 23:27:56.251137 containerd[1594]: time="2026-04-17T23:27:56.251110121Z" level=info msg="RemoveContainer for \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\"" Apr 17 23:27:56.255010 containerd[1594]: time="2026-04-17T23:27:56.254977173Z" level=info msg="RemoveContainer for \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\" returns successfully" Apr 17 23:27:56.255326 kubelet[2680]: I0417 23:27:56.255274 2680 scope.go:117] "RemoveContainer" containerID="8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1" Apr 17 23:27:56.256780 containerd[1594]: time="2026-04-17T23:27:56.256754562Z" level=info msg="RemoveContainer for \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\"" Apr 17 23:27:56.262633 containerd[1594]: time="2026-04-17T23:27:56.262564661Z" level=info msg="RemoveContainer for \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\" returns successfully" Apr 17 23:27:56.262907 kubelet[2680]: I0417 23:27:56.262868 2680 scope.go:117] "RemoveContainer" containerID="04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e" Apr 17 23:27:56.264473 containerd[1594]: time="2026-04-17T23:27:56.264308087Z" level=info msg="RemoveContainer for \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\"" Apr 17 23:27:56.269092 containerd[1594]: time="2026-04-17T23:27:56.268973224Z" level=info msg="RemoveContainer for \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\" returns successfully" Apr 17 23:27:56.269388 kubelet[2680]: I0417 23:27:56.269272 2680 scope.go:117] "RemoveContainer" containerID="a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78" Apr 17 23:27:56.270866 containerd[1594]: time="2026-04-17T23:27:56.270821301Z" level=info msg="RemoveContainer for \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\"" Apr 17 23:27:56.278522 containerd[1594]: time="2026-04-17T23:27:56.278291376Z" level=info msg="RemoveContainer for \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\" returns successfully" Apr 17 23:27:56.278877 kubelet[2680]: I0417 23:27:56.278848 2680 scope.go:117] "RemoveContainer" containerID="48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7" Apr 17 23:27:56.281652 containerd[1594]: time="2026-04-17T23:27:56.280718555Z" level=error msg="ContainerStatus for \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\": not found" Apr 17 23:27:56.281767 kubelet[2680]: E0417 23:27:56.281346 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\": not found" containerID="48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7" Apr 17 23:27:56.281767 kubelet[2680]: I0417 23:27:56.281383 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7"} err="failed to get container status \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\": rpc error: code = NotFound desc = an error occurred when try to find container \"48c2e7a9895c343243afad0cb2688a299d3ab2f6ebeef6bd1e3e87a9fb2b13d7\": not found" Apr 17 23:27:56.281767 kubelet[2680]: I0417 23:27:56.281437 2680 scope.go:117] "RemoveContainer" containerID="e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b" Apr 17 23:27:56.282450 containerd[1594]: time="2026-04-17T23:27:56.282407295Z" level=error msg="ContainerStatus for \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\": not found" Apr 17 23:27:56.282800 kubelet[2680]: E0417 23:27:56.282664 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\": not found" containerID="e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b" Apr 17 23:27:56.282878 kubelet[2680]: I0417 23:27:56.282805 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b"} err="failed to get container status \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\": rpc error: code = NotFound desc = an error occurred when try to find container \"e1f0aff26c68af3147aba2479ad807b5771f3f23ef829878b78dcdba7840a12b\": not found" Apr 17 23:27:56.282878 kubelet[2680]: I0417 23:27:56.282825 2680 scope.go:117] "RemoveContainer" containerID="8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1" Apr 17 23:27:56.283572 containerd[1594]: time="2026-04-17T23:27:56.283526454Z" level=error msg="ContainerStatus for \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\": not found" Apr 17 23:27:56.283984 kubelet[2680]: E0417 23:27:56.283939 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\": not found" containerID="8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1" Apr 17 23:27:56.284042 kubelet[2680]: I0417 23:27:56.283991 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1"} err="failed to get container status \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\": rpc error: code = NotFound desc = an error occurred when try to find container \"8f1895906a8d8e6fd75f1f04d794fdb1637aadff5bdf174a06871dbdda8f45a1\": not found" Apr 17 23:27:56.284042 kubelet[2680]: I0417 23:27:56.284008 2680 scope.go:117] "RemoveContainer" containerID="04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e" Apr 17 23:27:56.285507 containerd[1594]: time="2026-04-17T23:27:56.285393013Z" level=error msg="ContainerStatus for \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\": not found" Apr 17 23:27:56.286103 kubelet[2680]: E0417 23:27:56.285957 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\": not found" containerID="04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e" Apr 17 23:27:56.286103 kubelet[2680]: I0417 23:27:56.285997 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e"} err="failed to get container status \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\": rpc error: code = NotFound desc = an error occurred when try to find container \"04a0c1ffdc77aaf4aeba3a7dd145232139b27f0ab14c532655dd1a20568cb44e\": not found" Apr 17 23:27:56.286103 kubelet[2680]: I0417 23:27:56.286014 2680 scope.go:117] "RemoveContainer" containerID="a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78" Apr 17 23:27:56.286228 containerd[1594]: time="2026-04-17T23:27:56.286195738Z" level=error msg="ContainerStatus for \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\": not found" Apr 17 23:27:56.286710 kubelet[2680]: E0417 23:27:56.286300 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\": not found" containerID="a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78" Apr 17 23:27:56.286710 kubelet[2680]: I0417 23:27:56.286362 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78"} err="failed to get container status \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\": rpc error: code = NotFound desc = an error occurred when try to find container \"a3cb387858f24613c95c6ad0792e13dc1f8bbf387f0c3cdfa30fc922c949ad78\": not found" Apr 17 23:27:56.286710 kubelet[2680]: I0417 23:27:56.286380 2680 scope.go:117] "RemoveContainer" containerID="39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075" Apr 17 23:27:56.287538 containerd[1594]: time="2026-04-17T23:27:56.287447072Z" level=info msg="RemoveContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\"" Apr 17 23:27:56.294348 containerd[1594]: time="2026-04-17T23:27:56.294303082Z" level=info msg="RemoveContainer for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" returns successfully" Apr 17 23:27:56.294883 kubelet[2680]: I0417 23:27:56.294698 2680 scope.go:117] "RemoveContainer" containerID="39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075" Apr 17 23:27:56.295373 containerd[1594]: time="2026-04-17T23:27:56.295204938Z" level=error msg="ContainerStatus for \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\": not found" Apr 17 23:27:56.295583 kubelet[2680]: E0417 23:27:56.295509 2680 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\": not found" containerID="39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075" Apr 17 23:27:56.295583 kubelet[2680]: I0417 23:27:56.295563 2680 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075"} err="failed to get container status \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\": rpc error: code = NotFound desc = an error occurred when try to find container \"39a980bef7c582b4d31cafcb66472d017fa9ea8a009a9256d840e7f00cf86075\": not found" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.328876 2680 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-run\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.328934 2680 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-clustermesh-secrets\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.328957 2680 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hubble-tls\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.328977 2680 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-hostproc\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.328997 2680 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-config-path\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.329017 2680 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-kernel\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.329041 2680 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cni-path\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329331 kubelet[2680]: I0417 23:27:56.329069 2680 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-etc-cni-netd\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329090 2680 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-xtables-lock\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329111 2680 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-bpf-maps\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329130 2680 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-cilium-config-path\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329148 2680 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-lib-modules\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329170 2680 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-59zzq\" (UniqueName: \"kubernetes.io/projected/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-kube-api-access-59zzq\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329241 2680 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-cilium-cgroup\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329261 2680 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8j4f9\" (UniqueName: \"kubernetes.io/projected/1aa6e5bd-1e8e-4a18-b263-e64831e7b699-kube-api-access-8j4f9\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.329783 kubelet[2680]: I0417 23:27:56.329280 2680 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e561372a-e5ae-4fc2-ab53-44127aa5ca4a-host-proc-sys-net\") on node \"ci-4081-3-6-n-8954fb58c4\" DevicePath \"\"" Apr 17 23:27:56.935798 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e-rootfs.mount: Deactivated successfully. Apr 17 23:27:56.936234 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13-rootfs.mount: Deactivated successfully. Apr 17 23:27:56.936458 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13-shm.mount: Deactivated successfully. Apr 17 23:27:56.936642 systemd[1]: var-lib-kubelet-pods-1aa6e5bd\x2d1e8e\x2d4a18\x2db263\x2de64831e7b699-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8j4f9.mount: Deactivated successfully. Apr 17 23:27:56.936836 systemd[1]: var-lib-kubelet-pods-e561372a\x2de5ae\x2d4fc2\x2dab53\x2d44127aa5ca4a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d59zzq.mount: Deactivated successfully. Apr 17 23:27:56.937018 systemd[1]: var-lib-kubelet-pods-e561372a\x2de5ae\x2d4fc2\x2dab53\x2d44127aa5ca4a-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 17 23:27:56.937223 systemd[1]: var-lib-kubelet-pods-e561372a\x2de5ae\x2d4fc2\x2dab53\x2d44127aa5ca4a-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 17 23:27:56.942423 kubelet[2680]: I0417 23:27:56.942345 2680 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aa6e5bd-1e8e-4a18-b263-e64831e7b699" path="/var/lib/kubelet/pods/1aa6e5bd-1e8e-4a18-b263-e64831e7b699/volumes" Apr 17 23:27:56.942800 kubelet[2680]: I0417 23:27:56.942782 2680 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e561372a-e5ae-4fc2-ab53-44127aa5ca4a" path="/var/lib/kubelet/pods/e561372a-e5ae-4fc2-ab53-44127aa5ca4a/volumes" Apr 17 23:27:57.862105 sshd[4301]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:57.866976 systemd-logind[1565]: Session 22 logged out. Waiting for processes to exit. Apr 17 23:27:57.867513 systemd[1]: sshd@21-46.224.37.37:22-50.85.169.122:49784.service: Deactivated successfully. Apr 17 23:27:57.871168 systemd[1]: session-22.scope: Deactivated successfully. Apr 17 23:27:57.874138 systemd-logind[1565]: Removed session 22. Apr 17 23:27:57.887557 systemd[1]: Started sshd@22-46.224.37.37:22-50.85.169.122:49788.service - OpenSSH per-connection server daemon (50.85.169.122:49788). Apr 17 23:27:58.011326 sshd[4470]: Accepted publickey for core from 50.85.169.122 port 49788 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:58.013448 sshd[4470]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:58.019310 systemd-logind[1565]: New session 23 of user core. Apr 17 23:27:58.023511 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 17 23:27:58.990658 sshd[4470]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:58.997022 systemd-logind[1565]: Session 23 logged out. Waiting for processes to exit. Apr 17 23:27:58.999610 systemd[1]: sshd@22-46.224.37.37:22-50.85.169.122:49788.service: Deactivated successfully. Apr 17 23:27:59.014789 systemd[1]: session-23.scope: Deactivated successfully. Apr 17 23:27:59.017445 systemd-logind[1565]: Removed session 23. Apr 17 23:27:59.025669 systemd[1]: Started sshd@23-46.224.37.37:22-50.85.169.122:49798.service - OpenSSH per-connection server daemon (50.85.169.122:49798). Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147138 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-host-proc-sys-net\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147204 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-bpf-maps\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147227 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-hostproc\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147241 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-cni-path\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147255 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/fb1b3371-090e-458d-a18c-0b295a697b98-cilium-ipsec-secrets\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.147694 kubelet[2680]: I0417 23:27:59.147271 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkr4t\" (UniqueName: \"kubernetes.io/projected/fb1b3371-090e-458d-a18c-0b295a697b98-kube-api-access-hkr4t\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147317 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-cilium-run\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147365 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-cilium-cgroup\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147383 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-host-proc-sys-kernel\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147406 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-etc-cni-netd\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147422 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-xtables-lock\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148439 kubelet[2680]: I0417 23:27:59.147437 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/fb1b3371-090e-458d-a18c-0b295a697b98-clustermesh-secrets\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148723 kubelet[2680]: I0417 23:27:59.147459 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/fb1b3371-090e-458d-a18c-0b295a697b98-hubble-tls\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148723 kubelet[2680]: I0417 23:27:59.147475 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fb1b3371-090e-458d-a18c-0b295a697b98-cilium-config-path\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.148723 kubelet[2680]: I0417 23:27:59.147520 2680 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fb1b3371-090e-458d-a18c-0b295a697b98-lib-modules\") pod \"cilium-8w77f\" (UID: \"fb1b3371-090e-458d-a18c-0b295a697b98\") " pod="kube-system/cilium-8w77f" Apr 17 23:27:59.159240 sshd[4483]: Accepted publickey for core from 50.85.169.122 port 49798 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:59.161643 sshd[4483]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:59.167573 systemd-logind[1565]: New session 24 of user core. Apr 17 23:27:59.176944 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 17 23:27:59.290535 sshd[4483]: pam_unix(sshd:session): session closed for user core Apr 17 23:27:59.298057 systemd[1]: sshd@23-46.224.37.37:22-50.85.169.122:49798.service: Deactivated successfully. Apr 17 23:27:59.302072 systemd-logind[1565]: Session 24 logged out. Waiting for processes to exit. Apr 17 23:27:59.304508 systemd[1]: session-24.scope: Deactivated successfully. Apr 17 23:27:59.309649 systemd[1]: Started sshd@24-46.224.37.37:22-50.85.169.122:49810.service - OpenSSH per-connection server daemon (50.85.169.122:49810). Apr 17 23:27:59.311399 systemd-logind[1565]: Removed session 24. Apr 17 23:27:59.312503 containerd[1594]: time="2026-04-17T23:27:59.312466943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-8w77f,Uid:fb1b3371-090e-458d-a18c-0b295a697b98,Namespace:kube-system,Attempt:0,}" Apr 17 23:27:59.335870 containerd[1594]: time="2026-04-17T23:27:59.335705156Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 17 23:27:59.335870 containerd[1594]: time="2026-04-17T23:27:59.335771603Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 17 23:27:59.335870 containerd[1594]: time="2026-04-17T23:27:59.335786884Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:59.336212 containerd[1594]: time="2026-04-17T23:27:59.335967942Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 17 23:27:59.375808 containerd[1594]: time="2026-04-17T23:27:59.375766481Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-8w77f,Uid:fb1b3371-090e-458d-a18c-0b295a697b98,Namespace:kube-system,Attempt:0,} returns sandbox id \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\"" Apr 17 23:27:59.381505 containerd[1594]: time="2026-04-17T23:27:59.381386226Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 17 23:27:59.392987 containerd[1594]: time="2026-04-17T23:27:59.392894902Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e637e1d3854d384244ce16eb8cfcb87a0d4557959f38fc361a443fddbd32f252\"" Apr 17 23:27:59.395416 containerd[1594]: time="2026-04-17T23:27:59.395383384Z" level=info msg="StartContainer for \"e637e1d3854d384244ce16eb8cfcb87a0d4557959f38fc361a443fddbd32f252\"" Apr 17 23:27:59.439802 sshd[4496]: Accepted publickey for core from 50.85.169.122 port 49810 ssh2: RSA SHA256:VfypDX1RTsDok1DcKRgqFkknflSVDpDNB07R6ghJc68 Apr 17 23:27:59.442577 sshd[4496]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 17 23:27:59.453598 systemd-logind[1565]: New session 25 of user core. Apr 17 23:27:59.454911 containerd[1594]: time="2026-04-17T23:27:59.454227050Z" level=info msg="StartContainer for \"e637e1d3854d384244ce16eb8cfcb87a0d4557959f38fc361a443fddbd32f252\" returns successfully" Apr 17 23:27:59.465543 systemd[1]: Started session-25.scope - Session 25 of User core. Apr 17 23:27:59.510698 containerd[1594]: time="2026-04-17T23:27:59.510315010Z" level=info msg="shim disconnected" id=e637e1d3854d384244ce16eb8cfcb87a0d4557959f38fc361a443fddbd32f252 namespace=k8s.io Apr 17 23:27:59.510698 containerd[1594]: time="2026-04-17T23:27:59.510389017Z" level=warning msg="cleaning up after shim disconnected" id=e637e1d3854d384244ce16eb8cfcb87a0d4557959f38fc361a443fddbd32f252 namespace=k8s.io Apr 17 23:27:59.510698 containerd[1594]: time="2026-04-17T23:27:59.510408019Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:00.285388 containerd[1594]: time="2026-04-17T23:28:00.284137727Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 17 23:28:00.308548 containerd[1594]: time="2026-04-17T23:28:00.308457253Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744\"" Apr 17 23:28:00.309631 containerd[1594]: time="2026-04-17T23:28:00.309479589Z" level=info msg="StartContainer for \"9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744\"" Apr 17 23:28:00.372602 containerd[1594]: time="2026-04-17T23:28:00.372559478Z" level=info msg="StartContainer for \"9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744\" returns successfully" Apr 17 23:28:00.409998 containerd[1594]: time="2026-04-17T23:28:00.409848583Z" level=info msg="shim disconnected" id=9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744 namespace=k8s.io Apr 17 23:28:00.409998 containerd[1594]: time="2026-04-17T23:28:00.409966114Z" level=warning msg="cleaning up after shim disconnected" id=9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744 namespace=k8s.io Apr 17 23:28:00.409998 containerd[1594]: time="2026-04-17T23:28:00.409975155Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:01.024065 kubelet[2680]: E0417 23:28:01.024029 2680 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 17 23:28:01.258063 systemd[1]: run-containerd-runc-k8s.io-9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744-runc.lXfIf9.mount: Deactivated successfully. Apr 17 23:28:01.258345 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9d368990075d4676390bec32a607e83266de47e14c37540985f8d8a7bd7f8744-rootfs.mount: Deactivated successfully. Apr 17 23:28:01.292830 containerd[1594]: time="2026-04-17T23:28:01.291574899Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 17 23:28:01.312167 containerd[1594]: time="2026-04-17T23:28:01.312124131Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0\"" Apr 17 23:28:01.313946 containerd[1594]: time="2026-04-17T23:28:01.313903573Z" level=info msg="StartContainer for \"260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0\"" Apr 17 23:28:01.381950 containerd[1594]: time="2026-04-17T23:28:01.381675227Z" level=info msg="StartContainer for \"260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0\" returns successfully" Apr 17 23:28:01.407975 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0-rootfs.mount: Deactivated successfully. Apr 17 23:28:01.411772 containerd[1594]: time="2026-04-17T23:28:01.411713884Z" level=info msg="shim disconnected" id=260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0 namespace=k8s.io Apr 17 23:28:01.411772 containerd[1594]: time="2026-04-17T23:28:01.411768609Z" level=warning msg="cleaning up after shim disconnected" id=260421cc93c97581a8a330d91efe9a120c02cfc9b3018403a3b13bdbb6b109e0 namespace=k8s.io Apr 17 23:28:01.411772 containerd[1594]: time="2026-04-17T23:28:01.411777770Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:02.298225 containerd[1594]: time="2026-04-17T23:28:02.298093288Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 17 23:28:02.322327 containerd[1594]: time="2026-04-17T23:28:02.322275544Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69\"" Apr 17 23:28:02.323014 containerd[1594]: time="2026-04-17T23:28:02.322977486Z" level=info msg="StartContainer for \"e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69\"" Apr 17 23:28:02.383059 containerd[1594]: time="2026-04-17T23:28:02.382203876Z" level=info msg="StartContainer for \"e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69\" returns successfully" Apr 17 23:28:02.407154 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69-rootfs.mount: Deactivated successfully. Apr 17 23:28:02.414221 containerd[1594]: time="2026-04-17T23:28:02.409349353Z" level=info msg="shim disconnected" id=e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69 namespace=k8s.io Apr 17 23:28:02.414221 containerd[1594]: time="2026-04-17T23:28:02.409401037Z" level=warning msg="cleaning up after shim disconnected" id=e5735a40232bf2ba8fa279b4c9bb14c1a85d84502a3b5964c04447a15e30ac69 namespace=k8s.io Apr 17 23:28:02.414221 containerd[1594]: time="2026-04-17T23:28:02.409409518Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:02.829810 kubelet[2680]: I0417 23:28:02.828123 2680 setters.go:618] "Node became not ready" node="ci-4081-3-6-n-8954fb58c4" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-04-17T23:28:02Z","lastTransitionTime":"2026-04-17T23:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Apr 17 23:28:03.312352 containerd[1594]: time="2026-04-17T23:28:03.312202481Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 17 23:28:03.330251 containerd[1594]: time="2026-04-17T23:28:03.330148657Z" level=info msg="CreateContainer within sandbox \"565c236c7cfbb6444a0825bb18019bca12b091407c378042b3b6465c4b9daedd\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"47520655533f4f87e6fd1add294075d9e8b48932f2588b3f6d564124030515bc\"" Apr 17 23:28:03.332291 containerd[1594]: time="2026-04-17T23:28:03.332241996Z" level=info msg="StartContainer for \"47520655533f4f87e6fd1add294075d9e8b48932f2588b3f6d564124030515bc\"" Apr 17 23:28:03.386510 containerd[1594]: time="2026-04-17T23:28:03.386308143Z" level=info msg="StartContainer for \"47520655533f4f87e6fd1add294075d9e8b48932f2588b3f6d564124030515bc\" returns successfully" Apr 17 23:28:03.698604 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Apr 17 23:28:04.324259 kubelet[2680]: I0417 23:28:04.323455 2680 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-8w77f" podStartSLOduration=6.323424147 podStartE2EDuration="6.323424147s" podCreationTimestamp="2026-04-17 23:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 23:28:04.322898384 +0000 UTC m=+93.533892717" watchObservedRunningTime="2026-04-17 23:28:04.323424147 +0000 UTC m=+93.534418520" Apr 17 23:28:06.709964 systemd-networkd[1242]: lxc_health: Link UP Apr 17 23:28:06.740045 systemd-networkd[1242]: lxc_health: Gained carrier Apr 17 23:28:08.084210 kubelet[2680]: E0417 23:28:08.083573 2680 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:42966->127.0.0.1:45747: write tcp 127.0.0.1:42966->127.0.0.1:45747: write: broken pipe Apr 17 23:28:08.105441 systemd-networkd[1242]: lxc_health: Gained IPv6LL Apr 17 23:28:12.419900 sshd[4496]: pam_unix(sshd:session): session closed for user core Apr 17 23:28:12.425983 systemd[1]: sshd@24-46.224.37.37:22-50.85.169.122:49810.service: Deactivated successfully. Apr 17 23:28:12.431715 systemd[1]: session-25.scope: Deactivated successfully. Apr 17 23:28:12.434308 systemd-logind[1565]: Session 25 logged out. Waiting for processes to exit. Apr 17 23:28:12.438047 systemd-logind[1565]: Removed session 25. Apr 17 23:28:26.989639 kubelet[2680]: E0417 23:28:26.989583 2680 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:33474->10.0.0.2:2379: read: connection timed out" Apr 17 23:28:27.018405 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351-rootfs.mount: Deactivated successfully. Apr 17 23:28:27.035587 containerd[1594]: time="2026-04-17T23:28:27.035406617Z" level=info msg="shim disconnected" id=29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351 namespace=k8s.io Apr 17 23:28:27.036397 containerd[1594]: time="2026-04-17T23:28:27.036129127Z" level=warning msg="cleaning up after shim disconnected" id=29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351 namespace=k8s.io Apr 17 23:28:27.036397 containerd[1594]: time="2026-04-17T23:28:27.036153968Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:27.374744 kubelet[2680]: I0417 23:28:27.374152 2680 scope.go:117] "RemoveContainer" containerID="29f342a278d26a1239bbd2ad47a48e90480108b6860a2bbe6ea73dc470d9a351" Apr 17 23:28:27.376942 containerd[1594]: time="2026-04-17T23:28:27.376758609Z" level=info msg="CreateContainer within sandbox \"ce55e4a2842f0c78d57831dfe00541366bea6bcd37e6108fd7f3fd4103c7a975\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 17 23:28:27.397891 containerd[1594]: time="2026-04-17T23:28:27.397821867Z" level=info msg="CreateContainer within sandbox \"ce55e4a2842f0c78d57831dfe00541366bea6bcd37e6108fd7f3fd4103c7a975\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"105efcbf683e4a14e917b4a1be49261261853610a4a1d6412912338fcaeb6d86\"" Apr 17 23:28:27.398753 containerd[1594]: time="2026-04-17T23:28:27.398712983Z" level=info msg="StartContainer for \"105efcbf683e4a14e917b4a1be49261261853610a4a1d6412912338fcaeb6d86\"" Apr 17 23:28:27.469433 containerd[1594]: time="2026-04-17T23:28:27.469368143Z" level=info msg="StartContainer for \"105efcbf683e4a14e917b4a1be49261261853610a4a1d6412912338fcaeb6d86\" returns successfully" Apr 17 23:28:27.933196 containerd[1594]: time="2026-04-17T23:28:27.932509538Z" level=info msg="shim disconnected" id=40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75 namespace=k8s.io Apr 17 23:28:27.933492 containerd[1594]: time="2026-04-17T23:28:27.933298170Z" level=warning msg="cleaning up after shim disconnected" id=40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75 namespace=k8s.io Apr 17 23:28:27.934208 containerd[1594]: time="2026-04-17T23:28:27.933319531Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 17 23:28:28.018044 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75-rootfs.mount: Deactivated successfully. Apr 17 23:28:28.383068 kubelet[2680]: I0417 23:28:28.382961 2680 scope.go:117] "RemoveContainer" containerID="40e53a4c18a8242eed7fc783d31f2e6464cba14b413177ec0fe618dfbeb0cd75" Apr 17 23:28:28.387241 containerd[1594]: time="2026-04-17T23:28:28.387206195Z" level=info msg="CreateContainer within sandbox \"dca3a8ba705d3cf59da649c4b608edf73517c9030022b715a26c8a6d086a2fed\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 17 23:28:28.403512 containerd[1594]: time="2026-04-17T23:28:28.403462837Z" level=info msg="CreateContainer within sandbox \"dca3a8ba705d3cf59da649c4b608edf73517c9030022b715a26c8a6d086a2fed\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"f7f44bd1b4e19adda43eb7b389f943898450de21e7ab639c07487bfce37d3fef\"" Apr 17 23:28:28.405187 containerd[1594]: time="2026-04-17T23:28:28.404336072Z" level=info msg="StartContainer for \"f7f44bd1b4e19adda43eb7b389f943898450de21e7ab639c07487bfce37d3fef\"" Apr 17 23:28:28.496675 containerd[1594]: time="2026-04-17T23:28:28.496621520Z" level=info msg="StartContainer for \"f7f44bd1b4e19adda43eb7b389f943898450de21e7ab639c07487bfce37d3fef\" returns successfully" Apr 17 23:28:29.740275 kubelet[2680]: E0417 23:28:29.740016 2680 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:33274->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-3-6-n-8954fb58c4.18a748b01f54796a kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-3-6-n-8954fb58c4,UID:c6eb327a67cbb76d10da0426ea735d46,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-3-6-n-8954fb58c4,},FirstTimestamp:2026-04-17 23:28:19.298023786 +0000 UTC m=+108.509018119,LastTimestamp:2026-04-17 23:28:19.298023786 +0000 UTC m=+108.509018119,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-6-n-8954fb58c4,}" Apr 17 23:28:30.927698 containerd[1594]: time="2026-04-17T23:28:30.927642940Z" level=info msg="StopPodSandbox for \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\"" Apr 17 23:28:30.928316 containerd[1594]: time="2026-04-17T23:28:30.927811827Z" level=info msg="TearDown network for sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" successfully" Apr 17 23:28:30.928316 containerd[1594]: time="2026-04-17T23:28:30.927834187Z" level=info msg="StopPodSandbox for \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" returns successfully" Apr 17 23:28:30.928732 containerd[1594]: time="2026-04-17T23:28:30.928708500Z" level=info msg="RemovePodSandbox for \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\"" Apr 17 23:28:30.928790 containerd[1594]: time="2026-04-17T23:28:30.928739141Z" level=info msg="Forcibly stopping sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\"" Apr 17 23:28:30.928790 containerd[1594]: time="2026-04-17T23:28:30.928785903Z" level=info msg="TearDown network for sandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" successfully" Apr 17 23:28:30.932922 containerd[1594]: time="2026-04-17T23:28:30.932838694Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 17 23:28:30.933043 containerd[1594]: time="2026-04-17T23:28:30.932954818Z" level=info msg="RemovePodSandbox \"7d7d8b48eacb8ef5e1f1e334a6d0e1a214f6f2f4088a0cfcca96f30572bfaa13\" returns successfully" Apr 17 23:28:30.933992 containerd[1594]: time="2026-04-17T23:28:30.933764048Z" level=info msg="StopPodSandbox for \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\"" Apr 17 23:28:30.933992 containerd[1594]: time="2026-04-17T23:28:30.933883212Z" level=info msg="TearDown network for sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" successfully" Apr 17 23:28:30.933992 containerd[1594]: time="2026-04-17T23:28:30.933904213Z" level=info msg="StopPodSandbox for \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" returns successfully" Apr 17 23:28:30.934559 containerd[1594]: time="2026-04-17T23:28:30.934449673Z" level=info msg="RemovePodSandbox for \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\"" Apr 17 23:28:30.934559 containerd[1594]: time="2026-04-17T23:28:30.934528596Z" level=info msg="Forcibly stopping sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\"" Apr 17 23:28:30.934711 containerd[1594]: time="2026-04-17T23:28:30.934670042Z" level=info msg="TearDown network for sandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" successfully" Apr 17 23:28:30.938532 containerd[1594]: time="2026-04-17T23:28:30.938500544Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 17 23:28:30.938799 containerd[1594]: time="2026-04-17T23:28:30.938670350Z" level=info msg="RemovePodSandbox \"694d828c2d583071c5a0d408b1d26ec2463b65a130819fa7e76488b853895e9e\" returns successfully"