Apr 23 23:57:57.874096 kernel: Linux version 6.12.81-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu Apr 23 22:08:58 -00 2026 Apr 23 23:57:57.874115 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=35bf60e399c7fbdab9d27e362bd719e7cadd795a3fa26a4f30de01ccc70fba7e Apr 23 23:57:57.874122 kernel: BIOS-provided physical RAM map: Apr 23 23:57:57.874127 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Apr 23 23:57:57.874135 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ed3efff] usable Apr 23 23:57:57.874139 kernel: BIOS-e820: [mem 0x000000007ed3f000-0x000000007edfffff] reserved Apr 23 23:57:57.874145 kernel: BIOS-e820: [mem 0x000000007ee00000-0x000000007f8ecfff] usable Apr 23 23:57:57.874149 kernel: BIOS-e820: [mem 0x000000007f8ed000-0x000000007fb6cfff] reserved Apr 23 23:57:57.874154 kernel: BIOS-e820: [mem 0x000000007fb6d000-0x000000007fb7efff] ACPI data Apr 23 23:57:57.874159 kernel: BIOS-e820: [mem 0x000000007fb7f000-0x000000007fbfefff] ACPI NVS Apr 23 23:57:57.874164 kernel: BIOS-e820: [mem 0x000000007fbff000-0x000000007ff7bfff] usable Apr 23 23:57:57.874168 kernel: BIOS-e820: [mem 0x000000007ff7c000-0x000000007fffffff] reserved Apr 23 23:57:57.874173 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Apr 23 23:57:57.874180 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 23 23:57:57.874185 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Apr 23 23:57:57.874190 kernel: BIOS-e820: [mem 0x0000000100000000-0x0000000179ffffff] usable Apr 23 23:57:57.874195 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Apr 23 23:57:57.874200 kernel: NX (Execute Disable) protection: active Apr 23 23:57:57.874209 kernel: APIC: Static calls initialized Apr 23 23:57:57.874214 kernel: e820: update [mem 0x7dfab018-0x7dfb4a57] usable ==> usable Apr 23 23:57:57.874219 kernel: e820: update [mem 0x7df6f018-0x7dfaa657] usable ==> usable Apr 23 23:57:57.874224 kernel: e820: update [mem 0x7dc01018-0x7dc3c657] usable ==> usable Apr 23 23:57:57.874229 kernel: extended physical RAM map: Apr 23 23:57:57.874234 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Apr 23 23:57:57.874239 kernel: reserve setup_data: [mem 0x0000000000100000-0x000000007dc01017] usable Apr 23 23:57:57.874244 kernel: reserve setup_data: [mem 0x000000007dc01018-0x000000007dc3c657] usable Apr 23 23:57:57.874249 kernel: reserve setup_data: [mem 0x000000007dc3c658-0x000000007df6f017] usable Apr 23 23:57:57.874254 kernel: reserve setup_data: [mem 0x000000007df6f018-0x000000007dfaa657] usable Apr 23 23:57:57.874259 kernel: reserve setup_data: [mem 0x000000007dfaa658-0x000000007dfab017] usable Apr 23 23:57:57.874266 kernel: reserve setup_data: [mem 0x000000007dfab018-0x000000007dfb4a57] usable Apr 23 23:57:57.874271 kernel: reserve setup_data: [mem 0x000000007dfb4a58-0x000000007ed3efff] usable Apr 23 23:57:57.874276 kernel: reserve setup_data: [mem 0x000000007ed3f000-0x000000007edfffff] reserved Apr 23 23:57:57.874280 kernel: reserve setup_data: [mem 0x000000007ee00000-0x000000007f8ecfff] usable Apr 23 23:57:57.874285 kernel: reserve setup_data: [mem 0x000000007f8ed000-0x000000007fb6cfff] reserved Apr 23 23:57:57.874290 kernel: reserve setup_data: [mem 0x000000007fb6d000-0x000000007fb7efff] ACPI data Apr 23 23:57:57.874295 kernel: reserve setup_data: [mem 0x000000007fb7f000-0x000000007fbfefff] ACPI NVS Apr 23 23:57:57.874300 kernel: reserve setup_data: [mem 0x000000007fbff000-0x000000007ff7bfff] usable Apr 23 23:57:57.874305 kernel: reserve setup_data: [mem 0x000000007ff7c000-0x000000007fffffff] reserved Apr 23 23:57:57.874309 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Apr 23 23:57:57.874315 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Apr 23 23:57:57.874324 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Apr 23 23:57:57.874329 kernel: reserve setup_data: [mem 0x0000000100000000-0x0000000179ffffff] usable Apr 23 23:57:57.874334 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Apr 23 23:57:57.874340 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Apr 23 23:57:57.874346 kernel: efi: SMBIOS=0x7f988000 SMBIOS 3.0=0x7f986000 ACPI=0x7fb7e000 ACPI 2.0=0x7fb7e014 MEMATTR=0x7e01b198 RNG=0x7fb73018 Apr 23 23:57:57.874353 kernel: random: crng init done Apr 23 23:57:57.874359 kernel: efi: Remove mem137: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Apr 23 23:57:57.874364 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Apr 23 23:57:57.874369 kernel: secureboot: Secure boot disabled Apr 23 23:57:57.874374 kernel: SMBIOS 3.0.0 present. Apr 23 23:57:57.874379 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Apr 23 23:57:57.874384 kernel: DMI: Memory slots populated: 1/1 Apr 23 23:57:57.874389 kernel: Hypervisor detected: KVM Apr 23 23:57:57.874394 kernel: last_pfn = 0x7ff7c max_arch_pfn = 0x10000000000 Apr 23 23:57:57.874399 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 23 23:57:57.874404 kernel: kvm-clock: using sched offset of 13939367947 cycles Apr 23 23:57:57.874412 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 23 23:57:57.874417 kernel: tsc: Detected 2396.398 MHz processor Apr 23 23:57:57.874423 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 23 23:57:57.874428 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 23 23:57:57.874433 kernel: last_pfn = 0x17a000 max_arch_pfn = 0x10000000000 Apr 23 23:57:57.874438 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Apr 23 23:57:57.874443 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 23 23:57:57.874449 kernel: last_pfn = 0x7ff7c max_arch_pfn = 0x10000000000 Apr 23 23:57:57.874454 kernel: Using GB pages for direct mapping Apr 23 23:57:57.874462 kernel: ACPI: Early table checksum verification disabled Apr 23 23:57:57.874467 kernel: ACPI: RSDP 0x000000007FB7E014 000024 (v02 BOCHS ) Apr 23 23:57:57.874472 kernel: ACPI: XSDT 0x000000007FB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Apr 23 23:57:57.874477 kernel: ACPI: FACP 0x000000007FB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874483 kernel: ACPI: DSDT 0x000000007FB7A000 002443 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874513 kernel: ACPI: FACS 0x000000007FBDD000 000040 Apr 23 23:57:57.874519 kernel: ACPI: APIC 0x000000007FB78000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874524 kernel: ACPI: HPET 0x000000007FB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874529 kernel: ACPI: MCFG 0x000000007FB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874537 kernel: ACPI: WAET 0x000000007FB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 23 23:57:57.874542 kernel: ACPI: BGRT 0x000000007FB74000 000038 (v01 INTEL EDK2 00000002 01000013) Apr 23 23:57:57.874547 kernel: ACPI: Reserving FACP table memory at [mem 0x7fb79000-0x7fb790f3] Apr 23 23:57:57.874553 kernel: ACPI: Reserving DSDT table memory at [mem 0x7fb7a000-0x7fb7c442] Apr 23 23:57:57.874559 kernel: ACPI: Reserving FACS table memory at [mem 0x7fbdd000-0x7fbdd03f] Apr 23 23:57:57.874564 kernel: ACPI: Reserving APIC table memory at [mem 0x7fb78000-0x7fb7807f] Apr 23 23:57:57.874569 kernel: ACPI: Reserving HPET table memory at [mem 0x7fb77000-0x7fb77037] Apr 23 23:57:57.874574 kernel: ACPI: Reserving MCFG table memory at [mem 0x7fb76000-0x7fb7603b] Apr 23 23:57:57.874580 kernel: ACPI: Reserving WAET table memory at [mem 0x7fb75000-0x7fb75027] Apr 23 23:57:57.874587 kernel: ACPI: Reserving BGRT table memory at [mem 0x7fb74000-0x7fb74037] Apr 23 23:57:57.874592 kernel: No NUMA configuration found Apr 23 23:57:57.874598 kernel: Faking a node at [mem 0x0000000000000000-0x0000000179ffffff] Apr 23 23:57:57.874603 kernel: NODE_DATA(0) allocated [mem 0x179ff8dc0-0x179ffffff] Apr 23 23:57:57.874608 kernel: Zone ranges: Apr 23 23:57:57.874613 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 23 23:57:57.874619 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Apr 23 23:57:57.874624 kernel: Normal [mem 0x0000000100000000-0x0000000179ffffff] Apr 23 23:57:57.874631 kernel: Device empty Apr 23 23:57:57.874639 kernel: Movable zone start for each node Apr 23 23:57:57.874644 kernel: Early memory node ranges Apr 23 23:57:57.874649 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Apr 23 23:57:57.874654 kernel: node 0: [mem 0x0000000000100000-0x000000007ed3efff] Apr 23 23:57:57.874659 kernel: node 0: [mem 0x000000007ee00000-0x000000007f8ecfff] Apr 23 23:57:57.874665 kernel: node 0: [mem 0x000000007fbff000-0x000000007ff7bfff] Apr 23 23:57:57.874670 kernel: node 0: [mem 0x0000000100000000-0x0000000179ffffff] Apr 23 23:57:57.874675 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x0000000179ffffff] Apr 23 23:57:57.874680 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 23 23:57:57.874685 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Apr 23 23:57:57.874693 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Apr 23 23:57:57.874698 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Apr 23 23:57:57.874703 kernel: On node 0, zone Normal: 132 pages in unavailable ranges Apr 23 23:57:57.874709 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Apr 23 23:57:57.874714 kernel: ACPI: PM-Timer IO Port: 0x608 Apr 23 23:57:57.874719 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 23 23:57:57.874724 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Apr 23 23:57:57.874730 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Apr 23 23:57:57.874735 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 23 23:57:57.874742 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 23 23:57:57.874748 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 23 23:57:57.874753 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 23 23:57:57.874758 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 23 23:57:57.874763 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 23 23:57:57.874768 kernel: CPU topo: Max. logical packages: 1 Apr 23 23:57:57.874774 kernel: CPU topo: Max. logical dies: 1 Apr 23 23:57:57.874787 kernel: CPU topo: Max. dies per package: 1 Apr 23 23:57:57.874793 kernel: CPU topo: Max. threads per core: 1 Apr 23 23:57:57.874798 kernel: CPU topo: Num. cores per package: 2 Apr 23 23:57:57.874803 kernel: CPU topo: Num. threads per package: 2 Apr 23 23:57:57.874809 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Apr 23 23:57:57.874816 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 23 23:57:57.874822 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Apr 23 23:57:57.874827 kernel: Booting paravirtualized kernel on KVM Apr 23 23:57:57.874832 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 23 23:57:57.874838 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Apr 23 23:57:57.874846 kernel: percpu: Embedded 60 pages/cpu s207448 r8192 d30120 u1048576 Apr 23 23:57:57.874851 kernel: pcpu-alloc: s207448 r8192 d30120 u1048576 alloc=1*2097152 Apr 23 23:57:57.874857 kernel: pcpu-alloc: [0] 0 1 Apr 23 23:57:57.874862 kernel: kvm-guest: PV spinlocks disabled, no host support Apr 23 23:57:57.874868 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=35bf60e399c7fbdab9d27e362bd719e7cadd795a3fa26a4f30de01ccc70fba7e Apr 23 23:57:57.874874 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 23 23:57:57.874879 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 23 23:57:57.874884 kernel: Fallback order for Node 0: 0 Apr 23 23:57:57.874892 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1022792 Apr 23 23:57:57.874897 kernel: Policy zone: Normal Apr 23 23:57:57.874903 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 23 23:57:57.874908 kernel: software IO TLB: area num 2. Apr 23 23:57:57.874913 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 23 23:57:57.874919 kernel: ftrace: allocating 40126 entries in 157 pages Apr 23 23:57:57.874924 kernel: ftrace: allocated 157 pages with 5 groups Apr 23 23:57:57.874930 kernel: Dynamic Preempt: voluntary Apr 23 23:57:57.874935 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 23 23:57:57.874943 kernel: rcu: RCU event tracing is enabled. Apr 23 23:57:57.874949 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 23 23:57:57.874954 kernel: Trampoline variant of Tasks RCU enabled. Apr 23 23:57:57.874960 kernel: Rude variant of Tasks RCU enabled. Apr 23 23:57:57.874965 kernel: Tracing variant of Tasks RCU enabled. Apr 23 23:57:57.874971 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 23 23:57:57.874976 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 23 23:57:57.874982 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 23 23:57:57.874987 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 23 23:57:57.874995 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 23 23:57:57.875000 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Apr 23 23:57:57.875005 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 23 23:57:57.875035 kernel: Console: colour dummy device 80x25 Apr 23 23:57:57.875040 kernel: printk: legacy console [tty0] enabled Apr 23 23:57:57.875046 kernel: printk: legacy console [ttyS0] enabled Apr 23 23:57:57.875051 kernel: ACPI: Core revision 20240827 Apr 23 23:57:57.875057 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Apr 23 23:57:57.875062 kernel: APIC: Switch to symmetric I/O mode setup Apr 23 23:57:57.875070 kernel: x2apic enabled Apr 23 23:57:57.875075 kernel: APIC: Switched APIC routing to: physical x2apic Apr 23 23:57:57.875081 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Apr 23 23:57:57.875086 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x228aecd6e18, max_idle_ns: 440795270957 ns Apr 23 23:57:57.875092 kernel: Calibrating delay loop (skipped) preset value.. 4792.79 BogoMIPS (lpj=2396398) Apr 23 23:57:57.875097 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Apr 23 23:57:57.875103 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Apr 23 23:57:57.875108 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Apr 23 23:57:57.875114 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 23 23:57:57.875121 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Apr 23 23:57:57.875127 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Apr 23 23:57:57.875132 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Apr 23 23:57:57.875138 kernel: active return thunk: srso_alias_return_thunk Apr 23 23:57:57.875143 kernel: Speculative Return Stack Overflow: Mitigation: Safe RET Apr 23 23:57:57.875148 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Apr 23 23:57:57.875154 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Apr 23 23:57:57.875159 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 23 23:57:57.875165 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 23 23:57:57.875172 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 23 23:57:57.875178 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 23 23:57:57.875183 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 23 23:57:57.875189 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 23 23:57:57.875194 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Apr 23 23:57:57.875200 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 23 23:57:57.875205 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Apr 23 23:57:57.875210 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Apr 23 23:57:57.875216 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Apr 23 23:57:57.875223 kernel: x86/fpu: xstate_offset[9]: 2432, xstate_sizes[9]: 8 Apr 23 23:57:57.875229 kernel: x86/fpu: Enabled xstate features 0x2e7, context size is 2440 bytes, using 'compacted' format. Apr 23 23:57:57.875234 kernel: Freeing SMP alternatives memory: 32K Apr 23 23:57:57.875239 kernel: pid_max: default: 32768 minimum: 301 Apr 23 23:57:57.875245 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Apr 23 23:57:57.875250 kernel: landlock: Up and running. Apr 23 23:57:57.875256 kernel: SELinux: Initializing. Apr 23 23:57:57.875261 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 23 23:57:57.875267 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 23 23:57:57.875275 kernel: smpboot: CPU0: AMD EPYC-Genoa Processor (family: 0x19, model: 0x11, stepping: 0x0) Apr 23 23:57:57.875280 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Apr 23 23:57:57.875285 kernel: ... version: 0 Apr 23 23:57:57.875291 kernel: ... bit width: 48 Apr 23 23:57:57.875296 kernel: ... generic registers: 6 Apr 23 23:57:57.875302 kernel: ... value mask: 0000ffffffffffff Apr 23 23:57:57.875307 kernel: ... max period: 00007fffffffffff Apr 23 23:57:57.875312 kernel: ... fixed-purpose events: 0 Apr 23 23:57:57.875318 kernel: ... event mask: 000000000000003f Apr 23 23:57:57.875325 kernel: signal: max sigframe size: 3376 Apr 23 23:57:57.875331 kernel: rcu: Hierarchical SRCU implementation. Apr 23 23:57:57.875336 kernel: rcu: Max phase no-delay instances is 400. Apr 23 23:57:57.875342 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Apr 23 23:57:57.875347 kernel: smp: Bringing up secondary CPUs ... Apr 23 23:57:57.875352 kernel: smpboot: x86: Booting SMP configuration: Apr 23 23:57:57.875358 kernel: .... node #0, CPUs: #1 Apr 23 23:57:57.875363 kernel: smp: Brought up 1 node, 2 CPUs Apr 23 23:57:57.875369 kernel: smpboot: Total of 2 processors activated (9585.59 BogoMIPS) Apr 23 23:57:57.875376 kernel: Memory: 3813628K/4091168K available (14336K kernel code, 2453K rwdata, 26076K rodata, 46224K init, 2524K bss, 271900K reserved, 0K cma-reserved) Apr 23 23:57:57.875382 kernel: devtmpfs: initialized Apr 23 23:57:57.875387 kernel: x86/mm: Memory block size: 128MB Apr 23 23:57:57.875393 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7fb7f000-0x7fbfefff] (524288 bytes) Apr 23 23:57:57.875398 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 23 23:57:57.875404 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 23 23:57:57.875409 kernel: pinctrl core: initialized pinctrl subsystem Apr 23 23:57:57.875414 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 23 23:57:57.875420 kernel: audit: initializing netlink subsys (disabled) Apr 23 23:57:57.875427 kernel: audit: type=2000 audit(1776988675.532:1): state=initialized audit_enabled=0 res=1 Apr 23 23:57:57.875433 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 23 23:57:57.875438 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 23 23:57:57.875444 kernel: cpuidle: using governor menu Apr 23 23:57:57.875449 kernel: efi: Freeing EFI boot services memory: 34884K Apr 23 23:57:57.875454 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 23 23:57:57.875460 kernel: dca service started, version 1.12.1 Apr 23 23:57:57.875465 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Apr 23 23:57:57.875471 kernel: PCI: Using configuration type 1 for base access Apr 23 23:57:57.875479 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 23 23:57:57.875484 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 23 23:57:57.875500 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 23 23:57:57.875505 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 23 23:57:57.875511 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 23 23:57:57.875516 kernel: ACPI: Added _OSI(Module Device) Apr 23 23:57:57.875521 kernel: ACPI: Added _OSI(Processor Device) Apr 23 23:57:57.875527 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 23 23:57:57.875532 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 23 23:57:57.875540 kernel: ACPI: Interpreter enabled Apr 23 23:57:57.875545 kernel: ACPI: PM: (supports S0 S5) Apr 23 23:57:57.875551 kernel: ACPI: Using IOAPIC for interrupt routing Apr 23 23:57:57.875557 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 23 23:57:57.875563 kernel: PCI: Using E820 reservations for host bridge windows Apr 23 23:57:57.875568 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Apr 23 23:57:57.875574 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 23 23:57:57.875729 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 23 23:57:57.875835 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Apr 23 23:57:57.875934 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Apr 23 23:57:57.875940 kernel: PCI host bridge to bus 0000:00 Apr 23 23:57:57.876057 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 23 23:57:57.876149 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 23 23:57:57.876237 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 23 23:57:57.876325 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xdfffffff window] Apr 23 23:57:57.876416 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Apr 23 23:57:57.876515 kernel: pci_bus 0000:00: root bus resource [mem 0xc000000000-0xc7ffffffff window] Apr 23 23:57:57.876604 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 23 23:57:57.876713 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Apr 23 23:57:57.876824 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint Apr 23 23:57:57.876921 kernel: pci 0000:00:01.0: BAR 0 [mem 0x80000000-0x807fffff pref] Apr 23 23:57:57.877031 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc060500000-0xc060503fff 64bit pref] Apr 23 23:57:57.877129 kernel: pci 0000:00:01.0: BAR 4 [mem 0x8138a000-0x8138afff] Apr 23 23:57:57.877224 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Apr 23 23:57:57.877321 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 23 23:57:57.877428 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.877532 kernel: pci 0000:00:02.0: BAR 0 [mem 0x81389000-0x81389fff] Apr 23 23:57:57.877628 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 23 23:57:57.877728 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Apr 23 23:57:57.877825 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 23 23:57:57.877928 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.878318 kernel: pci 0000:00:02.1: BAR 0 [mem 0x81388000-0x81388fff] Apr 23 23:57:57.878429 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 23 23:57:57.878538 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Apr 23 23:57:57.878641 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.878742 kernel: pci 0000:00:02.2: BAR 0 [mem 0x81387000-0x81387fff] Apr 23 23:57:57.878841 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 23 23:57:57.878936 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Apr 23 23:57:57.879058 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 23 23:57:57.879164 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.879261 kernel: pci 0000:00:02.3: BAR 0 [mem 0x81386000-0x81386fff] Apr 23 23:57:57.879357 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 23 23:57:57.879456 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 23 23:57:57.879568 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.879665 kernel: pci 0000:00:02.4: BAR 0 [mem 0x81385000-0x81385fff] Apr 23 23:57:57.879761 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 23 23:57:57.879856 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Apr 23 23:57:57.879951 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 23 23:57:57.880068 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.880168 kernel: pci 0000:00:02.5: BAR 0 [mem 0x81384000-0x81384fff] Apr 23 23:57:57.880264 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 23 23:57:57.880359 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Apr 23 23:57:57.880454 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 23 23:57:57.880680 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.880783 kernel: pci 0000:00:02.6: BAR 0 [mem 0x81383000-0x81383fff] Apr 23 23:57:57.880878 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 23 23:57:57.880978 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Apr 23 23:57:57.881087 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 23 23:57:57.881190 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.881285 kernel: pci 0000:00:02.7: BAR 0 [mem 0x81382000-0x81382fff] Apr 23 23:57:57.881380 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 23 23:57:57.881476 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Apr 23 23:57:57.881585 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 23 23:57:57.881721 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Apr 23 23:57:57.881818 kernel: pci 0000:00:03.0: BAR 0 [mem 0x81381000-0x81381fff] Apr 23 23:57:57.881913 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 23 23:57:57.882007 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Apr 23 23:57:57.882117 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 23 23:57:57.882219 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Apr 23 23:57:57.882317 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Apr 23 23:57:57.882419 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Apr 23 23:57:57.882522 kernel: pci 0000:00:1f.2: BAR 4 [io 0x6040-0x605f] Apr 23 23:57:57.882617 kernel: pci 0000:00:1f.2: BAR 5 [mem 0x81380000-0x81380fff] Apr 23 23:57:57.882718 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Apr 23 23:57:57.882814 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6000-0x603f] Apr 23 23:57:57.882923 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Apr 23 23:57:57.883046 kernel: pci 0000:01:00.0: BAR 1 [mem 0x81200000-0x81200fff] Apr 23 23:57:57.883148 kernel: pci 0000:01:00.0: BAR 4 [mem 0xc060000000-0xc060003fff 64bit pref] Apr 23 23:57:57.883249 kernel: pci 0000:01:00.0: ROM [mem 0xfff80000-0xffffffff pref] Apr 23 23:57:57.883345 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 23 23:57:57.883450 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 PCIe Endpoint Apr 23 23:57:57.883561 kernel: pci 0000:02:00.0: BAR 0 [mem 0x81100000-0x81103fff 64bit] Apr 23 23:57:57.883660 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 23 23:57:57.883767 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 PCIe Endpoint Apr 23 23:57:57.883875 kernel: pci 0000:03:00.0: BAR 1 [mem 0x81000000-0x81000fff] Apr 23 23:57:57.884194 kernel: pci 0000:03:00.0: BAR 4 [mem 0xc060100000-0xc060103fff 64bit pref] Apr 23 23:57:57.884304 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 23 23:57:57.884414 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 PCIe Endpoint Apr 23 23:57:57.884528 kernel: pci 0000:04:00.0: BAR 4 [mem 0xc060200000-0xc060203fff 64bit pref] Apr 23 23:57:57.884642 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 23 23:57:57.884762 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 PCIe Endpoint Apr 23 23:57:57.884863 kernel: pci 0000:05:00.0: BAR 1 [mem 0x80f00000-0x80f00fff] Apr 23 23:57:57.884963 kernel: pci 0000:05:00.0: BAR 4 [mem 0xc060300000-0xc060303fff 64bit pref] Apr 23 23:57:57.887397 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 23 23:57:57.887646 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 PCIe Endpoint Apr 23 23:57:57.887834 kernel: pci 0000:06:00.0: BAR 1 [mem 0x80e00000-0x80e00fff] Apr 23 23:57:57.887944 kernel: pci 0000:06:00.0: BAR 4 [mem 0xc060400000-0xc060403fff 64bit pref] Apr 23 23:57:57.888058 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 23 23:57:57.888066 kernel: acpiphp: Slot [0] registered Apr 23 23:57:57.888174 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Apr 23 23:57:57.888275 kernel: pci 0000:07:00.0: BAR 1 [mem 0x80c00000-0x80c00fff] Apr 23 23:57:57.888375 kernel: pci 0000:07:00.0: BAR 4 [mem 0xc000000000-0xc000003fff 64bit pref] Apr 23 23:57:57.888475 kernel: pci 0000:07:00.0: ROM [mem 0xfff80000-0xffffffff pref] Apr 23 23:57:57.888583 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 23 23:57:57.888590 kernel: acpiphp: Slot [0-2] registered Apr 23 23:57:57.888685 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 23 23:57:57.888693 kernel: acpiphp: Slot [0-3] registered Apr 23 23:57:57.888788 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 23 23:57:57.888812 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 23 23:57:57.888818 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 23 23:57:57.888824 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 23 23:57:57.888832 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 23 23:57:57.888837 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Apr 23 23:57:57.888843 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Apr 23 23:57:57.888848 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Apr 23 23:57:57.888854 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Apr 23 23:57:57.888860 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Apr 23 23:57:57.888865 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Apr 23 23:57:57.888871 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Apr 23 23:57:57.888877 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Apr 23 23:57:57.888885 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Apr 23 23:57:57.888891 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Apr 23 23:57:57.888899 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Apr 23 23:57:57.888905 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Apr 23 23:57:57.888910 kernel: iommu: Default domain type: Translated Apr 23 23:57:57.888916 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 23 23:57:57.888924 kernel: efivars: Registered efivars operations Apr 23 23:57:57.888930 kernel: PCI: Using ACPI for IRQ routing Apr 23 23:57:57.888935 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 23 23:57:57.888942 kernel: e820: reserve RAM buffer [mem 0x7dc01018-0x7fffffff] Apr 23 23:57:57.888947 kernel: e820: reserve RAM buffer [mem 0x7df6f018-0x7fffffff] Apr 23 23:57:57.888953 kernel: e820: reserve RAM buffer [mem 0x7dfab018-0x7fffffff] Apr 23 23:57:57.888959 kernel: e820: reserve RAM buffer [mem 0x7ed3f000-0x7fffffff] Apr 23 23:57:57.888964 kernel: e820: reserve RAM buffer [mem 0x7f8ed000-0x7fffffff] Apr 23 23:57:57.888970 kernel: e820: reserve RAM buffer [mem 0x7ff7c000-0x7fffffff] Apr 23 23:57:57.888978 kernel: e820: reserve RAM buffer [mem 0x17a000000-0x17bffffff] Apr 23 23:57:57.889087 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Apr 23 23:57:57.889183 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Apr 23 23:57:57.889278 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 23 23:57:57.889285 kernel: vgaarb: loaded Apr 23 23:57:57.889292 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Apr 23 23:57:57.889297 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Apr 23 23:57:57.889303 kernel: clocksource: Switched to clocksource kvm-clock Apr 23 23:57:57.889311 kernel: VFS: Disk quotas dquot_6.6.0 Apr 23 23:57:57.889317 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 23 23:57:57.889323 kernel: pnp: PnP ACPI init Apr 23 23:57:57.889429 kernel: system 00:04: [mem 0xe0000000-0xefffffff window] has been reserved Apr 23 23:57:57.889437 kernel: pnp: PnP ACPI: found 5 devices Apr 23 23:57:57.889443 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 23 23:57:57.889449 kernel: NET: Registered PF_INET protocol family Apr 23 23:57:57.889454 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 23 23:57:57.889462 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 23 23:57:57.889468 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 23 23:57:57.889474 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 23 23:57:57.889480 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 23 23:57:57.889486 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 23 23:57:57.889501 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 23 23:57:57.889506 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 23 23:57:57.889512 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 23 23:57:57.889518 kernel: NET: Registered PF_XDP protocol family Apr 23 23:57:57.889626 kernel: pci 0000:01:00.0: ROM [mem 0xfff80000-0xffffffff pref]: can't claim; no compatible bridge window Apr 23 23:57:57.889729 kernel: pci 0000:07:00.0: ROM [mem 0xfff80000-0xffffffff pref]: can't claim; no compatible bridge window Apr 23 23:57:57.889826 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Apr 23 23:57:57.889925 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Apr 23 23:57:57.892069 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Apr 23 23:57:57.892183 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff]: assigned Apr 23 23:57:57.892284 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff]: assigned Apr 23 23:57:57.892406 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff]: assigned Apr 23 23:57:57.892556 kernel: pci 0000:01:00.0: ROM [mem 0x81280000-0x812fffff pref]: assigned Apr 23 23:57:57.892671 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Apr 23 23:57:57.892770 kernel: pci 0000:00:02.0: bridge window [mem 0x81200000-0x812fffff] Apr 23 23:57:57.892867 kernel: pci 0000:00:02.0: bridge window [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 23 23:57:57.892963 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Apr 23 23:57:57.897146 kernel: pci 0000:00:02.1: bridge window [mem 0x81100000-0x811fffff] Apr 23 23:57:57.897259 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Apr 23 23:57:57.897358 kernel: pci 0000:00:02.2: bridge window [mem 0x81000000-0x810fffff] Apr 23 23:57:57.897473 kernel: pci 0000:00:02.2: bridge window [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 23 23:57:57.897606 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Apr 23 23:57:57.897703 kernel: pci 0000:00:02.3: bridge window [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 23 23:57:57.897799 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Apr 23 23:57:57.897895 kernel: pci 0000:00:02.4: bridge window [mem 0x80f00000-0x80ffffff] Apr 23 23:57:57.897990 kernel: pci 0000:00:02.4: bridge window [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 23 23:57:57.899078 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Apr 23 23:57:57.899184 kernel: pci 0000:00:02.5: bridge window [mem 0x80e00000-0x80efffff] Apr 23 23:57:57.899281 kernel: pci 0000:00:02.5: bridge window [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 23 23:57:57.899384 kernel: pci 0000:07:00.0: ROM [mem 0x80c80000-0x80cfffff pref]: assigned Apr 23 23:57:57.899502 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Apr 23 23:57:57.899628 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Apr 23 23:57:57.899728 kernel: pci 0000:00:02.6: bridge window [mem 0x80c00000-0x80dfffff] Apr 23 23:57:57.899823 kernel: pci 0000:00:02.6: bridge window [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 23 23:57:57.899922 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Apr 23 23:57:57.900085 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Apr 23 23:57:57.900194 kernel: pci 0000:00:02.7: bridge window [mem 0x80a00000-0x80bfffff] Apr 23 23:57:57.900293 kernel: pci 0000:00:02.7: bridge window [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 23 23:57:57.900391 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Apr 23 23:57:57.900499 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Apr 23 23:57:57.900596 kernel: pci 0000:00:03.0: bridge window [mem 0x80800000-0x809fffff] Apr 23 23:57:57.900692 kernel: pci 0000:00:03.0: bridge window [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 23 23:57:57.900787 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 23 23:57:57.900877 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 23 23:57:57.900970 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 23 23:57:57.901116 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xdfffffff window] Apr 23 23:57:57.901208 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Apr 23 23:57:57.901297 kernel: pci_bus 0000:00: resource 9 [mem 0xc000000000-0xc7ffffffff window] Apr 23 23:57:57.901399 kernel: pci_bus 0000:01: resource 1 [mem 0x81200000-0x812fffff] Apr 23 23:57:57.901507 kernel: pci_bus 0000:01: resource 2 [mem 0xc060000000-0xc0600fffff 64bit pref] Apr 23 23:57:57.901615 kernel: pci_bus 0000:02: resource 1 [mem 0x81100000-0x811fffff] Apr 23 23:57:57.901720 kernel: pci_bus 0000:03: resource 1 [mem 0x81000000-0x810fffff] Apr 23 23:57:57.901815 kernel: pci_bus 0000:03: resource 2 [mem 0xc060100000-0xc0601fffff 64bit pref] Apr 23 23:57:57.901915 kernel: pci_bus 0000:04: resource 2 [mem 0xc060200000-0xc0602fffff 64bit pref] Apr 23 23:57:57.902043 kernel: pci_bus 0000:05: resource 1 [mem 0x80f00000-0x80ffffff] Apr 23 23:57:57.902140 kernel: pci_bus 0000:05: resource 2 [mem 0xc060300000-0xc0603fffff 64bit pref] Apr 23 23:57:57.902242 kernel: pci_bus 0000:06: resource 1 [mem 0x80e00000-0x80efffff] Apr 23 23:57:57.902339 kernel: pci_bus 0000:06: resource 2 [mem 0xc060400000-0xc0604fffff 64bit pref] Apr 23 23:57:57.902438 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Apr 23 23:57:57.902544 kernel: pci_bus 0000:07: resource 1 [mem 0x80c00000-0x80dfffff] Apr 23 23:57:57.902648 kernel: pci_bus 0000:07: resource 2 [mem 0xc000000000-0xc01fffffff 64bit pref] Apr 23 23:57:57.902768 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Apr 23 23:57:57.902863 kernel: pci_bus 0000:08: resource 1 [mem 0x80a00000-0x80bfffff] Apr 23 23:57:57.902959 kernel: pci_bus 0000:08: resource 2 [mem 0xc020000000-0xc03fffffff 64bit pref] Apr 23 23:57:57.903078 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Apr 23 23:57:57.903172 kernel: pci_bus 0000:09: resource 1 [mem 0x80800000-0x809fffff] Apr 23 23:57:57.903267 kernel: pci_bus 0000:09: resource 2 [mem 0xc040000000-0xc05fffffff 64bit pref] Apr 23 23:57:57.903275 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Apr 23 23:57:57.903281 kernel: PCI: CLS 0 bytes, default 64 Apr 23 23:57:57.903288 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Apr 23 23:57:57.903293 kernel: software IO TLB: mapped [mem 0x0000000077ffd000-0x000000007bffd000] (64MB) Apr 23 23:57:57.903304 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x228aecd6e18, max_idle_ns: 440795270957 ns Apr 23 23:57:57.903310 kernel: Initialise system trusted keyrings Apr 23 23:57:57.903316 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 23 23:57:57.903322 kernel: Key type asymmetric registered Apr 23 23:57:57.903328 kernel: Asymmetric key parser 'x509' registered Apr 23 23:57:57.903334 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 23 23:57:57.903340 kernel: io scheduler mq-deadline registered Apr 23 23:57:57.903345 kernel: io scheduler kyber registered Apr 23 23:57:57.903351 kernel: io scheduler bfq registered Apr 23 23:57:57.903453 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Apr 23 23:57:57.903562 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Apr 23 23:57:57.903661 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Apr 23 23:57:57.903757 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Apr 23 23:57:57.903854 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Apr 23 23:57:57.903952 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Apr 23 23:57:57.904060 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Apr 23 23:57:57.904158 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Apr 23 23:57:57.904256 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Apr 23 23:57:57.904353 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Apr 23 23:57:57.904449 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Apr 23 23:57:57.904557 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Apr 23 23:57:57.904658 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Apr 23 23:57:57.904764 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Apr 23 23:57:57.904883 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Apr 23 23:57:57.905003 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Apr 23 23:57:57.905040 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Apr 23 23:57:57.905170 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Apr 23 23:57:57.905297 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Apr 23 23:57:57.905309 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 23 23:57:57.905317 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Apr 23 23:57:57.905325 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 23 23:57:57.905337 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 23 23:57:57.905345 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 23 23:57:57.905353 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 23 23:57:57.905361 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 23 23:57:57.905502 kernel: rtc_cmos 00:03: RTC can wake from S4 Apr 23 23:57:57.905516 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 23 23:57:57.905625 kernel: rtc_cmos 00:03: registered as rtc0 Apr 23 23:57:57.905719 kernel: rtc_cmos 00:03: setting system clock to 2026-04-23T23:57:57 UTC (1776988677) Apr 23 23:57:57.905814 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Apr 23 23:57:57.905822 kernel: amd_pstate: The CPPC feature is supported but currently disabled by the BIOS. Please enable it if your BIOS has the CPPC option. Apr 23 23:57:57.905828 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Apr 23 23:57:57.905834 kernel: efifb: probing for efifb Apr 23 23:57:57.905840 kernel: efifb: framebuffer at 0x80000000, using 4000k, total 4000k Apr 23 23:57:57.905846 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Apr 23 23:57:57.905852 kernel: efifb: scrolling: redraw Apr 23 23:57:57.905860 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 23 23:57:57.905868 kernel: Console: switching to colour frame buffer device 160x50 Apr 23 23:57:57.905880 kernel: fb0: EFI VGA frame buffer device Apr 23 23:57:57.905889 kernel: pstore: Using crash dump compression: deflate Apr 23 23:57:57.905896 kernel: pstore: Registered efi_pstore as persistent store backend Apr 23 23:57:57.905904 kernel: NET: Registered PF_INET6 protocol family Apr 23 23:57:57.905911 kernel: Segment Routing with IPv6 Apr 23 23:57:57.905918 kernel: In-situ OAM (IOAM) with IPv6 Apr 23 23:57:57.905926 kernel: NET: Registered PF_PACKET protocol family Apr 23 23:57:57.905934 kernel: Key type dns_resolver registered Apr 23 23:57:57.905942 kernel: IPI shorthand broadcast: enabled Apr 23 23:57:57.905952 kernel: sched_clock: Marking stable (2882011412, 272062198)->(3223607584, -69533974) Apr 23 23:57:57.905961 kernel: registered taskstats version 1 Apr 23 23:57:57.905969 kernel: Loading compiled-in X.509 certificates Apr 23 23:57:57.905979 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.81-flatcar: 09f9b319c99eb3f54e68ef799fdb2bce5b238ec0' Apr 23 23:57:57.905988 kernel: Demotion targets for Node 0: null Apr 23 23:57:57.905996 kernel: Key type .fscrypt registered Apr 23 23:57:57.906003 kernel: Key type fscrypt-provisioning registered Apr 23 23:57:57.906026 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 23 23:57:57.906036 kernel: ima: Allocated hash algorithm: sha1 Apr 23 23:57:57.906046 kernel: ima: No architecture policies found Apr 23 23:57:57.906053 kernel: clk: Disabling unused clocks Apr 23 23:57:57.906062 kernel: Warning: unable to open an initial console. Apr 23 23:57:57.906070 kernel: Freeing unused kernel image (initmem) memory: 46224K Apr 23 23:57:57.906079 kernel: Write protecting the kernel read-only data: 40960k Apr 23 23:57:57.906086 kernel: Freeing unused kernel image (rodata/data gap) memory: 548K Apr 23 23:57:57.906092 kernel: Run /init as init process Apr 23 23:57:57.906098 kernel: with arguments: Apr 23 23:57:57.906104 kernel: /init Apr 23 23:57:57.906115 kernel: with environment: Apr 23 23:57:57.906124 kernel: HOME=/ Apr 23 23:57:57.906131 kernel: TERM=linux Apr 23 23:57:57.906140 systemd[1]: Successfully made /usr/ read-only. Apr 23 23:57:57.906151 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 23 23:57:57.906161 systemd[1]: Detected virtualization kvm. Apr 23 23:57:57.906169 systemd[1]: Detected architecture x86-64. Apr 23 23:57:57.906179 systemd[1]: Running in initrd. Apr 23 23:57:57.906188 systemd[1]: No hostname configured, using default hostname. Apr 23 23:57:57.906196 systemd[1]: Hostname set to . Apr 23 23:57:57.906205 systemd[1]: Initializing machine ID from VM UUID. Apr 23 23:57:57.906213 systemd[1]: Queued start job for default target initrd.target. Apr 23 23:57:57.906221 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:57:57.906230 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:57:57.906240 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 23 23:57:57.906252 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 23 23:57:57.906261 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 23 23:57:57.906269 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 23 23:57:57.906276 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 23 23:57:57.906282 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 23 23:57:57.906288 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:57:57.906294 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:57:57.906303 systemd[1]: Reached target paths.target - Path Units. Apr 23 23:57:57.906309 systemd[1]: Reached target slices.target - Slice Units. Apr 23 23:57:57.906316 systemd[1]: Reached target swap.target - Swaps. Apr 23 23:57:57.906322 systemd[1]: Reached target timers.target - Timer Units. Apr 23 23:57:57.906328 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 23 23:57:57.906334 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 23 23:57:57.906340 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 23 23:57:57.906346 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Apr 23 23:57:57.906352 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:57:57.906361 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 23 23:57:57.906367 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:57:57.906372 systemd[1]: Reached target sockets.target - Socket Units. Apr 23 23:57:57.906379 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 23 23:57:57.906386 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 23 23:57:57.906395 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 23 23:57:57.906404 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Apr 23 23:57:57.906412 systemd[1]: Starting systemd-fsck-usr.service... Apr 23 23:57:57.906423 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 23 23:57:57.906431 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 23 23:57:57.906439 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:57:57.906447 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 23 23:57:57.906456 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:57:57.906467 systemd[1]: Finished systemd-fsck-usr.service. Apr 23 23:57:57.906476 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 23 23:57:57.906525 systemd-journald[199]: Collecting audit messages is disabled. Apr 23 23:57:57.906547 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 23 23:57:57.906556 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 23 23:57:57.906564 kernel: Bridge firewalling registered Apr 23 23:57:57.906572 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 23 23:57:57.906581 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:57:57.906591 systemd-journald[199]: Journal started Apr 23 23:57:57.906611 systemd-journald[199]: Runtime Journal (/run/log/journal/25e6482849544a6e86d55309efe6e421) is 8M, max 76.1M, 68.1M free. Apr 23 23:57:57.856451 systemd-modules-load[201]: Inserted module 'overlay' Apr 23 23:57:57.900063 systemd-modules-load[201]: Inserted module 'br_netfilter' Apr 23 23:57:57.910514 systemd[1]: Started systemd-journald.service - Journal Service. Apr 23 23:57:57.913752 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 23 23:57:57.916100 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:57:57.918483 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 23 23:57:57.926915 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 23 23:57:57.939483 systemd-tmpfiles[220]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Apr 23 23:57:57.940981 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:57:57.942635 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:57:57.946382 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:57:57.949345 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 23 23:57:57.950811 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 23 23:57:57.953105 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 23 23:57:57.968530 dracut-cmdline[237]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=35bf60e399c7fbdab9d27e362bd719e7cadd795a3fa26a4f30de01ccc70fba7e Apr 23 23:57:57.983589 systemd-resolved[235]: Positive Trust Anchors: Apr 23 23:57:57.983605 systemd-resolved[235]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 23 23:57:57.983626 systemd-resolved[235]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 23 23:57:57.987568 systemd-resolved[235]: Defaulting to hostname 'linux'. Apr 23 23:57:57.988518 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 23 23:57:57.989393 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:57:58.046058 kernel: SCSI subsystem initialized Apr 23 23:57:58.054038 kernel: Loading iSCSI transport class v2.0-870. Apr 23 23:57:58.063035 kernel: iscsi: registered transport (tcp) Apr 23 23:57:58.080253 kernel: iscsi: registered transport (qla4xxx) Apr 23 23:57:58.080283 kernel: QLogic iSCSI HBA Driver Apr 23 23:57:58.097391 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 23 23:57:58.112443 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:57:58.114319 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 23 23:57:58.158388 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 23 23:57:58.160084 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 23 23:57:58.207058 kernel: raid6: avx512x4 gen() 47929 MB/s Apr 23 23:57:58.225083 kernel: raid6: avx512x2 gen() 51446 MB/s Apr 23 23:57:58.243057 kernel: raid6: avx512x1 gen() 48358 MB/s Apr 23 23:57:58.261060 kernel: raid6: avx2x4 gen() 52716 MB/s Apr 23 23:57:58.279060 kernel: raid6: avx2x2 gen() 55675 MB/s Apr 23 23:57:58.298086 kernel: raid6: avx2x1 gen() 42205 MB/s Apr 23 23:57:58.298132 kernel: raid6: using algorithm avx2x2 gen() 55675 MB/s Apr 23 23:57:58.318138 kernel: raid6: .... xor() 36038 MB/s, rmw enabled Apr 23 23:57:58.318194 kernel: raid6: using avx512x2 recovery algorithm Apr 23 23:57:58.358051 kernel: xor: automatically using best checksumming function avx Apr 23 23:57:58.465050 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 23 23:57:58.474384 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 23 23:57:58.478809 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:57:58.500557 systemd-udevd[448]: Using default interface naming scheme 'v255'. Apr 23 23:57:58.505740 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:57:58.511130 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 23 23:57:58.537873 dracut-pre-trigger[462]: rd.md=0: removing MD RAID activation Apr 23 23:57:58.579589 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 23 23:57:58.583318 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 23 23:57:58.657597 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:57:58.665245 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 23 23:57:58.739722 kernel: ACPI: bus type USB registered Apr 23 23:57:58.739791 kernel: usbcore: registered new interface driver usbfs Apr 23 23:57:58.740133 kernel: usbcore: registered new interface driver hub Apr 23 23:57:58.740145 kernel: virtio_scsi virtio5: 2/0/0 default/read/poll queues Apr 23 23:57:58.744969 kernel: usbcore: registered new device driver usb Apr 23 23:57:58.760033 kernel: scsi host0: Virtio SCSI HBA Apr 23 23:57:58.781831 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 23 23:57:58.782078 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Apr 23 23:57:58.785039 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Apr 23 23:57:58.788183 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Apr 23 23:57:58.788332 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Apr 23 23:57:58.792349 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Apr 23 23:57:58.798042 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Apr 23 23:57:58.806164 kernel: cryptd: max_cpu_qlen set to 1000 Apr 23 23:57:58.806201 kernel: hub 1-0:1.0: USB hub found Apr 23 23:57:58.806708 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:57:58.807571 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:57:58.815081 kernel: hub 1-0:1.0: 4 ports detected Apr 23 23:57:58.815275 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Apr 23 23:57:58.811148 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:57:58.812129 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:57:58.825610 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Apr 23 23:57:58.828927 kernel: hub 2-0:1.0: USB hub found Apr 23 23:57:58.841354 kernel: hub 2-0:1.0: 4 ports detected Apr 23 23:57:58.841565 kernel: libata version 3.00 loaded. Apr 23 23:57:58.850239 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:57:58.852103 kernel: ahci 0000:00:1f.2: version 3.0 Apr 23 23:57:58.852268 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Apr 23 23:57:58.852433 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:57:58.852540 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:57:58.857469 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Apr 23 23:57:58.857630 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Apr 23 23:57:58.857747 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Apr 23 23:57:58.857861 kernel: scsi host1: ahci Apr 23 23:57:58.854371 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:57:58.863151 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:57:58.863688 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:57:58.868105 kernel: sd 0:0:0:0: Power-on or device reset occurred Apr 23 23:57:58.868318 kernel: scsi host2: ahci Apr 23 23:57:58.871285 kernel: sd 0:0:0:0: [sda] 160006144 512-byte logical blocks: (81.9 GB/76.3 GiB) Apr 23 23:57:58.871444 kernel: scsi host3: ahci Apr 23 23:57:58.874430 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 23 23:57:58.874598 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Apr 23 23:57:58.874726 kernel: AES CTR mode by8 optimization enabled Apr 23 23:57:58.878059 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Apr 23 23:57:58.878219 kernel: scsi host4: ahci Apr 23 23:57:58.880098 kernel: scsi host5: ahci Apr 23 23:57:58.883043 kernel: scsi host6: ahci Apr 23 23:57:58.890045 kernel: ata1: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380100 irq 51 lpm-pol 1 Apr 23 23:57:58.890073 kernel: ata2: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380180 irq 51 lpm-pol 1 Apr 23 23:57:58.890083 kernel: ata3: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380200 irq 51 lpm-pol 1 Apr 23 23:57:58.892793 kernel: ata4: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380280 irq 51 lpm-pol 1 Apr 23 23:57:58.892813 kernel: ata5: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380300 irq 51 lpm-pol 1 Apr 23 23:57:58.902722 kernel: ata6: SATA max UDMA/133 abar m4096@0x81380000 port 0x81380380 irq 51 lpm-pol 1 Apr 23 23:57:58.902771 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 23 23:57:58.902824 kernel: GPT:17805311 != 160006143 Apr 23 23:57:58.902838 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 23 23:57:58.903733 kernel: GPT:17805311 != 160006143 Apr 23 23:57:58.905455 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 23 23:57:58.909774 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:57:58.909796 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 23 23:57:58.929429 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:57:59.061349 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Apr 23 23:57:59.201053 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 23 23:57:59.219068 kernel: ata3: SATA link down (SStatus 0 SControl 300) Apr 23 23:57:59.219173 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Apr 23 23:57:59.219210 kernel: ata6: SATA link down (SStatus 0 SControl 300) Apr 23 23:57:59.219232 kernel: ata4: SATA link down (SStatus 0 SControl 300) Apr 23 23:57:59.226071 kernel: ata2: SATA link down (SStatus 0 SControl 300) Apr 23 23:57:59.226122 kernel: ata5: SATA link down (SStatus 0 SControl 300) Apr 23 23:57:59.228045 kernel: ata1.00: LPM support broken, forcing max_power Apr 23 23:57:59.232358 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Apr 23 23:57:59.236548 kernel: ata1.00: applying bridge limits Apr 23 23:57:59.240190 kernel: ata1.00: LPM support broken, forcing max_power Apr 23 23:57:59.247770 kernel: ata1.00: configured for UDMA/100 Apr 23 23:57:59.254148 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Apr 23 23:57:59.297261 kernel: usbcore: registered new interface driver usbhid Apr 23 23:57:59.297327 kernel: usbhid: USB HID core driver Apr 23 23:57:59.310052 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input4 Apr 23 23:57:59.316042 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Apr 23 23:57:59.323167 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Apr 23 23:57:59.323346 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 23 23:57:59.347836 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Apr 23 23:57:59.349765 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Apr 23 23:57:59.355729 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Apr 23 23:57:59.362615 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 23 23:57:59.369383 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Apr 23 23:57:59.370063 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Apr 23 23:57:59.372314 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 23 23:57:59.383171 disk-uuid[662]: Primary Header is updated. Apr 23 23:57:59.383171 disk-uuid[662]: Secondary Entries is updated. Apr 23 23:57:59.383171 disk-uuid[662]: Secondary Header is updated. Apr 23 23:57:59.392057 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:57:59.567724 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 23 23:57:59.568604 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 23 23:57:59.569173 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:57:59.569953 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 23 23:57:59.571588 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 23 23:57:59.600434 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 23 23:58:00.416002 disk-uuid[663]: The operation has completed successfully. Apr 23 23:58:00.420184 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 23 23:58:00.498212 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 23 23:58:00.498314 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 23 23:58:00.511353 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 23 23:58:00.523089 sh[694]: Success Apr 23 23:58:00.539879 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 23 23:58:00.539931 kernel: device-mapper: uevent: version 1.0.3 Apr 23 23:58:00.539943 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Apr 23 23:58:00.551032 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Apr 23 23:58:00.589867 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 23 23:58:00.593068 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 23 23:58:00.604774 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 23 23:58:00.615169 kernel: BTRFS: device fsid b0afcb9a-4dc6-42cc-b61f-b370046a03ca devid 1 transid 32 /dev/mapper/usr (254:0) scanned by mount (706) Apr 23 23:58:00.615215 kernel: BTRFS info (device dm-0): first mount of filesystem b0afcb9a-4dc6-42cc-b61f-b370046a03ca Apr 23 23:58:00.619735 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 23 23:58:00.629217 kernel: BTRFS info (device dm-0 state E): enabling ssd optimizations Apr 23 23:58:00.629257 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Apr 23 23:58:00.632980 kernel: BTRFS info (device dm-0 state E): enabling free space tree Apr 23 23:58:00.635091 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 23 23:58:00.636472 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Apr 23 23:58:00.637299 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 23 23:58:00.638494 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 23 23:58:00.641167 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 23 23:58:00.684036 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (741) Apr 23 23:58:00.688395 kernel: BTRFS info (device sda6): first mount of filesystem 198e7c3b-b6f6-48f6-8d3f-d053e5a12995 Apr 23 23:58:00.688483 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 23 23:58:00.703769 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 23 23:58:00.703807 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:58:00.703816 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:58:00.711042 kernel: BTRFS info (device sda6): last unmount of filesystem 198e7c3b-b6f6-48f6-8d3f-d053e5a12995 Apr 23 23:58:00.711060 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 23 23:58:00.713137 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 23 23:58:00.778658 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 23 23:58:00.782117 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 23 23:58:00.811794 ignition[802]: Ignition 2.22.0 Apr 23 23:58:00.812360 ignition[802]: Stage: fetch-offline Apr 23 23:58:00.812387 ignition[802]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:00.812394 ignition[802]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:00.812473 ignition[802]: parsed url from cmdline: "" Apr 23 23:58:00.812476 ignition[802]: no config URL provided Apr 23 23:58:00.812480 ignition[802]: reading system config file "/usr/lib/ignition/user.ign" Apr 23 23:58:00.812487 ignition[802]: no config at "/usr/lib/ignition/user.ign" Apr 23 23:58:00.812492 ignition[802]: failed to fetch config: resource requires networking Apr 23 23:58:00.812599 ignition[802]: Ignition finished successfully Apr 23 23:58:00.816258 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 23 23:58:00.823169 systemd-networkd[876]: lo: Link UP Apr 23 23:58:00.823179 systemd-networkd[876]: lo: Gained carrier Apr 23 23:58:00.825450 systemd-networkd[876]: Enumeration completed Apr 23 23:58:00.825637 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 23 23:58:00.826241 systemd-networkd[876]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:00.826245 systemd-networkd[876]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:58:00.826586 systemd[1]: Reached target network.target - Network. Apr 23 23:58:00.827030 systemd-networkd[876]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:00.827034 systemd-networkd[876]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:58:00.828206 systemd-networkd[876]: eth0: Link UP Apr 23 23:58:00.828349 systemd-networkd[876]: eth1: Link UP Apr 23 23:58:00.828507 systemd-networkd[876]: eth0: Gained carrier Apr 23 23:58:00.828515 systemd-networkd[876]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:00.830108 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 23 23:58:00.834417 systemd-networkd[876]: eth1: Gained carrier Apr 23 23:58:00.834430 systemd-networkd[876]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:00.850353 ignition[884]: Ignition 2.22.0 Apr 23 23:58:00.850856 ignition[884]: Stage: fetch Apr 23 23:58:00.850962 ignition[884]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:00.850971 ignition[884]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:00.851053 ignition[884]: parsed url from cmdline: "" Apr 23 23:58:00.851056 ignition[884]: no config URL provided Apr 23 23:58:00.851062 ignition[884]: reading system config file "/usr/lib/ignition/user.ign" Apr 23 23:58:00.851069 ignition[884]: no config at "/usr/lib/ignition/user.ign" Apr 23 23:58:00.851108 ignition[884]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Apr 23 23:58:00.851555 ignition[884]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Apr 23 23:58:00.871072 systemd-networkd[876]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 23 23:58:00.892057 systemd-networkd[876]: eth0: DHCPv4 address 135.181.109.116/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 23 23:58:01.052041 ignition[884]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Apr 23 23:58:01.059306 ignition[884]: GET result: OK Apr 23 23:58:01.059765 ignition[884]: parsing config with SHA512: 5d41c918ae243e1646355a96ec8e831cf54ab9ce560e424342b395abc4351c274cf13a00c619f21c180fd5492d1f3f7efd6ac1515938560b56838163511809a4 Apr 23 23:58:01.067196 unknown[884]: fetched base config from "system" Apr 23 23:58:01.067671 ignition[884]: fetch: fetch complete Apr 23 23:58:01.067205 unknown[884]: fetched base config from "system" Apr 23 23:58:01.067676 ignition[884]: fetch: fetch passed Apr 23 23:58:01.067216 unknown[884]: fetched user config from "hetzner" Apr 23 23:58:01.067733 ignition[884]: Ignition finished successfully Apr 23 23:58:01.073630 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 23 23:58:01.077201 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 23 23:58:01.123441 ignition[891]: Ignition 2.22.0 Apr 23 23:58:01.123480 ignition[891]: Stage: kargs Apr 23 23:58:01.123694 ignition[891]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:01.123715 ignition[891]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:01.124570 ignition[891]: kargs: kargs passed Apr 23 23:58:01.124624 ignition[891]: Ignition finished successfully Apr 23 23:58:01.127400 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 23 23:58:01.128899 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 23 23:58:01.154300 ignition[897]: Ignition 2.22.0 Apr 23 23:58:01.154310 ignition[897]: Stage: disks Apr 23 23:58:01.154486 ignition[897]: no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:01.154497 ignition[897]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:01.155270 ignition[897]: disks: disks passed Apr 23 23:58:01.155313 ignition[897]: Ignition finished successfully Apr 23 23:58:01.157826 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 23 23:58:01.158384 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 23 23:58:01.158851 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 23 23:58:01.159488 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 23 23:58:01.160095 systemd[1]: Reached target sysinit.target - System Initialization. Apr 23 23:58:01.160707 systemd[1]: Reached target basic.target - Basic System. Apr 23 23:58:01.162162 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 23 23:58:01.182038 systemd-fsck[906]: ROOT: clean, 15/1628000 files, 120826/1617920 blocks Apr 23 23:58:01.185561 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 23 23:58:01.188086 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 23 23:58:01.278040 kernel: EXT4-fs (sda9): mounted filesystem 8c3ace63-1728-4b5e-a7b6-4ef650e41ba1 r/w with ordered data mode. Quota mode: none. Apr 23 23:58:01.278390 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 23 23:58:01.279242 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 23 23:58:01.280848 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 23 23:58:01.282515 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 23 23:58:01.294632 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 23 23:58:01.295688 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 23 23:58:01.296692 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 23 23:58:01.298938 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 23 23:58:01.305042 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (914) Apr 23 23:58:01.308263 kernel: BTRFS info (device sda6): first mount of filesystem 198e7c3b-b6f6-48f6-8d3f-d053e5a12995 Apr 23 23:58:01.308295 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 23 23:58:01.311135 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 23 23:58:01.321035 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 23 23:58:01.321070 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:58:01.322244 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:58:01.326580 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 23 23:58:01.360274 coreos-metadata[916]: Apr 23 23:58:01.360 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Apr 23 23:58:01.361436 coreos-metadata[916]: Apr 23 23:58:01.361 INFO Fetch successful Apr 23 23:58:01.363673 coreos-metadata[916]: Apr 23 23:58:01.362 INFO wrote hostname ci-4459-2-4-n-09e3934cf7 to /sysroot/etc/hostname Apr 23 23:58:01.364427 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 23 23:58:01.372134 initrd-setup-root[942]: cut: /sysroot/etc/passwd: No such file or directory Apr 23 23:58:01.376844 initrd-setup-root[949]: cut: /sysroot/etc/group: No such file or directory Apr 23 23:58:01.381045 initrd-setup-root[956]: cut: /sysroot/etc/shadow: No such file or directory Apr 23 23:58:01.385710 initrd-setup-root[963]: cut: /sysroot/etc/gshadow: No such file or directory Apr 23 23:58:01.475806 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 23 23:58:01.478087 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 23 23:58:01.479300 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 23 23:58:01.501060 kernel: BTRFS info (device sda6): last unmount of filesystem 198e7c3b-b6f6-48f6-8d3f-d053e5a12995 Apr 23 23:58:01.512379 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 23 23:58:01.526225 ignition[1032]: INFO : Ignition 2.22.0 Apr 23 23:58:01.526225 ignition[1032]: INFO : Stage: mount Apr 23 23:58:01.527096 ignition[1032]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:01.527096 ignition[1032]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:01.527730 ignition[1032]: INFO : mount: mount passed Apr 23 23:58:01.527730 ignition[1032]: INFO : Ignition finished successfully Apr 23 23:58:01.529548 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 23 23:58:01.531942 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 23 23:58:01.613695 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 23 23:58:01.616280 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 23 23:58:01.658105 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (1042) Apr 23 23:58:01.665336 kernel: BTRFS info (device sda6): first mount of filesystem 198e7c3b-b6f6-48f6-8d3f-d053e5a12995 Apr 23 23:58:01.665387 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Apr 23 23:58:01.681205 kernel: BTRFS info (device sda6): enabling ssd optimizations Apr 23 23:58:01.681259 kernel: BTRFS info (device sda6): turning on async discard Apr 23 23:58:01.685091 kernel: BTRFS info (device sda6): enabling free space tree Apr 23 23:58:01.694107 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 23 23:58:01.739557 ignition[1059]: INFO : Ignition 2.22.0 Apr 23 23:58:01.739557 ignition[1059]: INFO : Stage: files Apr 23 23:58:01.741749 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:01.741749 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:01.741749 ignition[1059]: DEBUG : files: compiled without relabeling support, skipping Apr 23 23:58:01.741749 ignition[1059]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 23 23:58:01.741749 ignition[1059]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 23 23:58:01.747096 ignition[1059]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 23 23:58:01.747096 ignition[1059]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 23 23:58:01.749131 ignition[1059]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 23 23:58:01.747162 unknown[1059]: wrote ssh authorized keys file for user: core Apr 23 23:58:01.751839 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 23 23:58:01.751839 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Apr 23 23:58:02.083045 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 23 23:58:02.152288 systemd-networkd[876]: eth0: Gained IPv6LL Apr 23 23:58:02.408848 systemd-networkd[876]: eth1: Gained IPv6LL Apr 23 23:58:02.517792 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Apr 23 23:58:02.517792 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 23 23:58:02.520411 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Apr 23 23:58:02.835623 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 23 23:58:02.939519 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 23 23:58:02.939519 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 23 23:58:02.942313 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.35.1-x86-64.raw: attempt #1 Apr 23 23:58:03.222730 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Apr 23 23:58:03.519062 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.35.1-x86-64.raw" Apr 23 23:58:03.519062 ignition[1059]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 23 23:58:03.521855 ignition[1059]: INFO : files: files passed Apr 23 23:58:03.521855 ignition[1059]: INFO : Ignition finished successfully Apr 23 23:58:03.522703 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 23 23:58:03.526547 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 23 23:58:03.537143 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 23 23:58:03.540635 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 23 23:58:03.540805 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 23 23:58:03.561368 initrd-setup-root-after-ignition[1089]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:58:03.561368 initrd-setup-root-after-ignition[1089]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:58:03.563570 initrd-setup-root-after-ignition[1093]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 23 23:58:03.565688 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 23 23:58:03.567217 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 23 23:58:03.569644 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 23 23:58:03.621943 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 23 23:58:03.622146 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 23 23:58:03.624001 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 23 23:58:03.625203 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 23 23:58:03.626435 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 23 23:58:03.627658 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 23 23:58:03.653295 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 23 23:58:03.655072 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 23 23:58:03.687959 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:58:03.688880 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:58:03.689317 systemd[1]: Stopped target timers.target - Timer Units. Apr 23 23:58:03.689879 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 23 23:58:03.689959 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 23 23:58:03.691893 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 23 23:58:03.693133 systemd[1]: Stopped target basic.target - Basic System. Apr 23 23:58:03.694295 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 23 23:58:03.695390 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 23 23:58:03.696417 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 23 23:58:03.697469 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Apr 23 23:58:03.698513 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 23 23:58:03.699566 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 23 23:58:03.700613 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 23 23:58:03.701696 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 23 23:58:03.702712 systemd[1]: Stopped target swap.target - Swaps. Apr 23 23:58:03.703812 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 23 23:58:03.703913 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 23 23:58:03.705598 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:58:03.706685 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:58:03.707725 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 23 23:58:03.707846 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:58:03.708794 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 23 23:58:03.708866 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 23 23:58:03.710559 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 23 23:58:03.710667 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 23 23:58:03.711701 systemd[1]: ignition-files.service: Deactivated successfully. Apr 23 23:58:03.711773 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 23 23:58:03.712746 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 23 23:58:03.712839 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 23 23:58:03.715145 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 23 23:58:03.718144 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 23 23:58:03.718539 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 23 23:58:03.718648 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:58:03.719103 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 23 23:58:03.719195 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 23 23:58:03.724202 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 23 23:58:03.727085 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 23 23:58:03.740337 ignition[1113]: INFO : Ignition 2.22.0 Apr 23 23:58:03.740337 ignition[1113]: INFO : Stage: umount Apr 23 23:58:03.742399 ignition[1113]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 23 23:58:03.742399 ignition[1113]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Apr 23 23:58:03.742399 ignition[1113]: INFO : umount: umount passed Apr 23 23:58:03.742399 ignition[1113]: INFO : Ignition finished successfully Apr 23 23:58:03.742885 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 23 23:58:03.743004 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 23 23:58:03.743618 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 23 23:58:03.743661 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 23 23:58:03.744708 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 23 23:58:03.744746 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 23 23:58:03.746174 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 23 23:58:03.746229 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 23 23:58:03.746574 systemd[1]: Stopped target network.target - Network. Apr 23 23:58:03.746891 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 23 23:58:03.746926 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 23 23:58:03.747276 systemd[1]: Stopped target paths.target - Path Units. Apr 23 23:58:03.747594 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 23 23:58:03.750781 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:58:03.751440 systemd[1]: Stopped target slices.target - Slice Units. Apr 23 23:58:03.752076 systemd[1]: Stopped target sockets.target - Socket Units. Apr 23 23:58:03.752400 systemd[1]: iscsid.socket: Deactivated successfully. Apr 23 23:58:03.752442 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 23 23:58:03.753416 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 23 23:58:03.753469 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 23 23:58:03.754134 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 23 23:58:03.754176 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 23 23:58:03.754845 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 23 23:58:03.754879 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 23 23:58:03.755656 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 23 23:58:03.756667 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 23 23:58:03.758027 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 23 23:58:03.765381 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 23 23:58:03.765519 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 23 23:58:03.770790 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Apr 23 23:58:03.771067 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 23 23:58:03.771161 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 23 23:58:03.772518 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Apr 23 23:58:03.773059 systemd[1]: Stopped target network-pre.target - Preparation for Network. Apr 23 23:58:03.774130 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 23 23:58:03.774171 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:58:03.775640 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 23 23:58:03.775976 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 23 23:58:03.776048 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 23 23:58:03.776437 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 23 23:58:03.776475 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:58:03.780438 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 23 23:58:03.780482 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 23 23:58:03.780826 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 23 23:58:03.780857 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:58:03.784062 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:58:03.787077 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Apr 23 23:58:03.787136 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:58:03.794338 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 23 23:58:03.794456 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 23 23:58:03.795436 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 23 23:58:03.795527 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 23 23:58:03.801856 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 23 23:58:03.802364 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 23 23:58:03.804414 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 23 23:58:03.804584 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:58:03.805620 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 23 23:58:03.805696 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 23 23:58:03.806440 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 23 23:58:03.806470 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:58:03.807098 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 23 23:58:03.807139 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 23 23:58:03.808176 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 23 23:58:03.808212 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 23 23:58:03.809236 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 23 23:58:03.809278 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 23 23:58:03.811118 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 23 23:58:03.811486 systemd[1]: systemd-network-generator.service: Deactivated successfully. Apr 23 23:58:03.811544 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:58:03.813543 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 23 23:58:03.813580 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:58:03.814231 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Apr 23 23:58:03.814265 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 23 23:58:03.815497 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 23 23:58:03.815544 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:58:03.817090 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:58:03.817132 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:58:03.818870 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Apr 23 23:58:03.818923 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Apr 23 23:58:03.818957 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Apr 23 23:58:03.818993 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Apr 23 23:58:03.827673 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 23 23:58:03.827781 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 23 23:58:03.828619 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 23 23:58:03.829761 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 23 23:58:03.844345 systemd[1]: Switching root. Apr 23 23:58:03.877106 systemd-journald[199]: Journal stopped Apr 23 23:58:05.014668 systemd-journald[199]: Received SIGTERM from PID 1 (systemd). Apr 23 23:58:05.014734 kernel: SELinux: policy capability network_peer_controls=1 Apr 23 23:58:05.014745 kernel: SELinux: policy capability open_perms=1 Apr 23 23:58:05.014757 kernel: SELinux: policy capability extended_socket_class=1 Apr 23 23:58:05.014767 kernel: SELinux: policy capability always_check_network=0 Apr 23 23:58:05.014775 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 23 23:58:05.014784 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 23 23:58:05.014793 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 23 23:58:05.014801 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 23 23:58:05.014809 kernel: SELinux: policy capability userspace_initial_context=0 Apr 23 23:58:05.014818 kernel: audit: type=1403 audit(1776988684.031:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 23 23:58:05.014830 systemd[1]: Successfully loaded SELinux policy in 53.555ms. Apr 23 23:58:05.014852 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 6.477ms. Apr 23 23:58:05.014862 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 23 23:58:05.014872 systemd[1]: Detected virtualization kvm. Apr 23 23:58:05.014881 systemd[1]: Detected architecture x86-64. Apr 23 23:58:05.014889 systemd[1]: Detected first boot. Apr 23 23:58:05.014898 systemd[1]: Hostname set to . Apr 23 23:58:05.014909 systemd[1]: Initializing machine ID from VM UUID. Apr 23 23:58:05.014918 zram_generator::config[1156]: No configuration found. Apr 23 23:58:05.014930 kernel: Guest personality initialized and is inactive Apr 23 23:58:05.014938 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Apr 23 23:58:05.014950 kernel: Initialized host personality Apr 23 23:58:05.014958 kernel: NET: Registered PF_VSOCK protocol family Apr 23 23:58:05.014967 systemd[1]: Populated /etc with preset unit settings. Apr 23 23:58:05.014977 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Apr 23 23:58:05.014986 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 23 23:58:05.014994 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 23 23:58:05.015005 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 23 23:58:05.016992 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 23 23:58:05.017007 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 23 23:58:05.017035 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 23 23:58:05.017060 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 23 23:58:05.017069 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 23 23:58:05.017078 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 23 23:58:05.017087 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 23 23:58:05.017096 systemd[1]: Created slice user.slice - User and Session Slice. Apr 23 23:58:05.017109 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 23 23:58:05.017119 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 23 23:58:05.017128 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 23 23:58:05.017137 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 23 23:58:05.017146 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 23 23:58:05.017155 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 23 23:58:05.017167 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 23 23:58:05.017176 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 23 23:58:05.017185 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 23 23:58:05.017194 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 23 23:58:05.017208 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 23 23:58:05.017217 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 23 23:58:05.017226 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 23 23:58:05.017234 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 23 23:58:05.017243 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 23 23:58:05.017254 systemd[1]: Reached target slices.target - Slice Units. Apr 23 23:58:05.017263 systemd[1]: Reached target swap.target - Swaps. Apr 23 23:58:05.017272 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 23 23:58:05.017281 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 23 23:58:05.017290 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Apr 23 23:58:05.017303 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 23 23:58:05.017312 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 23 23:58:05.017320 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 23 23:58:05.017329 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 23 23:58:05.017340 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 23 23:58:05.017349 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 23 23:58:05.017358 systemd[1]: Mounting media.mount - External Media Directory... Apr 23 23:58:05.017367 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:05.017376 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 23 23:58:05.017385 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 23 23:58:05.017397 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 23 23:58:05.017406 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 23 23:58:05.017424 systemd[1]: Reached target machines.target - Containers. Apr 23 23:58:05.017434 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 23 23:58:05.017443 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:58:05.017452 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 23 23:58:05.017461 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 23 23:58:05.017469 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:58:05.017478 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 23 23:58:05.017487 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:58:05.017496 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 23 23:58:05.017507 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:58:05.017520 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 23 23:58:05.017530 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 23 23:58:05.017539 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 23 23:58:05.017550 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 23 23:58:05.017561 systemd[1]: Stopped systemd-fsck-usr.service. Apr 23 23:58:05.017570 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:58:05.017579 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 23 23:58:05.017588 kernel: fuse: init (API version 7.41) Apr 23 23:58:05.017597 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 23 23:58:05.017608 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 23 23:58:05.017618 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 23 23:58:05.017627 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Apr 23 23:58:05.017635 kernel: loop: module loaded Apr 23 23:58:05.017644 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 23 23:58:05.017653 systemd[1]: verity-setup.service: Deactivated successfully. Apr 23 23:58:05.017662 systemd[1]: Stopped verity-setup.service. Apr 23 23:58:05.017671 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:05.017679 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 23 23:58:05.017690 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 23 23:58:05.017700 systemd[1]: Mounted media.mount - External Media Directory. Apr 23 23:58:05.017709 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 23 23:58:05.017717 kernel: ACPI: bus type drm_connector registered Apr 23 23:58:05.017726 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 23 23:58:05.017735 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 23 23:58:05.017744 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 23 23:58:05.017752 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 23 23:58:05.017761 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 23 23:58:05.017773 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:58:05.017783 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:58:05.017792 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 23 23:58:05.017801 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 23 23:58:05.017810 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:58:05.017819 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:58:05.017852 systemd-journald[1233]: Collecting audit messages is disabled. Apr 23 23:58:05.017874 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 23 23:58:05.017883 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 23 23:58:05.017892 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:58:05.017901 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:58:05.017910 systemd-journald[1233]: Journal started Apr 23 23:58:05.017929 systemd-journald[1233]: Runtime Journal (/run/log/journal/25e6482849544a6e86d55309efe6e421) is 8M, max 76.1M, 68.1M free. Apr 23 23:58:04.620576 systemd[1]: Queued start job for default target multi-user.target. Apr 23 23:58:04.629346 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 23 23:58:04.630571 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 23 23:58:05.021039 systemd[1]: Started systemd-journald.service - Journal Service. Apr 23 23:58:05.022792 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 23 23:58:05.023511 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 23 23:58:05.024233 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 23 23:58:05.024903 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 23 23:58:05.025687 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Apr 23 23:58:05.036750 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 23 23:58:05.042086 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 23 23:58:05.044711 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 23 23:58:05.045139 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 23 23:58:05.045204 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 23 23:58:05.046759 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Apr 23 23:58:05.052155 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 23 23:58:05.052627 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:58:05.055228 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 23 23:58:05.058160 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 23 23:58:05.058541 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 23 23:58:05.060156 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 23 23:58:05.060532 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 23 23:58:05.063339 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:58:05.069105 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 23 23:58:05.071285 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 23 23:58:05.074175 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 23 23:58:05.074941 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 23 23:58:05.090741 systemd-journald[1233]: Time spent on flushing to /var/log/journal/25e6482849544a6e86d55309efe6e421 is 20.108ms for 1254 entries. Apr 23 23:58:05.090741 systemd-journald[1233]: System Journal (/var/log/journal/25e6482849544a6e86d55309efe6e421) is 8M, max 584.8M, 576.8M free. Apr 23 23:58:05.149908 systemd-journald[1233]: Received client request to flush runtime journal. Apr 23 23:58:05.150082 kernel: loop0: detected capacity change from 0 to 128560 Apr 23 23:58:05.150099 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 23 23:58:05.110814 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 23 23:58:05.112763 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 23 23:58:05.115122 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Apr 23 23:58:05.136171 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:58:05.154561 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 23 23:58:05.158037 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 23 23:58:05.170263 systemd-tmpfiles[1282]: ACLs are not supported, ignoring. Apr 23 23:58:05.170279 systemd-tmpfiles[1282]: ACLs are not supported, ignoring. Apr 23 23:58:05.171140 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Apr 23 23:58:05.178717 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 23 23:58:05.182082 kernel: loop1: detected capacity change from 0 to 110984 Apr 23 23:58:05.182476 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 23 23:58:05.212036 kernel: loop2: detected capacity change from 0 to 217752 Apr 23 23:58:05.229675 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 23 23:58:05.233320 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 23 23:58:05.249035 kernel: loop3: detected capacity change from 0 to 8 Apr 23 23:58:05.260271 systemd-tmpfiles[1305]: ACLs are not supported, ignoring. Apr 23 23:58:05.260747 systemd-tmpfiles[1305]: ACLs are not supported, ignoring. Apr 23 23:58:05.266459 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 23 23:58:05.279050 kernel: loop4: detected capacity change from 0 to 128560 Apr 23 23:58:05.301241 kernel: loop5: detected capacity change from 0 to 110984 Apr 23 23:58:05.322041 kernel: loop6: detected capacity change from 0 to 217752 Apr 23 23:58:05.345165 kernel: loop7: detected capacity change from 0 to 8 Apr 23 23:58:05.346064 (sd-merge)[1310]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Apr 23 23:58:05.348953 (sd-merge)[1310]: Merged extensions into '/usr'. Apr 23 23:58:05.353715 systemd[1]: Reload requested from client PID 1281 ('systemd-sysext') (unit systemd-sysext.service)... Apr 23 23:58:05.353802 systemd[1]: Reloading... Apr 23 23:58:05.480240 zram_generator::config[1338]: No configuration found. Apr 23 23:58:05.556131 ldconfig[1276]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 23 23:58:05.629626 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 23 23:58:05.629814 systemd[1]: Reloading finished in 275 ms. Apr 23 23:58:05.645937 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 23 23:58:05.646722 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 23 23:58:05.647333 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 23 23:58:05.663204 systemd[1]: Starting ensure-sysext.service... Apr 23 23:58:05.666109 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 23 23:58:05.671124 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 23 23:58:05.680436 systemd[1]: Reload requested from client PID 1380 ('systemctl') (unit ensure-sysext.service)... Apr 23 23:58:05.680514 systemd[1]: Reloading... Apr 23 23:58:05.688248 systemd-tmpfiles[1381]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Apr 23 23:58:05.688599 systemd-tmpfiles[1381]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Apr 23 23:58:05.688886 systemd-tmpfiles[1381]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 23 23:58:05.689690 systemd-tmpfiles[1381]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 23 23:58:05.690551 systemd-tmpfiles[1381]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 23 23:58:05.690806 systemd-tmpfiles[1381]: ACLs are not supported, ignoring. Apr 23 23:58:05.690898 systemd-tmpfiles[1381]: ACLs are not supported, ignoring. Apr 23 23:58:05.696213 systemd-tmpfiles[1381]: Detected autofs mount point /boot during canonicalization of boot. Apr 23 23:58:05.696286 systemd-tmpfiles[1381]: Skipping /boot Apr 23 23:58:05.705497 systemd-tmpfiles[1381]: Detected autofs mount point /boot during canonicalization of boot. Apr 23 23:58:05.705579 systemd-tmpfiles[1381]: Skipping /boot Apr 23 23:58:05.716334 systemd-udevd[1382]: Using default interface naming scheme 'v255'. Apr 23 23:58:05.766068 zram_generator::config[1417]: No configuration found. Apr 23 23:58:05.931365 kernel: mousedev: PS/2 mouse device common for all mice Apr 23 23:58:05.949037 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input5 Apr 23 23:58:05.976046 kernel: ACPI: button: Power Button [PWRF] Apr 23 23:58:05.982695 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Apr 23 23:58:05.983108 systemd[1]: Reloading finished in 302 ms. Apr 23 23:58:05.992877 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 23 23:58:05.995619 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 23 23:58:06.016475 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Apr 23 23:58:06.016566 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.019770 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 23 23:58:06.025411 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 23 23:58:06.027245 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:58:06.028148 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:58:06.029802 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:58:06.038233 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:58:06.038928 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:58:06.038972 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:58:06.045162 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 23 23:58:06.049063 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 23 23:58:06.054158 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 23 23:58:06.057232 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 23 23:58:06.057905 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.058684 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:58:06.059699 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:58:06.067236 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:58:06.067443 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:58:06.070688 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 23 23:58:06.074745 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 23 23:58:06.075716 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:58:06.077040 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:58:06.081529 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.081769 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:58:06.083874 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:58:06.092589 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 23 23:58:06.096627 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 23 23:58:06.097139 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:58:06.097278 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:58:06.097435 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.103241 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.103450 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:58:06.103571 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:58:06.103626 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:58:06.103678 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.104123 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:58:06.104784 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:58:06.109905 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.110574 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 23 23:58:06.112436 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 23 23:58:06.114268 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 23 23:58:06.114876 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 23 23:58:06.114951 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 23 23:58:06.115224 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 23 23:58:06.115828 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 23 23:58:06.122589 systemd[1]: Finished ensure-sysext.service. Apr 23 23:58:06.127149 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 23 23:58:06.144240 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 23 23:58:06.146349 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 23 23:58:06.161058 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Apr 23 23:58:06.161301 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Apr 23 23:58:06.164043 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Apr 23 23:58:06.171622 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 23 23:58:06.171813 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 23 23:58:06.172814 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 23 23:58:06.183519 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 23 23:58:06.184202 augenrules[1544]: No rules Apr 23 23:58:06.187173 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 23 23:58:06.187857 systemd[1]: audit-rules.service: Deactivated successfully. Apr 23 23:58:06.188258 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 23 23:58:06.189414 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 23 23:58:06.190297 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 23 23:58:06.190891 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 23 23:58:06.192504 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 23 23:58:06.200727 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 23 23:58:06.202055 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 23 23:58:06.207476 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 23 23:58:06.207616 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 23 23:58:06.243045 kernel: EDAC MC: Ver: 3.0.0 Apr 23 23:58:06.255047 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Apr 23 23:58:06.262037 kernel: Console: switching to colour dummy device 80x25 Apr 23 23:58:06.267584 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Apr 23 23:58:06.267800 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Apr 23 23:58:06.267816 kernel: [drm] features: -context_init Apr 23 23:58:06.273205 kernel: [drm] number of scanouts: 1 Apr 23 23:58:06.273239 kernel: [drm] number of cap sets: 0 Apr 23 23:58:06.274803 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:01.0 on minor 0 Apr 23 23:58:06.277103 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Apr 23 23:58:06.277128 kernel: Console: switching to colour frame buffer device 160x50 Apr 23 23:58:06.290038 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Apr 23 23:58:06.308929 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Apr 23 23:58:06.317276 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 23 23:58:06.321271 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:58:06.327868 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 23 23:58:06.337857 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:58:06.342507 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:58:06.348232 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:58:06.354055 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 23 23:58:06.403486 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 23 23:58:06.403752 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:58:06.409151 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 23 23:58:06.467678 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 23 23:58:06.519481 systemd-resolved[1504]: Positive Trust Anchors: Apr 23 23:58:06.519498 systemd-resolved[1504]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 23 23:58:06.519519 systemd-resolved[1504]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 23 23:58:06.521556 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 23 23:58:06.523202 systemd[1]: Reached target time-set.target - System Time Set. Apr 23 23:58:06.525512 systemd-resolved[1504]: Using system hostname 'ci-4459-2-4-n-09e3934cf7'. Apr 23 23:58:06.526653 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 23 23:58:06.526829 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 23 23:58:06.526932 systemd[1]: Reached target sysinit.target - System Initialization. Apr 23 23:58:06.527150 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 23 23:58:06.527274 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 23 23:58:06.527402 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Apr 23 23:58:06.527652 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 23 23:58:06.527782 systemd-networkd[1503]: lo: Link UP Apr 23 23:58:06.527787 systemd-networkd[1503]: lo: Gained carrier Apr 23 23:58:06.527852 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 23 23:58:06.527928 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 23 23:58:06.527996 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 23 23:58:06.528108 systemd[1]: Reached target paths.target - Path Units. Apr 23 23:58:06.529719 systemd[1]: Reached target timers.target - Timer Units. Apr 23 23:58:06.530968 systemd-timesyncd[1525]: No network connectivity, watching for changes. Apr 23 23:58:06.531531 systemd-networkd[1503]: Enumeration completed Apr 23 23:58:06.531975 systemd-networkd[1503]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:06.531984 systemd-networkd[1503]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:58:06.532128 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 23 23:58:06.535648 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 23 23:58:06.536807 systemd-networkd[1503]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:06.536816 systemd-networkd[1503]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 23 23:58:06.537989 systemd-networkd[1503]: eth0: Link UP Apr 23 23:58:06.538156 systemd-networkd[1503]: eth0: Gained carrier Apr 23 23:58:06.538169 systemd-networkd[1503]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:06.539960 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Apr 23 23:58:06.542181 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Apr 23 23:58:06.543290 systemd-networkd[1503]: eth1: Link UP Apr 23 23:58:06.543752 systemd-networkd[1503]: eth1: Gained carrier Apr 23 23:58:06.544138 systemd-networkd[1503]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 23 23:58:06.544467 systemd[1]: Reached target ssh-access.target - SSH Access Available. Apr 23 23:58:06.552754 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 23 23:58:06.554030 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Apr 23 23:58:06.555196 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 23 23:58:06.558862 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 23 23:58:06.560551 systemd[1]: Reached target network.target - Network. Apr 23 23:58:06.564345 systemd[1]: Reached target sockets.target - Socket Units. Apr 23 23:58:06.564826 systemd[1]: Reached target basic.target - Basic System. Apr 23 23:58:06.565382 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 23 23:58:06.565418 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 23 23:58:06.566330 systemd[1]: Starting containerd.service - containerd container runtime... Apr 23 23:58:06.570275 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 23 23:58:06.576106 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 23 23:58:06.580178 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 23 23:58:06.583244 systemd-networkd[1503]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Apr 23 23:58:06.583968 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 23 23:58:06.586100 systemd-timesyncd[1525]: Network configuration changed, trying to establish connection. Apr 23 23:58:06.590436 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 23 23:58:06.592426 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 23 23:58:06.596235 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Apr 23 23:58:06.596738 systemd-networkd[1503]: eth0: DHCPv4 address 135.181.109.116/32, gateway 172.31.1.1 acquired from 172.31.1.1 Apr 23 23:58:06.598991 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 23 23:58:06.600644 systemd-timesyncd[1525]: Network configuration changed, trying to establish connection. Apr 23 23:58:06.601082 jq[1600]: false Apr 23 23:58:06.607913 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 23 23:58:06.609033 coreos-metadata[1597]: Apr 23 23:58:06.608 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Apr 23 23:58:06.613244 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Apr 23 23:58:06.620090 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Refreshing passwd entry cache Apr 23 23:58:06.619711 oslogin_cache_refresh[1604]: Refreshing passwd entry cache Apr 23 23:58:06.625103 coreos-metadata[1597]: Apr 23 23:58:06.620 INFO Fetch successful Apr 23 23:58:06.625103 coreos-metadata[1597]: Apr 23 23:58:06.620 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Apr 23 23:58:06.623155 oslogin_cache_refresh[1604]: Failure getting users, quitting Apr 23 23:58:06.625197 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Failure getting users, quitting Apr 23 23:58:06.625197 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Apr 23 23:58:06.625197 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Refreshing group entry cache Apr 23 23:58:06.625197 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Failure getting groups, quitting Apr 23 23:58:06.625197 google_oslogin_nss_cache[1604]: oslogin_cache_refresh[1604]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Apr 23 23:58:06.621157 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 23 23:58:06.623169 oslogin_cache_refresh[1604]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Apr 23 23:58:06.623205 oslogin_cache_refresh[1604]: Refreshing group entry cache Apr 23 23:58:06.624123 oslogin_cache_refresh[1604]: Failure getting groups, quitting Apr 23 23:58:06.624150 oslogin_cache_refresh[1604]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Apr 23 23:58:06.627042 coreos-metadata[1597]: Apr 23 23:58:06.625 INFO Fetch successful Apr 23 23:58:06.630056 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 23 23:58:06.637066 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 23 23:58:06.643555 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Apr 23 23:58:06.649386 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 23 23:58:06.652512 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 23 23:58:06.653890 extend-filesystems[1603]: Found /dev/sda6 Apr 23 23:58:06.652852 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 23 23:58:06.659474 systemd[1]: Starting update-engine.service - Update Engine... Apr 23 23:58:06.666317 extend-filesystems[1603]: Found /dev/sda9 Apr 23 23:58:06.667248 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 23 23:58:06.678542 extend-filesystems[1603]: Checking size of /dev/sda9 Apr 23 23:58:06.680053 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 23 23:58:06.690112 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 23 23:58:06.694057 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 23 23:58:06.694452 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Apr 23 23:58:06.694658 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Apr 23 23:58:06.695732 systemd[1]: motdgen.service: Deactivated successfully. Apr 23 23:58:06.695947 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 23 23:58:06.697337 extend-filesystems[1603]: Resized partition /dev/sda9 Apr 23 23:58:06.709078 extend-filesystems[1636]: resize2fs 1.47.3 (8-Jul-2025) Apr 23 23:58:06.705538 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 23 23:58:06.715676 jq[1626]: true Apr 23 23:58:06.715766 update_engine[1623]: I20260423 23:58:06.702304 1623 main.cc:92] Flatcar Update Engine starting Apr 23 23:58:06.705731 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 23 23:58:06.723039 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 19393531 blocks Apr 23 23:58:06.736150 (ntainerd)[1639]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 23 23:58:06.745241 jq[1638]: true Apr 23 23:58:06.767378 tar[1637]: linux-amd64/LICENSE Apr 23 23:58:06.767995 tar[1637]: linux-amd64/helm Apr 23 23:58:06.778522 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Apr 23 23:58:06.813346 dbus-daemon[1598]: [system] SELinux support is enabled Apr 23 23:58:06.813492 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 23 23:58:06.819328 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 23 23:58:06.819351 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 23 23:58:06.823730 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 23 23:58:06.823747 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 23 23:58:06.842707 systemd[1]: Started update-engine.service - Update Engine. Apr 23 23:58:06.844991 update_engine[1623]: I20260423 23:58:06.844944 1623 update_check_scheduler.cc:74] Next update check in 5m0s Apr 23 23:58:06.845985 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 23 23:58:06.850948 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 23 23:58:06.858934 systemd-logind[1616]: New seat seat0. Apr 23 23:58:06.865567 systemd-logind[1616]: Watching system buttons on /dev/input/event3 (Power Button) Apr 23 23:58:06.865593 systemd-logind[1616]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 23 23:58:06.871509 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 23 23:58:06.875417 systemd[1]: Started systemd-logind.service - User Login Management. Apr 23 23:58:06.902820 bash[1682]: Updated "/home/core/.ssh/authorized_keys" Apr 23 23:58:06.905208 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 23 23:58:06.911660 systemd[1]: Starting sshkeys.service... Apr 23 23:58:06.965508 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 23 23:58:06.974045 containerd[1639]: time="2026-04-23T23:58:06Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Apr 23 23:58:06.974045 containerd[1639]: time="2026-04-23T23:58:06.972056551Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Apr 23 23:58:06.975643 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 23 23:58:07.001669 containerd[1639]: time="2026-04-23T23:58:07.001622452Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="10.836µs" Apr 23 23:58:07.001669 containerd[1639]: time="2026-04-23T23:58:07.001663604Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Apr 23 23:58:07.001774 containerd[1639]: time="2026-04-23T23:58:07.001682612Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Apr 23 23:58:07.001856 containerd[1639]: time="2026-04-23T23:58:07.001834740Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Apr 23 23:58:07.001856 containerd[1639]: time="2026-04-23T23:58:07.001852817Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Apr 23 23:58:07.001905 containerd[1639]: time="2026-04-23T23:58:07.001875221Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 23 23:58:07.001957 containerd[1639]: time="2026-04-23T23:58:07.001929382Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Apr 23 23:58:07.001957 containerd[1639]: time="2026-04-23T23:58:07.001943493Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003293 containerd[1639]: time="2026-04-23T23:58:07.003267800Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003293 containerd[1639]: time="2026-04-23T23:58:07.003290705Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003337 containerd[1639]: time="2026-04-23T23:58:07.003303003Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003337 containerd[1639]: time="2026-04-23T23:58:07.003310735Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003491 containerd[1639]: time="2026-04-23T23:58:07.003401321Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003657 containerd[1639]: time="2026-04-23T23:58:07.003585898Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003657 containerd[1639]: time="2026-04-23T23:58:07.003616474Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Apr 23 23:58:07.003657 containerd[1639]: time="2026-04-23T23:58:07.003626158Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Apr 23 23:58:07.003657 containerd[1639]: time="2026-04-23T23:58:07.003648822Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Apr 23 23:58:07.005374 containerd[1639]: time="2026-04-23T23:58:07.005325398Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Apr 23 23:58:07.005411 containerd[1639]: time="2026-04-23T23:58:07.005398908Z" level=info msg="metadata content store policy set" policy=shared Apr 23 23:58:07.029531 kernel: EXT4-fs (sda9): resized filesystem to 19393531 Apr 23 23:58:07.044302 coreos-metadata[1687]: Apr 23 23:58:07.044 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Apr 23 23:58:07.045297 coreos-metadata[1687]: Apr 23 23:58:07.045 INFO Fetch successful Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050083263Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050134360Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050146208Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050155652Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050165537Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Apr 23 23:58:07.050178 containerd[1639]: time="2026-04-23T23:58:07.050180739Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050190103Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050199608Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050208802Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050217014Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050224135Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Apr 23 23:58:07.050278 containerd[1639]: time="2026-04-23T23:58:07.050233178Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Apr 23 23:58:07.050400 containerd[1639]: time="2026-04-23T23:58:07.050340750Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Apr 23 23:58:07.050400 containerd[1639]: time="2026-04-23T23:58:07.050354390Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Apr 23 23:58:07.050400 containerd[1639]: time="2026-04-23T23:58:07.050369793Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Apr 23 23:58:07.050400 containerd[1639]: time="2026-04-23T23:58:07.050382572Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Apr 23 23:58:07.050400 containerd[1639]: time="2026-04-23T23:58:07.050401701Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050411105Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050422652Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050431165Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050443243Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050450714Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Apr 23 23:58:07.050466 containerd[1639]: time="2026-04-23T23:58:07.050462372Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Apr 23 23:58:07.050547 containerd[1639]: time="2026-04-23T23:58:07.050503103Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Apr 23 23:58:07.050547 containerd[1639]: time="2026-04-23T23:58:07.050513859Z" level=info msg="Start snapshots syncer" Apr 23 23:58:07.050547 containerd[1639]: time="2026-04-23T23:58:07.050534781Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Apr 23 23:58:07.054241 containerd[1639]: time="2026-04-23T23:58:07.051369924Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Apr 23 23:58:07.054241 containerd[1639]: time="2026-04-23T23:58:07.051446899Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Apr 23 23:58:07.053912 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 23 23:58:07.054552 extend-filesystems[1636]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Apr 23 23:58:07.054552 extend-filesystems[1636]: old_desc_blocks = 1, new_desc_blocks = 10 Apr 23 23:58:07.054552 extend-filesystems[1636]: The filesystem on /dev/sda9 is now 19393531 (4k) blocks long. Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.051482753Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052067510Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052085628Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052095552Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052103595Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052112207Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.052119829Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053084957Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053103725Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053148002Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053156935Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053187471Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053242914Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 23 23:58:07.079699 containerd[1639]: time="2026-04-23T23:58:07.053249404Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 23 23:58:07.054133 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 23 23:58:07.079992 extend-filesystems[1603]: Resized filesystem in /dev/sda9 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053256845Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053262774Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053269674Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053283224Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053296855Z" level=info msg="runtime interface created" Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053301051Z" level=info msg="created NRI interface" Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053321141Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053329614Z" level=info msg="Connect containerd service" Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.053352699Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 23 23:58:07.082399 containerd[1639]: time="2026-04-23T23:58:07.059417709Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 23 23:58:07.055067 unknown[1687]: wrote ssh authorized keys file for user: core Apr 23 23:58:07.099989 update-ssh-keys[1700]: Updated "/home/core/.ssh/authorized_keys" Apr 23 23:58:07.101395 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 23 23:58:07.104208 systemd[1]: Finished sshkeys.service. Apr 23 23:58:07.140101 locksmithd[1681]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 23 23:58:07.147005 sshd_keygen[1632]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 23 23:58:07.164481 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 23 23:58:07.167566 containerd[1639]: time="2026-04-23T23:58:07.167539186Z" level=info msg="Start subscribing containerd event" Apr 23 23:58:07.167679 containerd[1639]: time="2026-04-23T23:58:07.167658815Z" level=info msg="Start recovering state" Apr 23 23:58:07.168173 containerd[1639]: time="2026-04-23T23:58:07.167811815Z" level=info msg="Start event monitor" Apr 23 23:58:07.168221 containerd[1639]: time="2026-04-23T23:58:07.168211425Z" level=info msg="Start cni network conf syncer for default" Apr 23 23:58:07.168285 containerd[1639]: time="2026-04-23T23:58:07.168274760Z" level=info msg="Start streaming server" Apr 23 23:58:07.168329 containerd[1639]: time="2026-04-23T23:58:07.168321570Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Apr 23 23:58:07.168744 containerd[1639]: time="2026-04-23T23:58:07.168732947Z" level=info msg="runtime interface starting up..." Apr 23 23:58:07.168947 containerd[1639]: time="2026-04-23T23:58:07.168935822Z" level=info msg="starting plugins..." Apr 23 23:58:07.169117 containerd[1639]: time="2026-04-23T23:58:07.168982151Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Apr 23 23:58:07.169668 containerd[1639]: time="2026-04-23T23:58:07.169399177Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 23 23:58:07.169668 containerd[1639]: time="2026-04-23T23:58:07.169444324Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 23 23:58:07.170520 containerd[1639]: time="2026-04-23T23:58:07.170509613Z" level=info msg="containerd successfully booted in 0.201626s" Apr 23 23:58:07.171161 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 23 23:58:07.172265 systemd[1]: Started containerd.service - containerd container runtime. Apr 23 23:58:07.189952 systemd[1]: issuegen.service: Deactivated successfully. Apr 23 23:58:07.190197 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 23 23:58:07.196171 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 23 23:58:07.216224 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 23 23:58:07.224220 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 23 23:58:07.226280 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 23 23:58:07.229629 systemd[1]: Reached target getty.target - Login Prompts. Apr 23 23:58:07.284991 tar[1637]: linux-amd64/README.md Apr 23 23:58:07.301232 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 23 23:58:07.441573 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 23 23:58:07.446697 systemd[1]: Started sshd@0-135.181.109.116:22-20.229.252.112:52078.service - OpenSSH per-connection server daemon (20.229.252.112:52078). Apr 23 23:58:07.686333 sshd[1736]: Accepted publickey for core from 20.229.252.112 port 52078 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:07.689779 sshd-session[1736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:07.702170 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 23 23:58:07.707418 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 23 23:58:07.727996 systemd-logind[1616]: New session 1 of user core. Apr 23 23:58:07.740612 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 23 23:58:07.750086 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 23 23:58:07.770357 (systemd)[1741]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 23 23:58:07.774576 systemd-logind[1616]: New session c1 of user core. Apr 23 23:58:07.902584 systemd[1741]: Queued start job for default target default.target. Apr 23 23:58:07.909147 systemd[1741]: Created slice app.slice - User Application Slice. Apr 23 23:58:07.909169 systemd[1741]: Reached target paths.target - Paths. Apr 23 23:58:07.909207 systemd[1741]: Reached target timers.target - Timers. Apr 23 23:58:07.910630 systemd[1741]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 23 23:58:07.926314 systemd[1741]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 23 23:58:07.926566 systemd[1741]: Reached target sockets.target - Sockets. Apr 23 23:58:07.926661 systemd[1741]: Reached target basic.target - Basic System. Apr 23 23:58:07.926878 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 23 23:58:07.927263 systemd[1741]: Reached target default.target - Main User Target. Apr 23 23:58:07.927351 systemd[1741]: Startup finished in 142ms. Apr 23 23:58:07.942256 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 23 23:58:08.041188 systemd-networkd[1503]: eth0: Gained IPv6LL Apr 23 23:58:08.042251 systemd-timesyncd[1525]: Network configuration changed, trying to establish connection. Apr 23 23:58:08.047165 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 23 23:58:08.053088 systemd[1]: Reached target network-online.target - Network is Online. Apr 23 23:58:08.058445 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:08.062173 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 23 23:58:08.067290 systemd[1]: Started sshd@1-135.181.109.116:22-20.229.252.112:52506.service - OpenSSH per-connection server daemon (20.229.252.112:52506). Apr 23 23:58:08.086811 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 23 23:58:08.168153 systemd-networkd[1503]: eth1: Gained IPv6LL Apr 23 23:58:08.168815 systemd-timesyncd[1525]: Network configuration changed, trying to establish connection. Apr 23 23:58:08.261457 sshd[1755]: Accepted publickey for core from 20.229.252.112 port 52506 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:08.264181 sshd-session[1755]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:08.271341 systemd-logind[1616]: New session 2 of user core. Apr 23 23:58:08.276116 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 23 23:58:08.357144 sshd[1767]: Connection closed by 20.229.252.112 port 52506 Apr 23 23:58:08.358318 sshd-session[1755]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:08.367153 systemd-logind[1616]: Session 2 logged out. Waiting for processes to exit. Apr 23 23:58:08.368562 systemd[1]: sshd@1-135.181.109.116:22-20.229.252.112:52506.service: Deactivated successfully. Apr 23 23:58:08.371882 systemd[1]: session-2.scope: Deactivated successfully. Apr 23 23:58:08.374853 systemd-logind[1616]: Removed session 2. Apr 23 23:58:08.397070 systemd[1]: Started sshd@2-135.181.109.116:22-20.229.252.112:52510.service - OpenSSH per-connection server daemon (20.229.252.112:52510). Apr 23 23:58:08.600197 sshd[1773]: Accepted publickey for core from 20.229.252.112 port 52510 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:08.602793 sshd-session[1773]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:08.609516 systemd-logind[1616]: New session 3 of user core. Apr 23 23:58:08.613122 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 23 23:58:08.702669 sshd[1776]: Connection closed by 20.229.252.112 port 52510 Apr 23 23:58:08.703813 sshd-session[1773]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:08.710112 systemd[1]: sshd@2-135.181.109.116:22-20.229.252.112:52510.service: Deactivated successfully. Apr 23 23:58:08.714575 systemd[1]: session-3.scope: Deactivated successfully. Apr 23 23:58:08.717118 systemd-logind[1616]: Session 3 logged out. Waiting for processes to exit. Apr 23 23:58:08.720644 systemd-logind[1616]: Removed session 3. Apr 23 23:58:09.003501 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:09.008776 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 23 23:58:09.011010 systemd[1]: Startup finished in 2.946s (kernel) + 6.373s (initrd) + 5.033s (userspace) = 14.353s. Apr 23 23:58:09.011609 (kubelet)[1786]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:58:09.569204 kubelet[1786]: E0423 23:58:09.568967 1786 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:58:09.573712 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:58:09.574108 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:58:09.574771 systemd[1]: kubelet.service: Consumed 887ms CPU time, 255.1M memory peak. Apr 23 23:58:18.746794 systemd[1]: Started sshd@3-135.181.109.116:22-20.229.252.112:38212.service - OpenSSH per-connection server daemon (20.229.252.112:38212). Apr 23 23:58:18.963988 sshd[1797]: Accepted publickey for core from 20.229.252.112 port 38212 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:18.966667 sshd-session[1797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:18.976099 systemd-logind[1616]: New session 4 of user core. Apr 23 23:58:18.983238 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 23 23:58:19.059088 sshd[1800]: Connection closed by 20.229.252.112 port 38212 Apr 23 23:58:19.061326 sshd-session[1797]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:19.066653 systemd[1]: sshd@3-135.181.109.116:22-20.229.252.112:38212.service: Deactivated successfully. Apr 23 23:58:19.069826 systemd[1]: session-4.scope: Deactivated successfully. Apr 23 23:58:19.072468 systemd-logind[1616]: Session 4 logged out. Waiting for processes to exit. Apr 23 23:58:19.075271 systemd-logind[1616]: Removed session 4. Apr 23 23:58:19.103745 systemd[1]: Started sshd@4-135.181.109.116:22-20.229.252.112:38214.service - OpenSSH per-connection server daemon (20.229.252.112:38214). Apr 23 23:58:19.307687 sshd[1806]: Accepted publickey for core from 20.229.252.112 port 38214 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:19.310203 sshd-session[1806]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:19.319094 systemd-logind[1616]: New session 5 of user core. Apr 23 23:58:19.331266 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 23 23:58:19.402356 sshd[1809]: Connection closed by 20.229.252.112 port 38214 Apr 23 23:58:19.402872 sshd-session[1806]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:19.407516 systemd[1]: sshd@4-135.181.109.116:22-20.229.252.112:38214.service: Deactivated successfully. Apr 23 23:58:19.410722 systemd[1]: session-5.scope: Deactivated successfully. Apr 23 23:58:19.414354 systemd-logind[1616]: Session 5 logged out. Waiting for processes to exit. Apr 23 23:58:19.415689 systemd-logind[1616]: Removed session 5. Apr 23 23:58:19.444795 systemd[1]: Started sshd@5-135.181.109.116:22-20.229.252.112:38222.service - OpenSSH per-connection server daemon (20.229.252.112:38222). Apr 23 23:58:19.618788 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 23 23:58:19.623311 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:19.660117 sshd[1815]: Accepted publickey for core from 20.229.252.112 port 38222 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:19.663664 sshd-session[1815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:19.673769 systemd-logind[1616]: New session 6 of user core. Apr 23 23:58:19.687221 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 23 23:58:19.772759 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:19.777466 sshd[1821]: Connection closed by 20.229.252.112 port 38222 Apr 23 23:58:19.778908 sshd-session[1815]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:19.782320 (kubelet)[1827]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:58:19.783182 systemd[1]: sshd@5-135.181.109.116:22-20.229.252.112:38222.service: Deactivated successfully. Apr 23 23:58:19.785621 systemd[1]: session-6.scope: Deactivated successfully. Apr 23 23:58:19.788464 systemd-logind[1616]: Session 6 logged out. Waiting for processes to exit. Apr 23 23:58:19.790630 systemd-logind[1616]: Removed session 6. Apr 23 23:58:19.820382 kubelet[1827]: E0423 23:58:19.820024 1827 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:58:19.820792 systemd[1]: Started sshd@6-135.181.109.116:22-20.229.252.112:38230.service - OpenSSH per-connection server daemon (20.229.252.112:38230). Apr 23 23:58:19.827162 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:58:19.827523 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:58:19.832205 systemd[1]: kubelet.service: Consumed 155ms CPU time, 110.1M memory peak. Apr 23 23:58:20.016174 sshd[1838]: Accepted publickey for core from 20.229.252.112 port 38230 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:20.018951 sshd-session[1838]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:20.029117 systemd-logind[1616]: New session 7 of user core. Apr 23 23:58:20.036297 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 23 23:58:20.101272 sudo[1843]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 23 23:58:20.102209 sudo[1843]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:58:20.130695 sudo[1843]: pam_unix(sudo:session): session closed for user root Apr 23 23:58:20.162208 sshd[1842]: Connection closed by 20.229.252.112 port 38230 Apr 23 23:58:20.164437 sshd-session[1838]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:20.170885 systemd[1]: sshd@6-135.181.109.116:22-20.229.252.112:38230.service: Deactivated successfully. Apr 23 23:58:20.175706 systemd[1]: session-7.scope: Deactivated successfully. Apr 23 23:58:20.180436 systemd-logind[1616]: Session 7 logged out. Waiting for processes to exit. Apr 23 23:58:20.182792 systemd-logind[1616]: Removed session 7. Apr 23 23:58:20.211541 systemd[1]: Started sshd@7-135.181.109.116:22-20.229.252.112:38238.service - OpenSSH per-connection server daemon (20.229.252.112:38238). Apr 23 23:58:20.431727 sshd[1849]: Accepted publickey for core from 20.229.252.112 port 38238 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:20.434258 sshd-session[1849]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:20.442117 systemd-logind[1616]: New session 8 of user core. Apr 23 23:58:20.453266 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 23 23:58:20.504638 sudo[1854]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 23 23:58:20.505884 sudo[1854]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:58:20.512719 sudo[1854]: pam_unix(sudo:session): session closed for user root Apr 23 23:58:20.523824 sudo[1853]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Apr 23 23:58:20.524519 sudo[1853]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:58:20.542532 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 23 23:58:20.611441 augenrules[1876]: No rules Apr 23 23:58:20.612773 systemd[1]: audit-rules.service: Deactivated successfully. Apr 23 23:58:20.613301 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 23 23:58:20.615370 sudo[1853]: pam_unix(sudo:session): session closed for user root Apr 23 23:58:20.646679 sshd[1852]: Connection closed by 20.229.252.112 port 38238 Apr 23 23:58:20.648329 sshd-session[1849]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:20.655701 systemd-logind[1616]: Session 8 logged out. Waiting for processes to exit. Apr 23 23:58:20.657371 systemd[1]: sshd@7-135.181.109.116:22-20.229.252.112:38238.service: Deactivated successfully. Apr 23 23:58:20.660932 systemd[1]: session-8.scope: Deactivated successfully. Apr 23 23:58:20.664103 systemd-logind[1616]: Removed session 8. Apr 23 23:58:20.694079 systemd[1]: Started sshd@8-135.181.109.116:22-20.229.252.112:38250.service - OpenSSH per-connection server daemon (20.229.252.112:38250). Apr 23 23:58:20.912075 sshd[1885]: Accepted publickey for core from 20.229.252.112 port 38250 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 23 23:58:20.913898 sshd-session[1885]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 23 23:58:20.922130 systemd-logind[1616]: New session 9 of user core. Apr 23 23:58:20.933268 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 23 23:58:20.986732 sudo[1889]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 23 23:58:20.987602 sudo[1889]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 23 23:58:21.330955 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 23 23:58:21.345723 (dockerd)[1907]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 23 23:58:21.565391 dockerd[1907]: time="2026-04-23T23:58:21.565134917Z" level=info msg="Starting up" Apr 23 23:58:21.567655 dockerd[1907]: time="2026-04-23T23:58:21.567606716Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Apr 23 23:58:21.581900 dockerd[1907]: time="2026-04-23T23:58:21.581754633Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Apr 23 23:58:21.609749 systemd[1]: var-lib-docker-metacopy\x2dcheck69025761-merged.mount: Deactivated successfully. Apr 23 23:58:21.639034 dockerd[1907]: time="2026-04-23T23:58:21.638861357Z" level=info msg="Loading containers: start." Apr 23 23:58:21.651039 kernel: Initializing XFRM netlink socket Apr 23 23:58:21.861631 systemd-timesyncd[1525]: Network configuration changed, trying to establish connection. Apr 23 23:58:22.844455 systemd-resolved[1504]: Clock change detected. Flushing caches. Apr 23 23:58:22.845099 systemd-timesyncd[1525]: Contacted time server 77.37.65.181:123 (2.flatcar.pool.ntp.org). Apr 23 23:58:22.845143 systemd-timesyncd[1525]: Initial clock synchronization to Thu 2026-04-23 23:58:22.844184 UTC. Apr 23 23:58:22.857409 systemd-networkd[1503]: docker0: Link UP Apr 23 23:58:22.861734 dockerd[1907]: time="2026-04-23T23:58:22.861703887Z" level=info msg="Loading containers: done." Apr 23 23:58:22.871982 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2032549456-merged.mount: Deactivated successfully. Apr 23 23:58:22.876373 dockerd[1907]: time="2026-04-23T23:58:22.876343392Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 23 23:58:22.876469 dockerd[1907]: time="2026-04-23T23:58:22.876408810Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Apr 23 23:58:22.876489 dockerd[1907]: time="2026-04-23T23:58:22.876471594Z" level=info msg="Initializing buildkit" Apr 23 23:58:22.896499 dockerd[1907]: time="2026-04-23T23:58:22.896475328Z" level=info msg="Completed buildkit initialization" Apr 23 23:58:22.901317 dockerd[1907]: time="2026-04-23T23:58:22.900527438Z" level=info msg="Daemon has completed initialization" Apr 23 23:58:22.901317 dockerd[1907]: time="2026-04-23T23:58:22.900644604Z" level=info msg="API listen on /run/docker.sock" Apr 23 23:58:22.900690 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 23 23:58:23.358621 containerd[1639]: time="2026-04-23T23:58:23.358519907Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.4\"" Apr 23 23:58:24.036093 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1260945439.mount: Deactivated successfully. Apr 23 23:58:25.199412 containerd[1639]: time="2026-04-23T23:58:25.199361907Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.35.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:25.200386 containerd[1639]: time="2026-04-23T23:58:25.200167075Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.35.4: active requests=0, bytes read=27579523" Apr 23 23:58:25.201293 containerd[1639]: time="2026-04-23T23:58:25.201269209Z" level=info msg="ImageCreate event name:\"sha256:840f22aa169cc9a11114a874832f60c2d4a4f7767d107303cd1ca6d9c228ee8b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:25.203240 containerd[1639]: time="2026-04-23T23:58:25.203216361Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:06b4bb208634a107ab9e6c50cdb9df178d05166a700c0cc448d59522091074b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:25.203930 containerd[1639]: time="2026-04-23T23:58:25.203903572Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.35.4\" with image id \"sha256:840f22aa169cc9a11114a874832f60c2d4a4f7767d107303cd1ca6d9c228ee8b\", repo tag \"registry.k8s.io/kube-apiserver:v1.35.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:06b4bb208634a107ab9e6c50cdb9df178d05166a700c0cc448d59522091074b5\", size \"27576022\" in 1.845293269s" Apr 23 23:58:25.203967 containerd[1639]: time="2026-04-23T23:58:25.203931484Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.35.4\" returns image reference \"sha256:840f22aa169cc9a11114a874832f60c2d4a4f7767d107303cd1ca6d9c228ee8b\"" Apr 23 23:58:25.204734 containerd[1639]: time="2026-04-23T23:58:25.204699055Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.4\"" Apr 23 23:58:26.405491 containerd[1639]: time="2026-04-23T23:58:26.405431294Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.35.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:26.406431 containerd[1639]: time="2026-04-23T23:58:26.406283132Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.35.4: active requests=0, bytes read=21451681" Apr 23 23:58:26.407352 containerd[1639]: time="2026-04-23T23:58:26.407330133Z" level=info msg="ImageCreate event name:\"sha256:96ce7469899d4d3ccad56b1a80b91609cb2203287112d73818296004948bb667\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:26.409505 containerd[1639]: time="2026-04-23T23:58:26.409478386Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7b036c805d57f203e9efaf43672cff6019b9083a9c0eb107ea8500eace29d8fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:26.410351 containerd[1639]: time="2026-04-23T23:58:26.410310575Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.35.4\" with image id \"sha256:96ce7469899d4d3ccad56b1a80b91609cb2203287112d73818296004948bb667\", repo tag \"registry.k8s.io/kube-controller-manager:v1.35.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7b036c805d57f203e9efaf43672cff6019b9083a9c0eb107ea8500eace29d8fd\", size \"23018006\" in 1.205556217s" Apr 23 23:58:26.410351 containerd[1639]: time="2026-04-23T23:58:26.410335893Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.35.4\" returns image reference \"sha256:96ce7469899d4d3ccad56b1a80b91609cb2203287112d73818296004948bb667\"" Apr 23 23:58:26.410964 containerd[1639]: time="2026-04-23T23:58:26.410932568Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.4\"" Apr 23 23:58:27.472310 containerd[1639]: time="2026-04-23T23:58:27.472260792Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.35.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:27.473443 containerd[1639]: time="2026-04-23T23:58:27.473297828Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.35.4: active requests=0, bytes read=15555312" Apr 23 23:58:27.474323 containerd[1639]: time="2026-04-23T23:58:27.474300943Z" level=info msg="ImageCreate event name:\"sha256:a0eecd9b69a38f829c29b535f73c1a3de3c7cc9f1294a44dc42c808faf0a23ff\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:27.476468 containerd[1639]: time="2026-04-23T23:58:27.476444279Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9054fecb4fa04cc63aec47b0913c8deb3487d414190cd15211f864cfe0d0b4d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:27.477198 containerd[1639]: time="2026-04-23T23:58:27.477173183Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.35.4\" with image id \"sha256:a0eecd9b69a38f829c29b535f73c1a3de3c7cc9f1294a44dc42c808faf0a23ff\", repo tag \"registry.k8s.io/kube-scheduler:v1.35.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9054fecb4fa04cc63aec47b0913c8deb3487d414190cd15211f864cfe0d0b4d6\", size \"17121655\" in 1.066221066s" Apr 23 23:58:27.477241 containerd[1639]: time="2026-04-23T23:58:27.477198741Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.35.4\" returns image reference \"sha256:a0eecd9b69a38f829c29b535f73c1a3de3c7cc9f1294a44dc42c808faf0a23ff\"" Apr 23 23:58:27.477623 containerd[1639]: time="2026-04-23T23:58:27.477589848Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.4\"" Apr 23 23:58:28.482740 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount568752115.mount: Deactivated successfully. Apr 23 23:58:28.677470 containerd[1639]: time="2026-04-23T23:58:28.677421656Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.35.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:28.678560 containerd[1639]: time="2026-04-23T23:58:28.678395657Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.35.4: active requests=0, bytes read=25699953" Apr 23 23:58:28.679384 containerd[1639]: time="2026-04-23T23:58:28.679361937Z" level=info msg="ImageCreate event name:\"sha256:f21f27cddb23d0d7131dc7c59666b3b0e0b5ca4c3f003225f90307ab6211b6e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:28.681079 containerd[1639]: time="2026-04-23T23:58:28.681062218Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c5daa23c72474e5e4062c320177d3b485fd42e7010f052bc80d657c4c00a0672\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:28.681358 containerd[1639]: time="2026-04-23T23:58:28.681337161Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.35.4\" with image id \"sha256:f21f27cddb23d0d7131dc7c59666b3b0e0b5ca4c3f003225f90307ab6211b6e1\", repo tag \"registry.k8s.io/kube-proxy:v1.35.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:c5daa23c72474e5e4062c320177d3b485fd42e7010f052bc80d657c4c00a0672\", size \"25698944\" in 1.203725039s" Apr 23 23:58:28.681387 containerd[1639]: time="2026-04-23T23:58:28.681362188Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.35.4\" returns image reference \"sha256:f21f27cddb23d0d7131dc7c59666b3b0e0b5ca4c3f003225f90307ab6211b6e1\"" Apr 23 23:58:28.681876 containerd[1639]: time="2026-04-23T23:58:28.681855749Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\"" Apr 23 23:58:29.193931 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1451421517.mount: Deactivated successfully. Apr 23 23:58:30.251952 containerd[1639]: time="2026-04-23T23:58:30.251900249Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.13.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.252904 containerd[1639]: time="2026-04-23T23:58:30.252864576Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.13.1: active requests=0, bytes read=23556642" Apr 23 23:58:30.253873 containerd[1639]: time="2026-04-23T23:58:30.253848322Z" level=info msg="ImageCreate event name:\"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.256409 containerd[1639]: time="2026-04-23T23:58:30.256392099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.257117 containerd[1639]: time="2026-04-23T23:58:30.257100712Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.13.1\" with image id \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\", repo tag \"registry.k8s.io/coredns/coredns:v1.13.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9b9128672209474da07c91439bf15ed704ae05ad918dd6454e5b6ae14e35fee6\", size \"23553139\" in 1.575225855s" Apr 23 23:58:30.257250 containerd[1639]: time="2026-04-23T23:58:30.257182024Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.13.1\" returns image reference \"sha256:aa5e3ebc0dfed0566805186b9e47110d8f9122291d8bad1497e78873ad291139\"" Apr 23 23:58:30.257808 containerd[1639]: time="2026-04-23T23:58:30.257776997Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Apr 23 23:58:30.738239 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount375340994.mount: Deactivated successfully. Apr 23 23:58:30.748198 containerd[1639]: time="2026-04-23T23:58:30.748084382Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.749368 containerd[1639]: time="2026-04-23T23:58:30.749221869Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321240" Apr 23 23:58:30.750455 containerd[1639]: time="2026-04-23T23:58:30.750410683Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.753706 containerd[1639]: time="2026-04-23T23:58:30.753601831Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:30.754594 containerd[1639]: time="2026-04-23T23:58:30.754519217Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 496.701269ms" Apr 23 23:58:30.754692 containerd[1639]: time="2026-04-23T23:58:30.754596413Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Apr 23 23:58:30.755844 containerd[1639]: time="2026-04-23T23:58:30.755811626Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\"" Apr 23 23:58:31.036746 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 23 23:58:31.040952 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:31.221296 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:31.236339 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1666188139.mount: Deactivated successfully. Apr 23 23:58:31.238027 (kubelet)[2263]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 23 23:58:31.279295 kubelet[2263]: E0423 23:58:31.279240 2263 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 23 23:58:31.282185 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 23 23:58:31.282347 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 23 23:58:31.282744 systemd[1]: kubelet.service: Consumed 195ms CPU time, 110M memory peak. Apr 23 23:58:31.955271 containerd[1639]: time="2026-04-23T23:58:31.955069633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.6-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:31.956018 containerd[1639]: time="2026-04-23T23:58:31.955907630Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.6-0: active requests=0, bytes read=23644553" Apr 23 23:58:31.956590 containerd[1639]: time="2026-04-23T23:58:31.956570745Z" level=info msg="ImageCreate event name:\"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:31.958355 containerd[1639]: time="2026-04-23T23:58:31.958336164Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:31.958890 containerd[1639]: time="2026-04-23T23:58:31.958873711Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.6-0\" with image id \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\", repo tag \"registry.k8s.io/etcd:3.6.6-0\", repo digest \"registry.k8s.io/etcd@sha256:60a30b5d81b2217555e2cfb9537f655b7ba97220b99c39ee2e162a7127225890\", size \"23641797\" in 1.203027733s" Apr 23 23:58:31.958948 containerd[1639]: time="2026-04-23T23:58:31.958938197Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.6-0\" returns image reference \"sha256:0a108f7189562e99793bdecab61fdf1a7c9d913af3385de9da17fb9d6ff430e2\"" Apr 23 23:58:32.983861 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:32.984004 systemd[1]: kubelet.service: Consumed 195ms CPU time, 110M memory peak. Apr 23 23:58:32.987727 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:33.018310 systemd[1]: Reload requested from client PID 2356 ('systemctl') (unit session-9.scope)... Apr 23 23:58:33.018320 systemd[1]: Reloading... Apr 23 23:58:33.115572 zram_generator::config[2400]: No configuration found. Apr 23 23:58:33.292493 systemd[1]: Reloading finished in 273 ms. Apr 23 23:58:33.337062 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 23 23:58:33.337146 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 23 23:58:33.337408 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:33.337443 systemd[1]: kubelet.service: Consumed 106ms CPU time, 98.3M memory peak. Apr 23 23:58:33.338710 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:33.505027 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:33.512372 (kubelet)[2454]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 23 23:58:33.558905 kubelet[2454]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 23 23:58:33.696695 kubelet[2454]: I0423 23:58:33.696631 2454 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Apr 23 23:58:33.696695 kubelet[2454]: I0423 23:58:33.696662 2454 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 23 23:58:33.696695 kubelet[2454]: I0423 23:58:33.696678 2454 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 23 23:58:33.696695 kubelet[2454]: I0423 23:58:33.696682 2454 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 23 23:58:33.696965 kubelet[2454]: I0423 23:58:33.696905 2454 server.go:951] "Client rotation is on, will bootstrap in background" Apr 23 23:58:33.704574 kubelet[2454]: I0423 23:58:33.704222 2454 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 23 23:58:33.706331 kubelet[2454]: E0423 23:58:33.706302 2454 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://135.181.109.116:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 135.181.109.116:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Apr 23 23:58:33.707258 kubelet[2454]: I0423 23:58:33.707236 2454 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 23 23:58:33.710461 kubelet[2454]: I0423 23:58:33.710413 2454 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 23 23:58:33.711192 kubelet[2454]: I0423 23:58:33.711153 2454 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 23 23:58:33.711286 kubelet[2454]: I0423 23:58:33.711175 2454 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-2-4-n-09e3934cf7","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 23 23:58:33.711286 kubelet[2454]: I0423 23:58:33.711276 2454 topology_manager.go:143] "Creating topology manager with none policy" Apr 23 23:58:33.711286 kubelet[2454]: I0423 23:58:33.711282 2454 container_manager_linux.go:308] "Creating device plugin manager" Apr 23 23:58:33.711583 kubelet[2454]: I0423 23:58:33.711340 2454 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Apr 23 23:58:33.713588 kubelet[2454]: I0423 23:58:33.713542 2454 state_mem.go:41] "Initialized" logger="CPUManager state memory" Apr 23 23:58:33.713754 kubelet[2454]: I0423 23:58:33.713722 2454 kubelet.go:482] "Attempting to sync node with API server" Apr 23 23:58:33.713754 kubelet[2454]: I0423 23:58:33.713754 2454 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 23 23:58:33.713961 kubelet[2454]: I0423 23:58:33.713771 2454 kubelet.go:394] "Adding apiserver pod source" Apr 23 23:58:33.713961 kubelet[2454]: I0423 23:58:33.713778 2454 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 23 23:58:33.717269 kubelet[2454]: I0423 23:58:33.717239 2454 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Apr 23 23:58:33.719579 kubelet[2454]: I0423 23:58:33.719482 2454 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 23 23:58:33.719579 kubelet[2454]: I0423 23:58:33.719505 2454 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 23 23:58:33.719579 kubelet[2454]: W0423 23:58:33.719562 2454 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 23 23:58:33.722159 kubelet[2454]: I0423 23:58:33.722134 2454 server.go:1257] "Started kubelet" Apr 23 23:58:33.730743 kubelet[2454]: I0423 23:58:33.730681 2454 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Apr 23 23:58:33.737990 kubelet[2454]: I0423 23:58:33.737961 2454 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 23 23:58:33.739799 kubelet[2454]: I0423 23:58:33.739775 2454 volume_manager.go:311] "Starting Kubelet Volume Manager" Apr 23 23:58:33.739903 kubelet[2454]: I0423 23:58:33.735191 2454 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Apr 23 23:58:33.740449 kubelet[2454]: I0423 23:58:33.740421 2454 server.go:317] "Adding debug handlers to kubelet server" Apr 23 23:58:33.741034 kubelet[2454]: E0423 23:58:33.741007 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:33.741350 kubelet[2454]: I0423 23:58:33.741322 2454 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 23 23:58:33.741403 kubelet[2454]: I0423 23:58:33.741358 2454 reconciler.go:29] "Reconciler: start to sync state" Apr 23 23:58:33.741591 kubelet[2454]: I0423 23:58:33.741502 2454 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 23 23:58:33.741591 kubelet[2454]: I0423 23:58:33.741564 2454 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 23 23:58:33.741760 kubelet[2454]: I0423 23:58:33.741739 2454 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 23 23:58:33.742243 kubelet[2454]: E0423 23:58:33.741867 2454 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://135.181.109.116:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-4-n-09e3934cf7?timeout=10s\": dial tcp 135.181.109.116:6443: connect: connection refused" interval="200ms" Apr 23 23:58:33.743240 kubelet[2454]: E0423 23:58:33.741943 2454 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://135.181.109.116:6443/api/v1/namespaces/default/events\": dial tcp 135.181.109.116:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459-2-4-n-09e3934cf7.18a921d1fb48ff54 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459-2-4-n-09e3934cf7,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459-2-4-n-09e3934cf7,},FirstTimestamp:2026-04-23 23:58:33.722117972 +0000 UTC m=+0.202863141,LastTimestamp:2026-04-23 23:58:33.722117972 +0000 UTC m=+0.202863141,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-2-4-n-09e3934cf7,}" Apr 23 23:58:33.744198 kubelet[2454]: I0423 23:58:33.744165 2454 factory.go:223] Registration of the systemd container factory successfully Apr 23 23:58:33.744322 kubelet[2454]: I0423 23:58:33.744224 2454 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 23 23:58:33.745075 kubelet[2454]: I0423 23:58:33.745050 2454 factory.go:223] Registration of the containerd container factory successfully Apr 23 23:58:33.746877 kubelet[2454]: I0423 23:58:33.746642 2454 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 23 23:58:33.750012 kubelet[2454]: E0423 23:58:33.749988 2454 kubelet.go:1656] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 23 23:58:33.771660 kubelet[2454]: I0423 23:58:33.771283 2454 cpu_manager.go:225] "Starting" policy="none" Apr 23 23:58:33.771660 kubelet[2454]: I0423 23:58:33.771300 2454 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Apr 23 23:58:33.771660 kubelet[2454]: I0423 23:58:33.771321 2454 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Apr 23 23:58:33.772187 kubelet[2454]: I0423 23:58:33.772165 2454 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 23 23:58:33.772187 kubelet[2454]: I0423 23:58:33.772183 2454 status_manager.go:249] "Starting to sync pod status with apiserver" Apr 23 23:58:33.772271 kubelet[2454]: I0423 23:58:33.772197 2454 kubelet.go:2501] "Starting kubelet main sync loop" Apr 23 23:58:33.772271 kubelet[2454]: E0423 23:58:33.772247 2454 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 23 23:58:33.776365 kubelet[2454]: I0423 23:58:33.776348 2454 policy_none.go:50] "Start" Apr 23 23:58:33.776571 kubelet[2454]: I0423 23:58:33.776445 2454 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 23 23:58:33.776571 kubelet[2454]: I0423 23:58:33.776462 2454 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 23 23:58:33.778382 kubelet[2454]: I0423 23:58:33.778346 2454 policy_none.go:44] "Start" Apr 23 23:58:33.784082 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 23 23:58:33.796855 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 23 23:58:33.801177 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 23 23:58:33.808739 kubelet[2454]: E0423 23:58:33.808717 2454 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 23 23:58:33.808966 kubelet[2454]: I0423 23:58:33.808896 2454 eviction_manager.go:194] "Eviction manager: starting control loop" Apr 23 23:58:33.808966 kubelet[2454]: I0423 23:58:33.808908 2454 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 23 23:58:33.810334 kubelet[2454]: I0423 23:58:33.809295 2454 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Apr 23 23:58:33.810334 kubelet[2454]: E0423 23:58:33.810189 2454 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 23 23:58:33.810334 kubelet[2454]: E0423 23:58:33.810232 2454 eviction_manager.go:297] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:33.885684 systemd[1]: Created slice kubepods-burstable-pod53c5f6c6d24b0d84b374cb52cf7e654c.slice - libcontainer container kubepods-burstable-pod53c5f6c6d24b0d84b374cb52cf7e654c.slice. Apr 23 23:58:33.887602 kubelet[2454]: W0423 23:58:33.887569 2454 helpers.go:246] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53c5f6c6d24b0d84b374cb52cf7e654c.slice/cpuset.cpus.effective": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53c5f6c6d24b0d84b374cb52cf7e654c.slice/cpuset.cpus.effective: no such device Apr 23 23:58:33.895825 kubelet[2454]: E0423 23:58:33.895708 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.899491 systemd[1]: Created slice kubepods-burstable-pod1c983f43ebc04ba27c56d85868a5d483.slice - libcontainer container kubepods-burstable-pod1c983f43ebc04ba27c56d85868a5d483.slice. Apr 23 23:58:33.902891 kubelet[2454]: E0423 23:58:33.902858 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.906776 systemd[1]: Created slice kubepods-burstable-pod546ddbe225645b866e6103335656f783.slice - libcontainer container kubepods-burstable-pod546ddbe225645b866e6103335656f783.slice. Apr 23 23:58:33.911674 kubelet[2454]: E0423 23:58:33.911648 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.913748 kubelet[2454]: I0423 23:58:33.913722 2454 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.914251 kubelet[2454]: E0423 23:58:33.914223 2454 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://135.181.109.116:6443/api/v1/nodes\": dial tcp 135.181.109.116:6443: connect: connection refused" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.941895 kubelet[2454]: I0423 23:58:33.941852 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-ca-certs\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.941895 kubelet[2454]: I0423 23:58:33.941894 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-ca-certs\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942341 kubelet[2454]: I0423 23:58:33.941923 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-kubeconfig\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942341 kubelet[2454]: I0423 23:58:33.941954 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/546ddbe225645b866e6103335656f783-kubeconfig\") pod \"kube-scheduler-ci-4459-2-4-n-09e3934cf7\" (UID: \"546ddbe225645b866e6103335656f783\") " pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942341 kubelet[2454]: I0423 23:58:33.941977 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-k8s-certs\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942341 kubelet[2454]: I0423 23:58:33.942000 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942341 kubelet[2454]: I0423 23:58:33.942020 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942579 kubelet[2454]: I0423 23:58:33.942041 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-k8s-certs\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942579 kubelet[2454]: I0423 23:58:33.942063 2454 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:33.942579 kubelet[2454]: E0423 23:58:33.942309 2454 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://135.181.109.116:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-4-n-09e3934cf7?timeout=10s\": dial tcp 135.181.109.116:6443: connect: connection refused" interval="400ms" Apr 23 23:58:34.117260 kubelet[2454]: I0423 23:58:34.117099 2454 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:34.118644 kubelet[2454]: E0423 23:58:34.118496 2454 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://135.181.109.116:6443/api/v1/nodes\": dial tcp 135.181.109.116:6443: connect: connection refused" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:34.201047 containerd[1639]: time="2026-04-23T23:58:34.200946213Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-2-4-n-09e3934cf7,Uid:53c5f6c6d24b0d84b374cb52cf7e654c,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:34.207233 containerd[1639]: time="2026-04-23T23:58:34.207171804Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-2-4-n-09e3934cf7,Uid:1c983f43ebc04ba27c56d85868a5d483,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:34.218461 containerd[1639]: time="2026-04-23T23:58:34.218388434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-2-4-n-09e3934cf7,Uid:546ddbe225645b866e6103335656f783,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:34.342888 kubelet[2454]: E0423 23:58:34.342840 2454 controller.go:201] "Failed to ensure lease exists, will retry" err="Get \"https://135.181.109.116:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-4-n-09e3934cf7?timeout=10s\": dial tcp 135.181.109.116:6443: connect: connection refused" interval="800ms" Apr 23 23:58:34.521852 kubelet[2454]: I0423 23:58:34.521666 2454 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:34.522384 kubelet[2454]: E0423 23:58:34.522325 2454 kubelet_node_status.go:106] "Unable to register node with API server" err="Post \"https://135.181.109.116:6443/api/v1/nodes\": dial tcp 135.181.109.116:6443: connect: connection refused" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:34.678863 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount62429453.mount: Deactivated successfully. Apr 23 23:58:34.688815 containerd[1639]: time="2026-04-23T23:58:34.688735631Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:58:34.690904 containerd[1639]: time="2026-04-23T23:58:34.690846638Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:58:34.693141 containerd[1639]: time="2026-04-23T23:58:34.693060249Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321160" Apr 23 23:58:34.694182 containerd[1639]: time="2026-04-23T23:58:34.694126259Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Apr 23 23:58:34.696579 containerd[1639]: time="2026-04-23T23:58:34.696500932Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:58:34.698636 containerd[1639]: time="2026-04-23T23:58:34.698471559Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:58:34.698803 containerd[1639]: time="2026-04-23T23:58:34.698765920Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Apr 23 23:58:34.701773 containerd[1639]: time="2026-04-23T23:58:34.701735877Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 23 23:58:34.703739 containerd[1639]: time="2026-04-23T23:58:34.703676368Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 483.589015ms" Apr 23 23:58:34.705452 containerd[1639]: time="2026-04-23T23:58:34.705374296Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 496.623192ms" Apr 23 23:58:34.707012 containerd[1639]: time="2026-04-23T23:58:34.706961617Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 503.42645ms" Apr 23 23:58:34.750561 containerd[1639]: time="2026-04-23T23:58:34.750137800Z" level=info msg="connecting to shim 754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f" address="unix:///run/containerd/s/a89b511aed31b4b13dce274d6ebfa1286ada34a47b8a79bb77acf539f027b834" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:34.754983 containerd[1639]: time="2026-04-23T23:58:34.754955388Z" level=info msg="connecting to shim a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d" address="unix:///run/containerd/s/ec2d4c223ba2b2c0ebe8c336c7bbfc133864dd5ec3a7b768364029eb920afdf2" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:34.760067 containerd[1639]: time="2026-04-23T23:58:34.760030523Z" level=info msg="connecting to shim c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad" address="unix:///run/containerd/s/ecfddf6f77c4f6e0c148e9c1ab5bb9e1113bf51c86d51e263e3714c31a455133" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:34.772654 systemd[1]: Started cri-containerd-754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f.scope - libcontainer container 754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f. Apr 23 23:58:34.793668 systemd[1]: Started cri-containerd-a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d.scope - libcontainer container a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d. Apr 23 23:58:34.795183 systemd[1]: Started cri-containerd-c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad.scope - libcontainer container c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad. Apr 23 23:58:34.850549 containerd[1639]: time="2026-04-23T23:58:34.849914166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459-2-4-n-09e3934cf7,Uid:53c5f6c6d24b0d84b374cb52cf7e654c,Namespace:kube-system,Attempt:0,} returns sandbox id \"a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d\"" Apr 23 23:58:34.851051 containerd[1639]: time="2026-04-23T23:58:34.851037391Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459-2-4-n-09e3934cf7,Uid:546ddbe225645b866e6103335656f783,Namespace:kube-system,Attempt:0,} returns sandbox id \"754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f\"" Apr 23 23:58:34.857162 containerd[1639]: time="2026-04-23T23:58:34.857051334Z" level=info msg="CreateContainer within sandbox \"a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 23 23:58:34.857788 containerd[1639]: time="2026-04-23T23:58:34.857767729Z" level=info msg="CreateContainer within sandbox \"754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 23 23:58:34.863184 containerd[1639]: time="2026-04-23T23:58:34.863164947Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459-2-4-n-09e3934cf7,Uid:1c983f43ebc04ba27c56d85868a5d483,Namespace:kube-system,Attempt:0,} returns sandbox id \"c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad\"" Apr 23 23:58:34.865628 containerd[1639]: time="2026-04-23T23:58:34.865595364Z" level=info msg="Container 3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:34.866498 containerd[1639]: time="2026-04-23T23:58:34.866477117Z" level=info msg="CreateContainer within sandbox \"c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 23 23:58:34.868695 containerd[1639]: time="2026-04-23T23:58:34.868672952Z" level=info msg="Container 6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:34.872640 containerd[1639]: time="2026-04-23T23:58:34.872585622Z" level=info msg="CreateContainer within sandbox \"754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c\"" Apr 23 23:58:34.873106 containerd[1639]: time="2026-04-23T23:58:34.873093434Z" level=info msg="StartContainer for \"3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c\"" Apr 23 23:58:34.874705 containerd[1639]: time="2026-04-23T23:58:34.874688357Z" level=info msg="connecting to shim 3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c" address="unix:///run/containerd/s/a89b511aed31b4b13dce274d6ebfa1286ada34a47b8a79bb77acf539f027b834" protocol=ttrpc version=3 Apr 23 23:58:34.879881 containerd[1639]: time="2026-04-23T23:58:34.879826356Z" level=info msg="CreateContainer within sandbox \"a060257d9cc11b9fa59c45f9dafdc628e1b3dd29d156ccadfda1100cb18ef28d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0\"" Apr 23 23:58:34.880892 containerd[1639]: time="2026-04-23T23:58:34.880856322Z" level=info msg="StartContainer for \"6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0\"" Apr 23 23:58:34.881519 containerd[1639]: time="2026-04-23T23:58:34.881500298Z" level=info msg="connecting to shim 6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0" address="unix:///run/containerd/s/ec2d4c223ba2b2c0ebe8c336c7bbfc133864dd5ec3a7b768364029eb920afdf2" protocol=ttrpc version=3 Apr 23 23:58:34.886270 containerd[1639]: time="2026-04-23T23:58:34.886255282Z" level=info msg="Container 00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:34.893794 systemd[1]: Started cri-containerd-3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c.scope - libcontainer container 3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c. Apr 23 23:58:34.897153 containerd[1639]: time="2026-04-23T23:58:34.897078422Z" level=info msg="CreateContainer within sandbox \"c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076\"" Apr 23 23:58:34.897860 containerd[1639]: time="2026-04-23T23:58:34.897463920Z" level=info msg="StartContainer for \"00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076\"" Apr 23 23:58:34.898211 containerd[1639]: time="2026-04-23T23:58:34.898190360Z" level=info msg="connecting to shim 00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076" address="unix:///run/containerd/s/ecfddf6f77c4f6e0c148e9c1ab5bb9e1113bf51c86d51e263e3714c31a455133" protocol=ttrpc version=3 Apr 23 23:58:34.905292 systemd[1]: Started cri-containerd-6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0.scope - libcontainer container 6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0. Apr 23 23:58:34.919689 systemd[1]: Started cri-containerd-00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076.scope - libcontainer container 00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076. Apr 23 23:58:34.966108 containerd[1639]: time="2026-04-23T23:58:34.966071881Z" level=info msg="StartContainer for \"3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c\" returns successfully" Apr 23 23:58:34.979887 containerd[1639]: time="2026-04-23T23:58:34.979858807Z" level=info msg="StartContainer for \"6e11e716fb56b34cdd322dfedd0330391fc27605ca505c0df8f7b107143db4e0\" returns successfully" Apr 23 23:58:34.991469 containerd[1639]: time="2026-04-23T23:58:34.991380304Z" level=info msg="StartContainer for \"00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076\" returns successfully" Apr 23 23:58:35.325608 kubelet[2454]: I0423 23:58:35.324702 2454 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.607491 kubelet[2454]: E0423 23:58:35.607396 2454 nodelease.go:50] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.786249 kubelet[2454]: I0423 23:58:35.786142 2454 kubelet_node_status.go:77] "Successfully registered node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.786249 kubelet[2454]: E0423 23:58:35.786163 2454 kubelet_node_status.go:474] "Error updating node status, will retry" err="error getting node \"ci-4459-2-4-n-09e3934cf7\": node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:35.788141 kubelet[2454]: E0423 23:58:35.788111 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.791710 kubelet[2454]: E0423 23:58:35.791680 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.793446 kubelet[2454]: E0423 23:58:35.793416 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:35.798038 kubelet[2454]: E0423 23:58:35.798006 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:35.899006 kubelet[2454]: E0423 23:58:35.898910 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:35.999838 kubelet[2454]: E0423 23:58:35.999770 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.101001 kubelet[2454]: E0423 23:58:36.100920 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.201906 kubelet[2454]: E0423 23:58:36.201748 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.302132 kubelet[2454]: E0423 23:58:36.302066 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.403080 kubelet[2454]: E0423 23:58:36.402994 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.504064 kubelet[2454]: E0423 23:58:36.503889 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.604898 kubelet[2454]: E0423 23:58:36.604855 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.705199 kubelet[2454]: E0423 23:58:36.705163 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.797935 kubelet[2454]: E0423 23:58:36.797748 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:36.798457 kubelet[2454]: E0423 23:58:36.798413 2454 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:36.805874 kubelet[2454]: E0423 23:58:36.805848 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:36.906873 kubelet[2454]: E0423 23:58:36.906788 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:37.007069 kubelet[2454]: E0423 23:58:37.007016 2454 kubelet_node_status.go:392] "Error getting the current node from lister" err="node \"ci-4459-2-4-n-09e3934cf7\" not found" Apr 23 23:58:37.141823 kubelet[2454]: I0423 23:58:37.141261 2454 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:37.151652 kubelet[2454]: I0423 23:58:37.151616 2454 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:37.157491 kubelet[2454]: I0423 23:58:37.157282 2454 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:37.719624 kubelet[2454]: I0423 23:58:37.719415 2454 apiserver.go:52] "Watching apiserver" Apr 23 23:58:37.741841 kubelet[2454]: I0423 23:58:37.741796 2454 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 23 23:58:37.798043 kubelet[2454]: I0423 23:58:37.797940 2454 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:37.804998 kubelet[2454]: E0423 23:58:37.804909 2454 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-4-n-09e3934cf7\" already exists" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:37.978163 systemd[1]: Reload requested from client PID 2732 ('systemctl') (unit session-9.scope)... Apr 23 23:58:37.978192 systemd[1]: Reloading... Apr 23 23:58:38.085594 zram_generator::config[2779]: No configuration found. Apr 23 23:58:38.268203 systemd[1]: Reloading finished in 289 ms. Apr 23 23:58:38.291988 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:38.313909 systemd[1]: kubelet.service: Deactivated successfully. Apr 23 23:58:38.314171 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:38.314223 systemd[1]: kubelet.service: Consumed 572ms CPU time, 123.3M memory peak. Apr 23 23:58:38.316477 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 23 23:58:38.488643 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 23 23:58:38.502118 (kubelet)[2827]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 23 23:58:38.544201 kubelet[2827]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 23 23:58:38.554157 kubelet[2827]: I0423 23:58:38.554115 2827 server.go:525] "Kubelet version" kubeletVersion="v1.35.1" Apr 23 23:58:38.554157 kubelet[2827]: I0423 23:58:38.554147 2827 server.go:527] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 23 23:58:38.554157 kubelet[2827]: I0423 23:58:38.554163 2827 watchdog_linux.go:95] "Systemd watchdog is not enabled" Apr 23 23:58:38.554157 kubelet[2827]: I0423 23:58:38.554168 2827 watchdog_linux.go:138] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 23 23:58:38.554376 kubelet[2827]: I0423 23:58:38.554343 2827 server.go:951] "Client rotation is on, will bootstrap in background" Apr 23 23:58:38.555218 kubelet[2827]: I0423 23:58:38.555200 2827 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Apr 23 23:58:38.556687 kubelet[2827]: I0423 23:58:38.556671 2827 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 23 23:58:38.559820 kubelet[2827]: I0423 23:58:38.559786 2827 server.go:1418] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 23 23:58:38.563487 kubelet[2827]: I0423 23:58:38.563428 2827 server.go:775] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Apr 23 23:58:38.563761 kubelet[2827]: I0423 23:58:38.563737 2827 container_manager_linux.go:272] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 23 23:58:38.563930 kubelet[2827]: I0423 23:58:38.563811 2827 container_manager_linux.go:277] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459-2-4-n-09e3934cf7","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 23 23:58:38.564250 kubelet[2827]: I0423 23:58:38.564051 2827 topology_manager.go:143] "Creating topology manager with none policy" Apr 23 23:58:38.564250 kubelet[2827]: I0423 23:58:38.564060 2827 container_manager_linux.go:308] "Creating device plugin manager" Apr 23 23:58:38.564250 kubelet[2827]: I0423 23:58:38.564077 2827 container_manager_linux.go:317] "Creating Dynamic Resource Allocation (DRA) manager" Apr 23 23:58:38.564250 kubelet[2827]: I0423 23:58:38.564219 2827 state_mem.go:41] "Initialized" logger="CPUManager state memory" Apr 23 23:58:38.564437 kubelet[2827]: I0423 23:58:38.564429 2827 kubelet.go:482] "Attempting to sync node with API server" Apr 23 23:58:38.564479 kubelet[2827]: I0423 23:58:38.564474 2827 kubelet.go:383] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 23 23:58:38.564519 kubelet[2827]: I0423 23:58:38.564513 2827 kubelet.go:394] "Adding apiserver pod source" Apr 23 23:58:38.564572 kubelet[2827]: I0423 23:58:38.564566 2827 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 23 23:58:38.565805 kubelet[2827]: I0423 23:58:38.565645 2827 kuberuntime_manager.go:294] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Apr 23 23:58:38.567710 kubelet[2827]: I0423 23:58:38.567390 2827 kubelet.go:943] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 23 23:58:38.567710 kubelet[2827]: I0423 23:58:38.567418 2827 kubelet.go:970] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Apr 23 23:58:38.579044 kubelet[2827]: I0423 23:58:38.579029 2827 server.go:1257] "Started kubelet" Apr 23 23:58:38.581278 kubelet[2827]: I0423 23:58:38.581257 2827 fs_resource_analyzer.go:69] "Starting FS ResourceAnalyzer" Apr 23 23:58:38.583081 kubelet[2827]: I0423 23:58:38.582955 2827 server.go:182] "Starting to listen" address="0.0.0.0" port=10250 Apr 23 23:58:38.584446 kubelet[2827]: I0423 23:58:38.584434 2827 server.go:317] "Adding debug handlers to kubelet server" Apr 23 23:58:38.588150 kubelet[2827]: I0423 23:58:38.588114 2827 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 23 23:58:38.588299 kubelet[2827]: I0423 23:58:38.588290 2827 server_v1.go:49] "podresources" method="list" useActivePods=true Apr 23 23:58:38.588456 kubelet[2827]: I0423 23:58:38.588448 2827 server.go:254] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 23 23:58:38.590022 kubelet[2827]: I0423 23:58:38.589614 2827 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 23 23:58:38.592773 kubelet[2827]: I0423 23:58:38.592380 2827 volume_manager.go:311] "Starting Kubelet Volume Manager" Apr 23 23:58:38.594834 kubelet[2827]: I0423 23:58:38.593621 2827 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 23 23:58:38.594834 kubelet[2827]: I0423 23:58:38.593715 2827 reconciler.go:29] "Reconciler: start to sync state" Apr 23 23:58:38.594834 kubelet[2827]: I0423 23:58:38.594168 2827 factory.go:223] Registration of the systemd container factory successfully Apr 23 23:58:38.594834 kubelet[2827]: I0423 23:58:38.594238 2827 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 23 23:58:38.597675 kubelet[2827]: I0423 23:58:38.597648 2827 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Apr 23 23:58:38.597768 kubelet[2827]: I0423 23:58:38.597756 2827 factory.go:223] Registration of the containerd container factory successfully Apr 23 23:58:38.599063 kubelet[2827]: I0423 23:58:38.599045 2827 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Apr 23 23:58:38.599063 kubelet[2827]: I0423 23:58:38.599062 2827 status_manager.go:249] "Starting to sync pod status with apiserver" Apr 23 23:58:38.599120 kubelet[2827]: I0423 23:58:38.599078 2827 kubelet.go:2501] "Starting kubelet main sync loop" Apr 23 23:58:38.599120 kubelet[2827]: E0423 23:58:38.599114 2827 kubelet.go:2525] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 23 23:58:38.644034 kubelet[2827]: I0423 23:58:38.644003 2827 cpu_manager.go:225] "Starting" policy="none" Apr 23 23:58:38.644034 kubelet[2827]: I0423 23:58:38.644020 2827 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Apr 23 23:58:38.644034 kubelet[2827]: I0423 23:58:38.644037 2827 state_mem.go:41] "Initialized" logger="CPUManager state checkpoint.CPUManager state memory" Apr 23 23:58:38.644188 kubelet[2827]: I0423 23:58:38.644165 2827 state_mem.go:94] "Updated default CPUSet" logger="CPUManager state checkpoint.CPUManager state memory" cpuSet="" Apr 23 23:58:38.644188 kubelet[2827]: I0423 23:58:38.644177 2827 state_mem.go:102] "Updated CPUSet assignments" logger="CPUManager state checkpoint.CPUManager state memory" assignments={} Apr 23 23:58:38.644224 kubelet[2827]: I0423 23:58:38.644193 2827 policy_none.go:50] "Start" Apr 23 23:58:38.644224 kubelet[2827]: I0423 23:58:38.644202 2827 memory_manager.go:187] "Starting memorymanager" policy="None" Apr 23 23:58:38.644224 kubelet[2827]: I0423 23:58:38.644212 2827 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Apr 23 23:58:38.644310 kubelet[2827]: I0423 23:58:38.644292 2827 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Apr 23 23:58:38.644310 kubelet[2827]: I0423 23:58:38.644306 2827 policy_none.go:44] "Start" Apr 23 23:58:38.649404 kubelet[2827]: E0423 23:58:38.649380 2827 manager.go:525] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 23 23:58:38.649544 kubelet[2827]: I0423 23:58:38.649514 2827 eviction_manager.go:194] "Eviction manager: starting control loop" Apr 23 23:58:38.649593 kubelet[2827]: I0423 23:58:38.649526 2827 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 23 23:58:38.649947 kubelet[2827]: I0423 23:58:38.649913 2827 plugin_manager.go:121] "Starting Kubelet Plugin Manager" Apr 23 23:58:38.652822 kubelet[2827]: E0423 23:58:38.652801 2827 eviction_manager.go:272] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 23 23:58:38.700421 kubelet[2827]: I0423 23:58:38.700394 2827 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.700840 kubelet[2827]: I0423 23:58:38.700602 2827 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.700840 kubelet[2827]: I0423 23:58:38.700473 2827 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.708224 kubelet[2827]: E0423 23:58:38.708204 2827 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" already exists" pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.708781 kubelet[2827]: E0423 23:58:38.708771 2827 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-4-n-09e3934cf7\" already exists" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.708959 kubelet[2827]: E0423 23:58:38.708949 2827 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" already exists" pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.758188 kubelet[2827]: I0423 23:58:38.758115 2827 kubelet_node_status.go:74] "Attempting to register node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.767584 kubelet[2827]: I0423 23:58:38.767322 2827 kubelet_node_status.go:123] "Node was previously registered" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.767584 kubelet[2827]: I0423 23:58:38.767409 2827 kubelet_node_status.go:77] "Successfully registered node" node="ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.895959 kubelet[2827]: I0423 23:58:38.895778 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-ca-certs\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.895959 kubelet[2827]: I0423 23:58:38.895805 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-flexvolume-dir\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.895959 kubelet[2827]: I0423 23:58:38.895816 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-k8s-certs\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.895959 kubelet[2827]: I0423 23:58:38.895837 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.895959 kubelet[2827]: I0423 23:58:38.895852 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/546ddbe225645b866e6103335656f783-kubeconfig\") pod \"kube-scheduler-ci-4459-2-4-n-09e3934cf7\" (UID: \"546ddbe225645b866e6103335656f783\") " pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.896244 kubelet[2827]: I0423 23:58:38.895862 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-ca-certs\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.896244 kubelet[2827]: I0423 23:58:38.895872 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.896244 kubelet[2827]: I0423 23:58:38.895882 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1c983f43ebc04ba27c56d85868a5d483-kubeconfig\") pod \"kube-controller-manager-ci-4459-2-4-n-09e3934cf7\" (UID: \"1c983f43ebc04ba27c56d85868a5d483\") " pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.896244 kubelet[2827]: I0423 23:58:38.895892 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/53c5f6c6d24b0d84b374cb52cf7e654c-k8s-certs\") pod \"kube-apiserver-ci-4459-2-4-n-09e3934cf7\" (UID: \"53c5f6c6d24b0d84b374cb52cf7e654c\") " pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:38.974163 sudo[2866]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Apr 23 23:58:38.974965 sudo[2866]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Apr 23 23:58:39.240026 sudo[2866]: pam_unix(sudo:session): session closed for user root Apr 23 23:58:39.566511 kubelet[2827]: I0423 23:58:39.565375 2827 apiserver.go:52] "Watching apiserver" Apr 23 23:58:39.594718 kubelet[2827]: I0423 23:58:39.594672 2827 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 23 23:58:39.633737 kubelet[2827]: I0423 23:58:39.633223 2827 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:39.648566 kubelet[2827]: E0423 23:58:39.647561 2827 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459-2-4-n-09e3934cf7\" already exists" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" Apr 23 23:58:39.683864 kubelet[2827]: I0423 23:58:39.683803 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459-2-4-n-09e3934cf7" podStartSLOduration=2.6837901029999998 podStartE2EDuration="2.683790103s" podCreationTimestamp="2026-04-23 23:58:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:39.67125795 +0000 UTC m=+1.162425424" watchObservedRunningTime="2026-04-23 23:58:39.683790103 +0000 UTC m=+1.174957537" Apr 23 23:58:39.693026 kubelet[2827]: I0423 23:58:39.692988 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459-2-4-n-09e3934cf7" podStartSLOduration=2.692981634 podStartE2EDuration="2.692981634s" podCreationTimestamp="2026-04-23 23:58:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:39.684681561 +0000 UTC m=+1.175848985" watchObservedRunningTime="2026-04-23 23:58:39.692981634 +0000 UTC m=+1.184149058" Apr 23 23:58:39.912683 kubelet[2827]: I0423 23:58:39.912454 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459-2-4-n-09e3934cf7" podStartSLOduration=2.9124352460000003 podStartE2EDuration="2.912435246s" podCreationTimestamp="2026-04-23 23:58:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:39.693335295 +0000 UTC m=+1.184502719" watchObservedRunningTime="2026-04-23 23:58:39.912435246 +0000 UTC m=+1.403602710" Apr 23 23:58:40.554961 sudo[1889]: pam_unix(sudo:session): session closed for user root Apr 23 23:58:40.584834 sshd[1888]: Connection closed by 20.229.252.112 port 38250 Apr 23 23:58:40.585299 sshd-session[1885]: pam_unix(sshd:session): session closed for user core Apr 23 23:58:40.588843 systemd-logind[1616]: Session 9 logged out. Waiting for processes to exit. Apr 23 23:58:40.589495 systemd[1]: sshd@8-135.181.109.116:22-20.229.252.112:38250.service: Deactivated successfully. Apr 23 23:58:40.591471 systemd[1]: session-9.scope: Deactivated successfully. Apr 23 23:58:40.591667 systemd[1]: session-9.scope: Consumed 2.720s CPU time, 270.3M memory peak. Apr 23 23:58:40.593401 systemd-logind[1616]: Removed session 9. Apr 23 23:58:42.635815 kubelet[2827]: I0423 23:58:42.635687 2827 kuberuntime_manager.go:2062] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 23 23:58:42.637692 containerd[1639]: time="2026-04-23T23:58:42.637602943Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 23 23:58:42.638718 kubelet[2827]: I0423 23:58:42.638070 2827 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 23 23:58:43.622947 systemd[1]: Created slice kubepods-besteffort-poda9157fb6_20d5_46ae_82e4_759c22b53154.slice - libcontainer container kubepods-besteffort-poda9157fb6_20d5_46ae_82e4_759c22b53154.slice. Apr 23 23:58:43.624044 kubelet[2827]: I0423 23:58:43.623337 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a9157fb6-20d5-46ae-82e4-759c22b53154-xtables-lock\") pod \"kube-proxy-5s9nh\" (UID: \"a9157fb6-20d5-46ae-82e4-759c22b53154\") " pod="kube-system/kube-proxy-5s9nh" Apr 23 23:58:43.624044 kubelet[2827]: I0423 23:58:43.623388 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9157fb6-20d5-46ae-82e4-759c22b53154-lib-modules\") pod \"kube-proxy-5s9nh\" (UID: \"a9157fb6-20d5-46ae-82e4-759c22b53154\") " pod="kube-system/kube-proxy-5s9nh" Apr 23 23:58:43.624044 kubelet[2827]: I0423 23:58:43.623412 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g956\" (UniqueName: \"kubernetes.io/projected/a9157fb6-20d5-46ae-82e4-759c22b53154-kube-api-access-5g956\") pod \"kube-proxy-5s9nh\" (UID: \"a9157fb6-20d5-46ae-82e4-759c22b53154\") " pod="kube-system/kube-proxy-5s9nh" Apr 23 23:58:43.624044 kubelet[2827]: I0423 23:58:43.623437 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a9157fb6-20d5-46ae-82e4-759c22b53154-kube-proxy\") pod \"kube-proxy-5s9nh\" (UID: \"a9157fb6-20d5-46ae-82e4-759c22b53154\") " pod="kube-system/kube-proxy-5s9nh" Apr 23 23:58:43.645258 systemd[1]: Created slice kubepods-burstable-pod4239b600_718c_4e02_9846_e3af0a0fc526.slice - libcontainer container kubepods-burstable-pod4239b600_718c_4e02_9846_e3af0a0fc526.slice. Apr 23 23:58:43.724597 kubelet[2827]: I0423 23:58:43.724569 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cni-path\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725146 kubelet[2827]: I0423 23:58:43.725135 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-lib-modules\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725226 kubelet[2827]: I0423 23:58:43.725218 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-kernel\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725275 kubelet[2827]: I0423 23:58:43.725264 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj57b\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-kube-api-access-jj57b\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725380 kubelet[2827]: I0423 23:58:43.725346 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-run\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725424 kubelet[2827]: I0423 23:58:43.725415 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-hostproc\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725507 kubelet[2827]: I0423 23:58:43.725473 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-cgroup\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725600 kubelet[2827]: I0423 23:58:43.725581 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-bpf-maps\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725664 kubelet[2827]: I0423 23:58:43.725657 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-etc-cni-netd\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725707 kubelet[2827]: I0423 23:58:43.725699 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-xtables-lock\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725763 kubelet[2827]: I0423 23:58:43.725755 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4239b600-718c-4e02-9846-e3af0a0fc526-clustermesh-secrets\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725799 kubelet[2827]: I0423 23:58:43.725792 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-net\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725918 kubelet[2827]: I0423 23:58:43.725841 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-hubble-tls\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.725918 kubelet[2827]: I0423 23:58:43.725852 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-config-path\") pod \"cilium-wsjjn\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " pod="kube-system/cilium-wsjjn" Apr 23 23:58:43.862590 systemd[1]: Created slice kubepods-besteffort-pod9d2a4eeb_2587_473e_b4d1_64cdde2220e0.slice - libcontainer container kubepods-besteffort-pod9d2a4eeb_2587_473e_b4d1_64cdde2220e0.slice. Apr 23 23:58:43.927866 kubelet[2827]: I0423 23:58:43.927707 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-cilium-config-path\") pod \"cilium-operator-78cf5644cb-96xmb\" (UID: \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\") " pod="kube-system/cilium-operator-78cf5644cb-96xmb" Apr 23 23:58:43.927866 kubelet[2827]: I0423 23:58:43.927807 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8p72\" (UniqueName: \"kubernetes.io/projected/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-kube-api-access-l8p72\") pod \"cilium-operator-78cf5644cb-96xmb\" (UID: \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\") " pod="kube-system/cilium-operator-78cf5644cb-96xmb" Apr 23 23:58:43.944919 containerd[1639]: time="2026-04-23T23:58:43.944870709Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5s9nh,Uid:a9157fb6-20d5-46ae-82e4-759c22b53154,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:43.951879 containerd[1639]: time="2026-04-23T23:58:43.951841478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wsjjn,Uid:4239b600-718c-4e02-9846-e3af0a0fc526,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:43.975059 containerd[1639]: time="2026-04-23T23:58:43.974962680Z" level=info msg="connecting to shim 5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017" address="unix:///run/containerd/s/c977360478f2965c56544a5f1e08e15dfa30ca8366f0744b510b0585e5b22906" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:43.982439 containerd[1639]: time="2026-04-23T23:58:43.982402873Z" level=info msg="connecting to shim ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:44.002658 systemd[1]: Started cri-containerd-5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017.scope - libcontainer container 5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017. Apr 23 23:58:44.007118 systemd[1]: Started cri-containerd-ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0.scope - libcontainer container ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0. Apr 23 23:58:44.044523 containerd[1639]: time="2026-04-23T23:58:44.044489248Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5s9nh,Uid:a9157fb6-20d5-46ae-82e4-759c22b53154,Namespace:kube-system,Attempt:0,} returns sandbox id \"5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017\"" Apr 23 23:58:44.050493 containerd[1639]: time="2026-04-23T23:58:44.050305455Z" level=info msg="CreateContainer within sandbox \"5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 23 23:58:44.051465 containerd[1639]: time="2026-04-23T23:58:44.051426227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wsjjn,Uid:4239b600-718c-4e02-9846-e3af0a0fc526,Namespace:kube-system,Attempt:0,} returns sandbox id \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\"" Apr 23 23:58:44.053800 containerd[1639]: time="2026-04-23T23:58:44.053689573Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Apr 23 23:58:44.064057 containerd[1639]: time="2026-04-23T23:58:44.064016758Z" level=info msg="Container d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:44.070990 containerd[1639]: time="2026-04-23T23:58:44.070959635Z" level=info msg="CreateContainer within sandbox \"5dd4fbfbb8030cd3af26787d3b023315aecf10020ec659d8a97b198c80157017\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038\"" Apr 23 23:58:44.071637 containerd[1639]: time="2026-04-23T23:58:44.071622550Z" level=info msg="StartContainer for \"d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038\"" Apr 23 23:58:44.072762 containerd[1639]: time="2026-04-23T23:58:44.072748599Z" level=info msg="connecting to shim d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038" address="unix:///run/containerd/s/c977360478f2965c56544a5f1e08e15dfa30ca8366f0744b510b0585e5b22906" protocol=ttrpc version=3 Apr 23 23:58:44.091650 systemd[1]: Started cri-containerd-d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038.scope - libcontainer container d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038. Apr 23 23:58:44.152064 containerd[1639]: time="2026-04-23T23:58:44.152015623Z" level=info msg="StartContainer for \"d850b89f99054ea531edd6ef4ca1e17a5a0a9b58cb07356b8f0a9e96a4a27038\" returns successfully" Apr 23 23:58:44.171610 containerd[1639]: time="2026-04-23T23:58:44.171525916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-78cf5644cb-96xmb,Uid:9d2a4eeb-2587-473e-b4d1-64cdde2220e0,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:44.190012 containerd[1639]: time="2026-04-23T23:58:44.189815829Z" level=info msg="connecting to shim 5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5" address="unix:///run/containerd/s/9716795a8638cddfa0813af5059ae18be96518f4d22fdaa7367609463768acf8" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:44.212699 systemd[1]: Started cri-containerd-5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5.scope - libcontainer container 5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5. Apr 23 23:58:44.256159 containerd[1639]: time="2026-04-23T23:58:44.256131570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-78cf5644cb-96xmb,Uid:9d2a4eeb-2587-473e-b4d1-64cdde2220e0,Namespace:kube-system,Attempt:0,} returns sandbox id \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\"" Apr 23 23:58:47.080178 kubelet[2827]: I0423 23:58:47.080057 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/kube-proxy-5s9nh" podStartSLOduration=4.080030645 podStartE2EDuration="4.080030645s" podCreationTimestamp="2026-04-23 23:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:44.671648218 +0000 UTC m=+6.162815642" watchObservedRunningTime="2026-04-23 23:58:47.080030645 +0000 UTC m=+8.571198079" Apr 23 23:58:47.964773 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount637018972.mount: Deactivated successfully. Apr 23 23:58:49.125141 containerd[1639]: time="2026-04-23T23:58:49.125085189Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:49.126174 containerd[1639]: time="2026-04-23T23:58:49.126087282Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Apr 23 23:58:49.126942 containerd[1639]: time="2026-04-23T23:58:49.126921394Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:49.128167 containerd[1639]: time="2026-04-23T23:58:49.127920903Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 5.074188467s" Apr 23 23:58:49.128167 containerd[1639]: time="2026-04-23T23:58:49.127944479Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Apr 23 23:58:49.129454 containerd[1639]: time="2026-04-23T23:58:49.129426002Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Apr 23 23:58:49.131672 containerd[1639]: time="2026-04-23T23:58:49.131655176Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 23 23:58:49.142033 containerd[1639]: time="2026-04-23T23:58:49.140739727Z" level=info msg="Container 227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:49.155605 containerd[1639]: time="2026-04-23T23:58:49.155578732Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\"" Apr 23 23:58:49.156969 containerd[1639]: time="2026-04-23T23:58:49.156939223Z" level=info msg="StartContainer for \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\"" Apr 23 23:58:49.157577 containerd[1639]: time="2026-04-23T23:58:49.157558472Z" level=info msg="connecting to shim 227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" protocol=ttrpc version=3 Apr 23 23:58:49.175641 systemd[1]: Started cri-containerd-227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9.scope - libcontainer container 227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9. Apr 23 23:58:49.200304 containerd[1639]: time="2026-04-23T23:58:49.200272250Z" level=info msg="StartContainer for \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" returns successfully" Apr 23 23:58:49.211642 containerd[1639]: time="2026-04-23T23:58:49.211491174Z" level=info msg="received container exit event container_id:\"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" id:\"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" pid:3250 exited_at:{seconds:1776988729 nanos:210333587}" Apr 23 23:58:49.211655 systemd[1]: cri-containerd-227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9.scope: Deactivated successfully. Apr 23 23:58:49.228179 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9-rootfs.mount: Deactivated successfully. Apr 23 23:58:49.675972 containerd[1639]: time="2026-04-23T23:58:49.675222614Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 23 23:58:49.688608 containerd[1639]: time="2026-04-23T23:58:49.688510782Z" level=info msg="Container fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:49.698605 containerd[1639]: time="2026-04-23T23:58:49.697676374Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\"" Apr 23 23:58:49.700607 containerd[1639]: time="2026-04-23T23:58:49.700527752Z" level=info msg="StartContainer for \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\"" Apr 23 23:58:49.704803 containerd[1639]: time="2026-04-23T23:58:49.704408685Z" level=info msg="connecting to shim fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" protocol=ttrpc version=3 Apr 23 23:58:49.734731 systemd[1]: Started cri-containerd-fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428.scope - libcontainer container fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428. Apr 23 23:58:49.781471 containerd[1639]: time="2026-04-23T23:58:49.781391391Z" level=info msg="StartContainer for \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" returns successfully" Apr 23 23:58:49.794226 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 23 23:58:49.794398 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:58:49.794707 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:58:49.799169 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 23 23:58:49.799384 systemd[1]: cri-containerd-fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428.scope: Deactivated successfully. Apr 23 23:58:49.800910 containerd[1639]: time="2026-04-23T23:58:49.800810578Z" level=info msg="received container exit event container_id:\"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" id:\"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" pid:3296 exited_at:{seconds:1776988729 nanos:800000342}" Apr 23 23:58:49.816190 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 23 23:58:50.690060 containerd[1639]: time="2026-04-23T23:58:50.690002266Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 23 23:58:50.716244 containerd[1639]: time="2026-04-23T23:58:50.714879152Z" level=info msg="Container 12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:50.722191 containerd[1639]: time="2026-04-23T23:58:50.722175039Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\"" Apr 23 23:58:50.722789 containerd[1639]: time="2026-04-23T23:58:50.722776221Z" level=info msg="StartContainer for \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\"" Apr 23 23:58:50.725103 containerd[1639]: time="2026-04-23T23:58:50.725086737Z" level=info msg="connecting to shim 12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" protocol=ttrpc version=3 Apr 23 23:58:50.744761 systemd[1]: Started cri-containerd-12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b.scope - libcontainer container 12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b. Apr 23 23:58:50.806034 containerd[1639]: time="2026-04-23T23:58:50.806004988Z" level=info msg="StartContainer for \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" returns successfully" Apr 23 23:58:50.809136 systemd[1]: cri-containerd-12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b.scope: Deactivated successfully. Apr 23 23:58:50.811499 containerd[1639]: time="2026-04-23T23:58:50.811277620Z" level=info msg="received container exit event container_id:\"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" id:\"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" pid:3341 exited_at:{seconds:1776988730 nanos:809309497}" Apr 23 23:58:51.139362 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b-rootfs.mount: Deactivated successfully. Apr 23 23:58:51.186424 containerd[1639]: time="2026-04-23T23:58:51.186367531Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:51.187523 containerd[1639]: time="2026-04-23T23:58:51.187345789Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Apr 23 23:58:51.188412 containerd[1639]: time="2026-04-23T23:58:51.188389024Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 23 23:58:51.189263 containerd[1639]: time="2026-04-23T23:58:51.189238729Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.059360658s" Apr 23 23:58:51.189325 containerd[1639]: time="2026-04-23T23:58:51.189312309Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Apr 23 23:58:51.192666 containerd[1639]: time="2026-04-23T23:58:51.192634133Z" level=info msg="CreateContainer within sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Apr 23 23:58:51.199352 containerd[1639]: time="2026-04-23T23:58:51.198987446Z" level=info msg="Container f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:51.209158 containerd[1639]: time="2026-04-23T23:58:51.209130925Z" level=info msg="CreateContainer within sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\"" Apr 23 23:58:51.209600 containerd[1639]: time="2026-04-23T23:58:51.209576263Z" level=info msg="StartContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\"" Apr 23 23:58:51.210466 containerd[1639]: time="2026-04-23T23:58:51.210413019Z" level=info msg="connecting to shim f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598" address="unix:///run/containerd/s/9716795a8638cddfa0813af5059ae18be96518f4d22fdaa7367609463768acf8" protocol=ttrpc version=3 Apr 23 23:58:51.229654 systemd[1]: Started cri-containerd-f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598.scope - libcontainer container f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598. Apr 23 23:58:51.254349 containerd[1639]: time="2026-04-23T23:58:51.254306077Z" level=info msg="StartContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" returns successfully" Apr 23 23:58:51.701954 containerd[1639]: time="2026-04-23T23:58:51.701899003Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 23 23:58:51.713808 containerd[1639]: time="2026-04-23T23:58:51.712672778Z" level=info msg="Container f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:51.718327 containerd[1639]: time="2026-04-23T23:58:51.718295785Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\"" Apr 23 23:58:51.721498 containerd[1639]: time="2026-04-23T23:58:51.721056747Z" level=info msg="StartContainer for \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\"" Apr 23 23:58:51.723129 containerd[1639]: time="2026-04-23T23:58:51.722715015Z" level=info msg="connecting to shim f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" protocol=ttrpc version=3 Apr 23 23:58:51.733112 kubelet[2827]: I0423 23:58:51.733070 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/cilium-operator-78cf5644cb-96xmb" podStartSLOduration=1.800774992 podStartE2EDuration="8.733059086s" podCreationTimestamp="2026-04-23 23:58:43 +0000 UTC" firstStartedPulling="2026-04-23 23:58:44.257592692 +0000 UTC m=+5.748760116" lastFinishedPulling="2026-04-23 23:58:51.189876776 +0000 UTC m=+12.681044210" observedRunningTime="2026-04-23 23:58:51.701823499 +0000 UTC m=+13.192990933" watchObservedRunningTime="2026-04-23 23:58:51.733059086 +0000 UTC m=+13.224226520" Apr 23 23:58:51.747842 systemd[1]: Started cri-containerd-f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d.scope - libcontainer container f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d. Apr 23 23:58:51.799736 systemd[1]: cri-containerd-f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d.scope: Deactivated successfully. Apr 23 23:58:51.804614 containerd[1639]: time="2026-04-23T23:58:51.803584714Z" level=info msg="StartContainer for \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" returns successfully" Apr 23 23:58:51.808823 containerd[1639]: time="2026-04-23T23:58:51.808766659Z" level=info msg="received container exit event container_id:\"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" id:\"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" pid:3431 exited_at:{seconds:1776988731 nanos:808637405}" Apr 23 23:58:52.616225 update_engine[1623]: I20260423 23:58:52.616122 1623 update_attempter.cc:509] Updating boot flags... Apr 23 23:58:52.733738 containerd[1639]: time="2026-04-23T23:58:52.733627435Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 23 23:58:52.759890 containerd[1639]: time="2026-04-23T23:58:52.759644932Z" level=info msg="Container 289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:52.769144 containerd[1639]: time="2026-04-23T23:58:52.769080459Z" level=info msg="CreateContainer within sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\"" Apr 23 23:58:52.770944 containerd[1639]: time="2026-04-23T23:58:52.770666960Z" level=info msg="StartContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\"" Apr 23 23:58:52.773673 containerd[1639]: time="2026-04-23T23:58:52.773648223Z" level=info msg="connecting to shim 289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52" address="unix:///run/containerd/s/925f3e8dd7b8faf68801fd6703cd033fb3a2690584610d126c1b72ae414f0626" protocol=ttrpc version=3 Apr 23 23:58:52.796681 systemd[1]: Started cri-containerd-289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52.scope - libcontainer container 289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52. Apr 23 23:58:52.891868 containerd[1639]: time="2026-04-23T23:58:52.891736234Z" level=info msg="StartContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" returns successfully" Apr 23 23:58:53.034906 kubelet[2827]: I0423 23:58:53.034871 2827 kubelet_node_status.go:427] "Fast updating node status as it just became ready" Apr 23 23:58:53.073462 systemd[1]: Created slice kubepods-burstable-pode848859f_83e3_40e8_86bc_d0e502363af6.slice - libcontainer container kubepods-burstable-pode848859f_83e3_40e8_86bc_d0e502363af6.slice. Apr 23 23:58:53.083013 systemd[1]: Created slice kubepods-burstable-pode9dc5ee0_a938_4f44_9e1c_0c2aa4032865.slice - libcontainer container kubepods-burstable-pode9dc5ee0_a938_4f44_9e1c_0c2aa4032865.slice. Apr 23 23:58:53.092417 kubelet[2827]: I0423 23:58:53.092362 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9dc5ee0-a938-4f44-9e1c-0c2aa4032865-config-volume\") pod \"coredns-7d764666f9-6m7gm\" (UID: \"e9dc5ee0-a938-4f44-9e1c-0c2aa4032865\") " pod="kube-system/coredns-7d764666f9-6m7gm" Apr 23 23:58:53.092417 kubelet[2827]: I0423 23:58:53.092413 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e848859f-83e3-40e8-86bc-d0e502363af6-config-volume\") pod \"coredns-7d764666f9-gpwsq\" (UID: \"e848859f-83e3-40e8-86bc-d0e502363af6\") " pod="kube-system/coredns-7d764666f9-gpwsq" Apr 23 23:58:53.092417 kubelet[2827]: I0423 23:58:53.092426 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkjdx\" (UniqueName: \"kubernetes.io/projected/e9dc5ee0-a938-4f44-9e1c-0c2aa4032865-kube-api-access-jkjdx\") pod \"coredns-7d764666f9-6m7gm\" (UID: \"e9dc5ee0-a938-4f44-9e1c-0c2aa4032865\") " pod="kube-system/coredns-7d764666f9-6m7gm" Apr 23 23:58:53.092610 kubelet[2827]: I0423 23:58:53.092439 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-667zp\" (UniqueName: \"kubernetes.io/projected/e848859f-83e3-40e8-86bc-d0e502363af6-kube-api-access-667zp\") pod \"coredns-7d764666f9-gpwsq\" (UID: \"e848859f-83e3-40e8-86bc-d0e502363af6\") " pod="kube-system/coredns-7d764666f9-gpwsq" Apr 23 23:58:53.380349 containerd[1639]: time="2026-04-23T23:58:53.379998491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-gpwsq,Uid:e848859f-83e3-40e8-86bc-d0e502363af6,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:53.390977 containerd[1639]: time="2026-04-23T23:58:53.390619877Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-6m7gm,Uid:e9dc5ee0-a938-4f44-9e1c-0c2aa4032865,Namespace:kube-system,Attempt:0,}" Apr 23 23:58:55.076989 systemd-networkd[1503]: cilium_host: Link UP Apr 23 23:58:55.077154 systemd-networkd[1503]: cilium_net: Link UP Apr 23 23:58:55.078691 systemd-networkd[1503]: cilium_net: Gained carrier Apr 23 23:58:55.079992 systemd-networkd[1503]: cilium_host: Gained carrier Apr 23 23:58:55.173866 systemd-networkd[1503]: cilium_vxlan: Link UP Apr 23 23:58:55.174454 systemd-networkd[1503]: cilium_vxlan: Gained carrier Apr 23 23:58:55.354603 kernel: NET: Registered PF_ALG protocol family Apr 23 23:58:55.558732 systemd-networkd[1503]: cilium_net: Gained IPv6LL Apr 23 23:58:55.782761 systemd-networkd[1503]: cilium_host: Gained IPv6LL Apr 23 23:58:55.895322 systemd-networkd[1503]: lxc_health: Link UP Apr 23 23:58:55.898569 systemd-networkd[1503]: lxc_health: Gained carrier Apr 23 23:58:55.964813 kubelet[2827]: I0423 23:58:55.964516 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/cilium-wsjjn" podStartSLOduration=4.29797765 podStartE2EDuration="12.964504781s" podCreationTimestamp="2026-04-23 23:58:43 +0000 UTC" firstStartedPulling="2026-04-23 23:58:44.053259808 +0000 UTC m=+5.544427242" lastFinishedPulling="2026-04-23 23:58:52.719786949 +0000 UTC m=+14.210954373" observedRunningTime="2026-04-23 23:58:53.748990663 +0000 UTC m=+15.240158127" watchObservedRunningTime="2026-04-23 23:58:55.964504781 +0000 UTC m=+17.455672215" Apr 23 23:58:56.433734 kernel: eth0: renamed from tmp90d91 Apr 23 23:58:56.438082 systemd-networkd[1503]: lxc3fb8e34c5eda: Link UP Apr 23 23:58:56.438299 systemd-networkd[1503]: lxc3fb8e34c5eda: Gained carrier Apr 23 23:58:56.453473 systemd-networkd[1503]: lxc7bdb9e4cc42f: Link UP Apr 23 23:58:56.462084 kernel: eth0: renamed from tmp99404 Apr 23 23:58:56.465017 systemd-networkd[1503]: lxc7bdb9e4cc42f: Gained carrier Apr 23 23:58:56.679003 systemd-networkd[1503]: cilium_vxlan: Gained IPv6LL Apr 23 23:58:57.831720 systemd-networkd[1503]: lxc_health: Gained IPv6LL Apr 23 23:58:58.024497 systemd-networkd[1503]: lxc3fb8e34c5eda: Gained IPv6LL Apr 23 23:58:58.278758 systemd-networkd[1503]: lxc7bdb9e4cc42f: Gained IPv6LL Apr 23 23:58:58.839556 containerd[1639]: time="2026-04-23T23:58:58.838775708Z" level=info msg="connecting to shim 9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5" address="unix:///run/containerd/s/3abd9e2d6d2efc0ab5d6555245b13794b163525e33d1e2c775a80a855b7fc95f" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:58.870015 containerd[1639]: time="2026-04-23T23:58:58.869648979Z" level=info msg="connecting to shim 90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd" address="unix:///run/containerd/s/2055e373cec5dd83696f3726c6a51f759e949207214c831e4515000529f9766e" namespace=k8s.io protocol=ttrpc version=3 Apr 23 23:58:58.874797 systemd[1]: Started cri-containerd-9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5.scope - libcontainer container 9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5. Apr 23 23:58:58.897696 systemd[1]: Started cri-containerd-90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd.scope - libcontainer container 90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd. Apr 23 23:58:58.944277 containerd[1639]: time="2026-04-23T23:58:58.944242201Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-6m7gm,Uid:e9dc5ee0-a938-4f44-9e1c-0c2aa4032865,Namespace:kube-system,Attempt:0,} returns sandbox id \"9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5\"" Apr 23 23:58:58.953616 containerd[1639]: time="2026-04-23T23:58:58.952420924Z" level=info msg="CreateContainer within sandbox \"9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 23 23:58:58.962766 containerd[1639]: time="2026-04-23T23:58:58.962746705Z" level=info msg="Container 27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:58.969898 containerd[1639]: time="2026-04-23T23:58:58.969866442Z" level=info msg="CreateContainer within sandbox \"9940433825c122fbc4c4c7db5402ed04be26713cdfd7a877f8cbd77337c4c7e5\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9\"" Apr 23 23:58:58.970888 containerd[1639]: time="2026-04-23T23:58:58.970870644Z" level=info msg="StartContainer for \"27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9\"" Apr 23 23:58:58.973958 containerd[1639]: time="2026-04-23T23:58:58.973763680Z" level=info msg="connecting to shim 27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9" address="unix:///run/containerd/s/3abd9e2d6d2efc0ab5d6555245b13794b163525e33d1e2c775a80a855b7fc95f" protocol=ttrpc version=3 Apr 23 23:58:58.984703 containerd[1639]: time="2026-04-23T23:58:58.984684847Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7d764666f9-gpwsq,Uid:e848859f-83e3-40e8-86bc-d0e502363af6,Namespace:kube-system,Attempt:0,} returns sandbox id \"90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd\"" Apr 23 23:58:58.991856 containerd[1639]: time="2026-04-23T23:58:58.991825267Z" level=info msg="CreateContainer within sandbox \"90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 23 23:58:58.997687 systemd[1]: Started cri-containerd-27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9.scope - libcontainer container 27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9. Apr 23 23:58:58.999155 containerd[1639]: time="2026-04-23T23:58:58.999130852Z" level=info msg="Container f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25: CDI devices from CRI Config.CDIDevices: []" Apr 23 23:58:59.005437 containerd[1639]: time="2026-04-23T23:58:59.005410869Z" level=info msg="CreateContainer within sandbox \"90d9105d8d11b7561e32ff1018f3fec2b767467b15cc4c5cd168b9c70481dccd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25\"" Apr 23 23:58:59.006846 containerd[1639]: time="2026-04-23T23:58:59.006815347Z" level=info msg="StartContainer for \"f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25\"" Apr 23 23:58:59.007599 containerd[1639]: time="2026-04-23T23:58:59.007528733Z" level=info msg="connecting to shim f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25" address="unix:///run/containerd/s/2055e373cec5dd83696f3726c6a51f759e949207214c831e4515000529f9766e" protocol=ttrpc version=3 Apr 23 23:58:59.026699 systemd[1]: Started cri-containerd-f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25.scope - libcontainer container f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25. Apr 23 23:58:59.036826 containerd[1639]: time="2026-04-23T23:58:59.036753133Z" level=info msg="StartContainer for \"27c0789a5b1a3ab124109b46e6cdba2fdaa42acf903d06ecb2b0334379aa81f9\" returns successfully" Apr 23 23:58:59.068172 containerd[1639]: time="2026-04-23T23:58:59.068042935Z" level=info msg="StartContainer for \"f4e98578fe82c58047e66aaec60d5ff20342879ff2885743f84d6433ed172e25\" returns successfully" Apr 23 23:58:59.786306 kubelet[2827]: I0423 23:58:59.786111 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-gpwsq" podStartSLOduration=16.786091004 podStartE2EDuration="16.786091004s" podCreationTimestamp="2026-04-23 23:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:59.766980512 +0000 UTC m=+21.258147976" watchObservedRunningTime="2026-04-23 23:58:59.786091004 +0000 UTC m=+21.277258468" Apr 23 23:59:05.511105 kubelet[2827]: I0423 23:59:05.510734 2827 prober_manager.go:356] "Failed to trigger a manual run" probe="Readiness" Apr 23 23:59:05.531282 kubelet[2827]: I0423 23:59:05.531179 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/coredns-7d764666f9-6m7gm" podStartSLOduration=22.531161827 podStartE2EDuration="22.531161827s" podCreationTimestamp="2026-04-23 23:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-23 23:58:59.810863768 +0000 UTC m=+21.302031222" watchObservedRunningTime="2026-04-23 23:59:05.531161827 +0000 UTC m=+27.022329291" Apr 24 00:01:00.273762 systemd[1]: Started sshd@9-135.181.109.116:22-20.229.252.112:50256.service - OpenSSH per-connection server daemon (20.229.252.112:50256). Apr 24 00:01:00.488592 sshd[4172]: Accepted publickey for core from 20.229.252.112 port 50256 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:00.490461 sshd-session[4172]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:00.500185 systemd-logind[1616]: New session 10 of user core. Apr 24 00:01:00.508749 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 24 00:01:00.700047 sshd[4175]: Connection closed by 20.229.252.112 port 50256 Apr 24 00:01:00.700691 sshd-session[4172]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:00.704852 systemd[1]: sshd@9-135.181.109.116:22-20.229.252.112:50256.service: Deactivated successfully. Apr 24 00:01:00.706870 systemd[1]: session-10.scope: Deactivated successfully. Apr 24 00:01:00.708104 systemd-logind[1616]: Session 10 logged out. Waiting for processes to exit. Apr 24 00:01:00.710188 systemd-logind[1616]: Removed session 10. Apr 24 00:01:05.745264 systemd[1]: Started sshd@10-135.181.109.116:22-20.229.252.112:50268.service - OpenSSH per-connection server daemon (20.229.252.112:50268). Apr 24 00:01:05.957000 sshd[4188]: Accepted publickey for core from 20.229.252.112 port 50268 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:05.959689 sshd-session[4188]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:05.969420 systemd-logind[1616]: New session 11 of user core. Apr 24 00:01:05.975786 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 24 00:01:06.132970 sshd[4191]: Connection closed by 20.229.252.112 port 50268 Apr 24 00:01:06.134858 sshd-session[4188]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:06.140598 systemd[1]: sshd@10-135.181.109.116:22-20.229.252.112:50268.service: Deactivated successfully. Apr 24 00:01:06.145005 systemd[1]: session-11.scope: Deactivated successfully. Apr 24 00:01:06.146466 systemd-logind[1616]: Session 11 logged out. Waiting for processes to exit. Apr 24 00:01:06.148377 systemd-logind[1616]: Removed session 11. Apr 24 00:01:11.181873 systemd[1]: Started sshd@11-135.181.109.116:22-20.229.252.112:59174.service - OpenSSH per-connection server daemon (20.229.252.112:59174). Apr 24 00:01:11.396798 sshd[4204]: Accepted publickey for core from 20.229.252.112 port 59174 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:11.399448 sshd-session[4204]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:11.407458 systemd-logind[1616]: New session 12 of user core. Apr 24 00:01:11.414806 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 24 00:01:11.595429 sshd[4207]: Connection closed by 20.229.252.112 port 59174 Apr 24 00:01:11.597887 sshd-session[4204]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:11.604131 systemd[1]: sshd@11-135.181.109.116:22-20.229.252.112:59174.service: Deactivated successfully. Apr 24 00:01:11.608062 systemd[1]: session-12.scope: Deactivated successfully. Apr 24 00:01:11.612777 systemd-logind[1616]: Session 12 logged out. Waiting for processes to exit. Apr 24 00:01:11.614634 systemd-logind[1616]: Removed session 12. Apr 24 00:01:16.634609 systemd[1]: Started sshd@12-135.181.109.116:22-20.229.252.112:43212.service - OpenSSH per-connection server daemon (20.229.252.112:43212). Apr 24 00:01:16.839391 sshd[4222]: Accepted publickey for core from 20.229.252.112 port 43212 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:16.841940 sshd-session[4222]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:16.851579 systemd-logind[1616]: New session 13 of user core. Apr 24 00:01:16.853762 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 24 00:01:17.033671 sshd[4225]: Connection closed by 20.229.252.112 port 43212 Apr 24 00:01:17.035836 sshd-session[4222]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:17.042845 systemd-logind[1616]: Session 13 logged out. Waiting for processes to exit. Apr 24 00:01:17.044204 systemd[1]: sshd@12-135.181.109.116:22-20.229.252.112:43212.service: Deactivated successfully. Apr 24 00:01:17.048100 systemd[1]: session-13.scope: Deactivated successfully. Apr 24 00:01:17.052011 systemd-logind[1616]: Removed session 13. Apr 24 00:01:17.074878 systemd[1]: Started sshd@13-135.181.109.116:22-20.229.252.112:43222.service - OpenSSH per-connection server daemon (20.229.252.112:43222). Apr 24 00:01:17.268449 sshd[4237]: Accepted publickey for core from 20.229.252.112 port 43222 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:17.271251 sshd-session[4237]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:17.281523 systemd-logind[1616]: New session 14 of user core. Apr 24 00:01:17.290826 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 24 00:01:17.465703 sshd[4240]: Connection closed by 20.229.252.112 port 43222 Apr 24 00:01:17.466388 sshd-session[4237]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:17.470169 systemd-logind[1616]: Session 14 logged out. Waiting for processes to exit. Apr 24 00:01:17.471007 systemd[1]: sshd@13-135.181.109.116:22-20.229.252.112:43222.service: Deactivated successfully. Apr 24 00:01:17.473033 systemd[1]: session-14.scope: Deactivated successfully. Apr 24 00:01:17.475250 systemd-logind[1616]: Removed session 14. Apr 24 00:01:17.505473 systemd[1]: Started sshd@14-135.181.109.116:22-20.229.252.112:43238.service - OpenSSH per-connection server daemon (20.229.252.112:43238). Apr 24 00:01:17.698937 sshd[4250]: Accepted publickey for core from 20.229.252.112 port 43238 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:17.701924 sshd-session[4250]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:17.710652 systemd-logind[1616]: New session 15 of user core. Apr 24 00:01:17.720975 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 24 00:01:17.897382 sshd[4253]: Connection closed by 20.229.252.112 port 43238 Apr 24 00:01:17.898418 sshd-session[4250]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:17.905560 systemd[1]: sshd@14-135.181.109.116:22-20.229.252.112:43238.service: Deactivated successfully. Apr 24 00:01:17.905674 systemd-logind[1616]: Session 15 logged out. Waiting for processes to exit. Apr 24 00:01:17.908491 systemd[1]: session-15.scope: Deactivated successfully. Apr 24 00:01:17.911402 systemd-logind[1616]: Removed session 15. Apr 24 00:01:22.942638 systemd[1]: Started sshd@15-135.181.109.116:22-20.229.252.112:43248.service - OpenSSH per-connection server daemon (20.229.252.112:43248). Apr 24 00:01:23.158331 sshd[4265]: Accepted publickey for core from 20.229.252.112 port 43248 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:23.160894 sshd-session[4265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:23.171500 systemd-logind[1616]: New session 16 of user core. Apr 24 00:01:23.173823 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 24 00:01:23.338229 sshd[4268]: Connection closed by 20.229.252.112 port 43248 Apr 24 00:01:23.338943 sshd-session[4265]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:23.346451 systemd-logind[1616]: Session 16 logged out. Waiting for processes to exit. Apr 24 00:01:23.347344 systemd[1]: sshd@15-135.181.109.116:22-20.229.252.112:43248.service: Deactivated successfully. Apr 24 00:01:23.351079 systemd[1]: session-16.scope: Deactivated successfully. Apr 24 00:01:23.353834 systemd-logind[1616]: Removed session 16. Apr 24 00:01:28.384940 systemd[1]: Started sshd@16-135.181.109.116:22-20.229.252.112:35994.service - OpenSSH per-connection server daemon (20.229.252.112:35994). Apr 24 00:01:28.606604 sshd[4280]: Accepted publickey for core from 20.229.252.112 port 35994 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:28.609133 sshd-session[4280]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:28.619093 systemd-logind[1616]: New session 17 of user core. Apr 24 00:01:28.623842 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 24 00:01:28.780454 sshd[4283]: Connection closed by 20.229.252.112 port 35994 Apr 24 00:01:28.781723 sshd-session[4280]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:28.787256 systemd[1]: sshd@16-135.181.109.116:22-20.229.252.112:35994.service: Deactivated successfully. Apr 24 00:01:28.791021 systemd[1]: session-17.scope: Deactivated successfully. Apr 24 00:01:28.792252 systemd-logind[1616]: Session 17 logged out. Waiting for processes to exit. Apr 24 00:01:28.794682 systemd-logind[1616]: Removed session 17. Apr 24 00:01:33.827011 systemd[1]: Started sshd@17-135.181.109.116:22-20.229.252.112:36008.service - OpenSSH per-connection server daemon (20.229.252.112:36008). Apr 24 00:01:34.039941 sshd[4294]: Accepted publickey for core from 20.229.252.112 port 36008 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:34.042599 sshd-session[4294]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:34.051697 systemd-logind[1616]: New session 18 of user core. Apr 24 00:01:34.058770 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 24 00:01:34.244302 sshd[4297]: Connection closed by 20.229.252.112 port 36008 Apr 24 00:01:34.245924 sshd-session[4294]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:34.252665 systemd[1]: sshd@17-135.181.109.116:22-20.229.252.112:36008.service: Deactivated successfully. Apr 24 00:01:34.257031 systemd[1]: session-18.scope: Deactivated successfully. Apr 24 00:01:34.261175 systemd-logind[1616]: Session 18 logged out. Waiting for processes to exit. Apr 24 00:01:34.263358 systemd-logind[1616]: Removed session 18. Apr 24 00:01:39.296682 systemd[1]: Started sshd@18-135.181.109.116:22-20.229.252.112:35536.service - OpenSSH per-connection server daemon (20.229.252.112:35536). Apr 24 00:01:39.510604 sshd[4311]: Accepted publickey for core from 20.229.252.112 port 35536 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:39.512318 sshd-session[4311]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:39.521903 systemd-logind[1616]: New session 19 of user core. Apr 24 00:01:39.527799 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 24 00:01:39.683955 sshd[4314]: Connection closed by 20.229.252.112 port 35536 Apr 24 00:01:39.685827 sshd-session[4311]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:39.689956 systemd[1]: sshd@18-135.181.109.116:22-20.229.252.112:35536.service: Deactivated successfully. Apr 24 00:01:39.693380 systemd[1]: session-19.scope: Deactivated successfully. Apr 24 00:01:39.695250 systemd-logind[1616]: Session 19 logged out. Waiting for processes to exit. Apr 24 00:01:39.696876 systemd-logind[1616]: Removed session 19. Apr 24 00:01:39.727846 systemd[1]: Started sshd@19-135.181.109.116:22-20.229.252.112:35540.service - OpenSSH per-connection server daemon (20.229.252.112:35540). Apr 24 00:01:39.950346 sshd[4326]: Accepted publickey for core from 20.229.252.112 port 35540 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:39.952529 sshd-session[4326]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:39.960998 systemd-logind[1616]: New session 20 of user core. Apr 24 00:01:39.967872 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 24 00:01:40.180239 sshd[4329]: Connection closed by 20.229.252.112 port 35540 Apr 24 00:01:40.181868 sshd-session[4326]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:40.188484 systemd[1]: sshd@19-135.181.109.116:22-20.229.252.112:35540.service: Deactivated successfully. Apr 24 00:01:40.192630 systemd[1]: session-20.scope: Deactivated successfully. Apr 24 00:01:40.194989 systemd-logind[1616]: Session 20 logged out. Waiting for processes to exit. Apr 24 00:01:40.197677 systemd-logind[1616]: Removed session 20. Apr 24 00:01:40.223472 systemd[1]: Started sshd@20-135.181.109.116:22-20.229.252.112:35546.service - OpenSSH per-connection server daemon (20.229.252.112:35546). Apr 24 00:01:40.417020 sshd[4339]: Accepted publickey for core from 20.229.252.112 port 35546 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:40.418801 sshd-session[4339]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:40.425473 systemd-logind[1616]: New session 21 of user core. Apr 24 00:01:40.433833 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 24 00:01:41.120672 sshd[4342]: Connection closed by 20.229.252.112 port 35546 Apr 24 00:01:41.121780 sshd-session[4339]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:41.128718 systemd[1]: sshd@20-135.181.109.116:22-20.229.252.112:35546.service: Deactivated successfully. Apr 24 00:01:41.132228 systemd[1]: session-21.scope: Deactivated successfully. Apr 24 00:01:41.134956 systemd-logind[1616]: Session 21 logged out. Waiting for processes to exit. Apr 24 00:01:41.139192 systemd-logind[1616]: Removed session 21. Apr 24 00:01:41.162720 systemd[1]: Started sshd@21-135.181.109.116:22-20.229.252.112:35552.service - OpenSSH per-connection server daemon (20.229.252.112:35552). Apr 24 00:01:41.353594 sshd[4357]: Accepted publickey for core from 20.229.252.112 port 35552 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:41.355258 sshd-session[4357]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:41.363655 systemd-logind[1616]: New session 22 of user core. Apr 24 00:01:41.374815 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 24 00:01:41.636321 sshd[4360]: Connection closed by 20.229.252.112 port 35552 Apr 24 00:01:41.637688 sshd-session[4357]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:41.641777 systemd-logind[1616]: Session 22 logged out. Waiting for processes to exit. Apr 24 00:01:41.642202 systemd[1]: sshd@21-135.181.109.116:22-20.229.252.112:35552.service: Deactivated successfully. Apr 24 00:01:41.644604 systemd[1]: session-22.scope: Deactivated successfully. Apr 24 00:01:41.646419 systemd-logind[1616]: Removed session 22. Apr 24 00:01:41.676374 systemd[1]: Started sshd@22-135.181.109.116:22-20.229.252.112:35568.service - OpenSSH per-connection server daemon (20.229.252.112:35568). Apr 24 00:01:41.863409 sshd[4370]: Accepted publickey for core from 20.229.252.112 port 35568 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:41.866442 sshd-session[4370]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:41.874562 systemd-logind[1616]: New session 23 of user core. Apr 24 00:01:41.883852 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 24 00:01:42.035988 sshd[4373]: Connection closed by 20.229.252.112 port 35568 Apr 24 00:01:42.037553 sshd-session[4370]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:42.044400 systemd-logind[1616]: Session 23 logged out. Waiting for processes to exit. Apr 24 00:01:42.045713 systemd[1]: sshd@22-135.181.109.116:22-20.229.252.112:35568.service: Deactivated successfully. Apr 24 00:01:42.048739 systemd[1]: session-23.scope: Deactivated successfully. Apr 24 00:01:42.051273 systemd-logind[1616]: Removed session 23. Apr 24 00:01:47.082719 systemd[1]: Started sshd@23-135.181.109.116:22-20.229.252.112:48760.service - OpenSSH per-connection server daemon (20.229.252.112:48760). Apr 24 00:01:47.308592 sshd[4391]: Accepted publickey for core from 20.229.252.112 port 48760 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:47.310528 sshd-session[4391]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:47.319779 systemd-logind[1616]: New session 24 of user core. Apr 24 00:01:47.327820 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 24 00:01:47.483805 sshd[4394]: Connection closed by 20.229.252.112 port 48760 Apr 24 00:01:47.484756 sshd-session[4391]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:47.488618 systemd-logind[1616]: Session 24 logged out. Waiting for processes to exit. Apr 24 00:01:47.489052 systemd[1]: sshd@23-135.181.109.116:22-20.229.252.112:48760.service: Deactivated successfully. Apr 24 00:01:47.490827 systemd[1]: session-24.scope: Deactivated successfully. Apr 24 00:01:47.492471 systemd-logind[1616]: Removed session 24. Apr 24 00:01:52.528895 systemd[1]: Started sshd@24-135.181.109.116:22-20.229.252.112:48762.service - OpenSSH per-connection server daemon (20.229.252.112:48762). Apr 24 00:01:52.750166 sshd[4406]: Accepted publickey for core from 20.229.252.112 port 48762 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:52.752782 sshd-session[4406]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:52.761955 systemd-logind[1616]: New session 25 of user core. Apr 24 00:01:52.766840 systemd[1]: Started session-25.scope - Session 25 of User core. Apr 24 00:01:52.922291 sshd[4409]: Connection closed by 20.229.252.112 port 48762 Apr 24 00:01:52.922829 sshd-session[4406]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:52.928498 systemd[1]: sshd@24-135.181.109.116:22-20.229.252.112:48762.service: Deactivated successfully. Apr 24 00:01:52.933495 systemd[1]: session-25.scope: Deactivated successfully. Apr 24 00:01:52.939593 systemd-logind[1616]: Session 25 logged out. Waiting for processes to exit. Apr 24 00:01:52.941670 systemd-logind[1616]: Removed session 25. Apr 24 00:01:52.968803 systemd[1]: Started sshd@25-135.181.109.116:22-20.229.252.112:48772.service - OpenSSH per-connection server daemon (20.229.252.112:48772). Apr 24 00:01:53.172767 sshd[4420]: Accepted publickey for core from 20.229.252.112 port 48772 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:53.175744 sshd-session[4420]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:53.184267 systemd-logind[1616]: New session 26 of user core. Apr 24 00:01:53.192781 systemd[1]: Started session-26.scope - Session 26 of User core. Apr 24 00:01:54.563631 containerd[1639]: time="2026-04-24T00:01:54.563211679Z" level=info msg="StopContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" with timeout 30 (s)" Apr 24 00:01:54.565791 containerd[1639]: time="2026-04-24T00:01:54.565766682Z" level=info msg="Stop container \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" with signal terminated" Apr 24 00:01:54.579778 containerd[1639]: time="2026-04-24T00:01:54.579708689Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 24 00:01:54.591718 systemd[1]: cri-containerd-f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598.scope: Deactivated successfully. Apr 24 00:01:54.592459 containerd[1639]: time="2026-04-24T00:01:54.591402065Z" level=info msg="StopContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" with timeout 2 (s)" Apr 24 00:01:54.592661 containerd[1639]: time="2026-04-24T00:01:54.592632776Z" level=info msg="Stop container \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" with signal terminated" Apr 24 00:01:54.596084 containerd[1639]: time="2026-04-24T00:01:54.595843856Z" level=info msg="received container exit event container_id:\"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" id:\"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" pid:3397 exited_at:{seconds:1776988914 nanos:594951693}" Apr 24 00:01:54.605666 systemd-networkd[1503]: lxc_health: Link DOWN Apr 24 00:01:54.605673 systemd-networkd[1503]: lxc_health: Lost carrier Apr 24 00:01:54.622098 systemd[1]: cri-containerd-289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52.scope: Deactivated successfully. Apr 24 00:01:54.622384 systemd[1]: cri-containerd-289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52.scope: Consumed 5.183s CPU time, 125.5M memory peak, 112K read from disk, 13.3M written to disk. Apr 24 00:01:54.626095 containerd[1639]: time="2026-04-24T00:01:54.626060119Z" level=info msg="received container exit event container_id:\"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" id:\"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" pid:3487 exited_at:{seconds:1776988914 nanos:625507477}" Apr 24 00:01:54.646647 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598-rootfs.mount: Deactivated successfully. Apr 24 00:01:54.657710 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52-rootfs.mount: Deactivated successfully. Apr 24 00:01:54.668442 containerd[1639]: time="2026-04-24T00:01:54.668369769Z" level=info msg="StopContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" returns successfully" Apr 24 00:01:54.669019 containerd[1639]: time="2026-04-24T00:01:54.669005005Z" level=info msg="StopPodSandbox for \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\"" Apr 24 00:01:54.669172 containerd[1639]: time="2026-04-24T00:01:54.669111414Z" level=info msg="Container to stop \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.669172 containerd[1639]: time="2026-04-24T00:01:54.669123212Z" level=info msg="Container to stop \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.669172 containerd[1639]: time="2026-04-24T00:01:54.669129932Z" level=info msg="Container to stop \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.669172 containerd[1639]: time="2026-04-24T00:01:54.669136111Z" level=info msg="Container to stop \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.669262 containerd[1639]: time="2026-04-24T00:01:54.669253438Z" level=info msg="Container to stop \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.670248 containerd[1639]: time="2026-04-24T00:01:54.670226752Z" level=info msg="StopContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" returns successfully" Apr 24 00:01:54.670653 containerd[1639]: time="2026-04-24T00:01:54.670529797Z" level=info msg="StopPodSandbox for \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\"" Apr 24 00:01:54.670725 containerd[1639]: time="2026-04-24T00:01:54.670716368Z" level=info msg="Container to stop \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 24 00:01:54.676083 systemd[1]: cri-containerd-ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0.scope: Deactivated successfully. Apr 24 00:01:54.680077 systemd[1]: cri-containerd-5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5.scope: Deactivated successfully. Apr 24 00:01:54.682436 containerd[1639]: time="2026-04-24T00:01:54.682360540Z" level=info msg="received sandbox exit event container_id:\"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" id:\"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" exit_status:137 exited_at:{seconds:1776988914 nanos:682134630}" monitor_name=podsandbox Apr 24 00:01:54.685914 containerd[1639]: time="2026-04-24T00:01:54.685863007Z" level=info msg="received sandbox exit event container_id:\"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" id:\"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" exit_status:137 exited_at:{seconds:1776988914 nanos:685708085}" monitor_name=podsandbox Apr 24 00:01:54.702345 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0-rootfs.mount: Deactivated successfully. Apr 24 00:01:54.709273 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5-rootfs.mount: Deactivated successfully. Apr 24 00:01:54.710979 containerd[1639]: time="2026-04-24T00:01:54.710950007Z" level=info msg="shim disconnected" id=ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0 namespace=k8s.io Apr 24 00:01:54.710979 containerd[1639]: time="2026-04-24T00:01:54.710978500Z" level=warning msg="cleaning up after shim disconnected" id=ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0 namespace=k8s.io Apr 24 00:01:54.711497 containerd[1639]: time="2026-04-24T00:01:54.710985711Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 00:01:54.714759 containerd[1639]: time="2026-04-24T00:01:54.714611062Z" level=info msg="shim disconnected" id=5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5 namespace=k8s.io Apr 24 00:01:54.714759 containerd[1639]: time="2026-04-24T00:01:54.714626876Z" level=warning msg="cleaning up after shim disconnected" id=5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5 namespace=k8s.io Apr 24 00:01:54.714759 containerd[1639]: time="2026-04-24T00:01:54.714633827Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 24 00:01:54.726686 containerd[1639]: time="2026-04-24T00:01:54.726636398Z" level=info msg="received sandbox container exit event sandbox_id:\"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" exit_status:137 exited_at:{seconds:1776988914 nanos:682134630}" monitor_name=criService Apr 24 00:01:54.727279 containerd[1639]: time="2026-04-24T00:01:54.727262299Z" level=info msg="received sandbox container exit event sandbox_id:\"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" exit_status:137 exited_at:{seconds:1776988914 nanos:685708085}" monitor_name=criService Apr 24 00:01:54.729943 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5-shm.mount: Deactivated successfully. Apr 24 00:01:54.730026 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0-shm.mount: Deactivated successfully. Apr 24 00:01:54.730968 containerd[1639]: time="2026-04-24T00:01:54.730951838Z" level=info msg="TearDown network for sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" successfully" Apr 24 00:01:54.731004 containerd[1639]: time="2026-04-24T00:01:54.730970566Z" level=info msg="StopPodSandbox for \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" returns successfully" Apr 24 00:01:54.731332 containerd[1639]: time="2026-04-24T00:01:54.731318590Z" level=info msg="TearDown network for sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" successfully" Apr 24 00:01:54.731445 containerd[1639]: time="2026-04-24T00:01:54.731419692Z" level=info msg="StopPodSandbox for \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" returns successfully" Apr 24 00:01:54.909664 kubelet[2827]: I0424 00:01:54.909311 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-kube-api-access-jj57b\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-kube-api-access-jj57b\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.909664 kubelet[2827]: I0424 00:01:54.909359 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-hubble-tls\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-hubble-tls\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.909664 kubelet[2827]: I0424 00:01:54.909386 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/projected/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-kube-api-access-l8p72\" (UniqueName: \"kubernetes.io/projected/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-kube-api-access-l8p72\") pod \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\" (UID: \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\") " Apr 24 00:01:54.909664 kubelet[2827]: I0424 00:01:54.909412 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-cgroup\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.909664 kubelet[2827]: I0424 00:01:54.909447 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-kernel\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.910500 kubelet[2827]: I0424 00:01:54.909468 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-hostproc\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-hostproc\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.910500 kubelet[2827]: I0424 00:01:54.909497 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/secret/4239b600-718c-4e02-9846-e3af0a0fc526-clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4239b600-718c-4e02-9846-e3af0a0fc526-clustermesh-secrets\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.910500 kubelet[2827]: I0424 00:01:54.909520 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-etc-cni-netd\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912188 kubelet[2827]: I0424 00:01:54.910629 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-cgroup" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.912188 kubelet[2827]: I0424 00:01:54.911285 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-run\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-run\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912188 kubelet[2827]: I0424 00:01:54.911349 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-config-path\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912188 kubelet[2827]: I0424 00:01:54.911374 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-net\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912188 kubelet[2827]: I0424 00:01:54.911404 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/configmap/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-cilium-config-path\") pod \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\" (UID: \"9d2a4eeb-2587-473e-b4d1-64cdde2220e0\") " Apr 24 00:01:54.912500 kubelet[2827]: I0424 00:01:54.911452 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-lib-modules\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-lib-modules\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912500 kubelet[2827]: I0424 00:01:54.911474 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-bpf-maps\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912500 kubelet[2827]: I0424 00:01:54.911500 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cni-path\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cni-path\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912500 kubelet[2827]: I0424 00:01:54.911523 2827 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-xtables-lock\") pod \"4239b600-718c-4e02-9846-e3af0a0fc526\" (UID: \"4239b600-718c-4e02-9846-e3af0a0fc526\") " Apr 24 00:01:54.912500 kubelet[2827]: I0424 00:01:54.911608 2827 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-cgroup\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:54.912739 kubelet[2827]: I0424 00:01:54.911641 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-xtables-lock" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.912739 kubelet[2827]: I0424 00:01:54.911670 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-kernel" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.912739 kubelet[2827]: I0424 00:01:54.911694 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-hostproc" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.913528 kubelet[2827]: I0424 00:01:54.913470 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-etc-cni-netd" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.913851 kubelet[2827]: I0424 00:01:54.913648 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-run" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.913953 kubelet[2827]: I0424 00:01:54.913932 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-net" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.916104 kubelet[2827]: I0424 00:01:54.916025 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-lib-modules" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.916324 kubelet[2827]: I0424 00:01:54.916281 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-bpf-maps" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.918737 kubelet[2827]: I0424 00:01:54.916657 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cni-path" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Apr 24 00:01:54.922707 kubelet[2827]: I0424 00:01:54.922671 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-kube-api-access-jj57b" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "kube-api-access-jj57b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 24 00:01:54.929709 kubelet[2827]: I0424 00:01:54.929632 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4239b600-718c-4e02-9846-e3af0a0fc526-clustermesh-secrets" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 24 00:01:54.930137 kubelet[2827]: I0424 00:01:54.930115 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-hubble-tls" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 24 00:01:54.930245 kubelet[2827]: I0424 00:01:54.930170 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-cilium-config-path" pod "9d2a4eeb-2587-473e-b4d1-64cdde2220e0" (UID: "9d2a4eeb-2587-473e-b4d1-64cdde2220e0"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 24 00:01:54.930860 kubelet[2827]: I0424 00:01:54.930688 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-config-path" pod "4239b600-718c-4e02-9846-e3af0a0fc526" (UID: "4239b600-718c-4e02-9846-e3af0a0fc526"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 24 00:01:54.931041 kubelet[2827]: I0424 00:01:54.931002 2827 operation_generator.go:779] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-kube-api-access-l8p72" pod "9d2a4eeb-2587-473e-b4d1-64cdde2220e0" (UID: "9d2a4eeb-2587-473e-b4d1-64cdde2220e0"). InnerVolumeSpecName "kube-api-access-l8p72". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 24 00:01:55.012197 kubelet[2827]: I0424 00:01:55.012139 2827 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-net\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012197 kubelet[2827]: I0424 00:01:55.012184 2827 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-cilium-config-path\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012197 kubelet[2827]: I0424 00:01:55.012200 2827 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-lib-modules\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012197 kubelet[2827]: I0424 00:01:55.012215 2827 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-bpf-maps\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012197 kubelet[2827]: I0424 00:01:55.012233 2827 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cni-path\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012247 2827 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-xtables-lock\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012260 2827 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jj57b\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-kube-api-access-jj57b\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012274 2827 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/4239b600-718c-4e02-9846-e3af0a0fc526-hubble-tls\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012287 2827 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l8p72\" (UniqueName: \"kubernetes.io/projected/9d2a4eeb-2587-473e-b4d1-64cdde2220e0-kube-api-access-l8p72\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012300 2827 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-host-proc-sys-kernel\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012314 2827 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-hostproc\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012327 2827 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4239b600-718c-4e02-9846-e3af0a0fc526-clustermesh-secrets\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012597 kubelet[2827]: I0424 00:01:55.012342 2827 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-etc-cni-netd\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012901 kubelet[2827]: I0424 00:01:55.012355 2827 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-run\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.012901 kubelet[2827]: I0424 00:01:55.012368 2827 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4239b600-718c-4e02-9846-e3af0a0fc526-cilium-config-path\") on node \"ci-4459-2-4-n-09e3934cf7\" DevicePath \"\"" Apr 24 00:01:55.207693 kubelet[2827]: I0424 00:01:55.207631 2827 scope.go:122] "RemoveContainer" containerID="f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598" Apr 24 00:01:55.214392 containerd[1639]: time="2026-04-24T00:01:55.212946671Z" level=info msg="RemoveContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\"" Apr 24 00:01:55.221200 systemd[1]: Removed slice kubepods-besteffort-pod9d2a4eeb_2587_473e_b4d1_64cdde2220e0.slice - libcontainer container kubepods-besteffort-pod9d2a4eeb_2587_473e_b4d1_64cdde2220e0.slice. Apr 24 00:01:55.224227 containerd[1639]: time="2026-04-24T00:01:55.224161256Z" level=info msg="RemoveContainer for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" returns successfully" Apr 24 00:01:55.224878 kubelet[2827]: I0424 00:01:55.224841 2827 scope.go:122] "RemoveContainer" containerID="f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598" Apr 24 00:01:55.228093 containerd[1639]: time="2026-04-24T00:01:55.228001972Z" level=error msg="ContainerStatus for \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\": not found" Apr 24 00:01:55.230011 kubelet[2827]: E0424 00:01:55.229968 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\": not found" containerID="f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598" Apr 24 00:01:55.230146 kubelet[2827]: I0424 00:01:55.230001 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598"} err="failed to get container status \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\": rpc error: code = NotFound desc = an error occurred when try to find container \"f23a29a08cfe52377748e389f3a1133341f61fa79224871b05a5c07d23643598\": not found" Apr 24 00:01:55.230146 kubelet[2827]: I0424 00:01:55.230043 2827 scope.go:122] "RemoveContainer" containerID="289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52" Apr 24 00:01:55.234902 containerd[1639]: time="2026-04-24T00:01:55.234836761Z" level=info msg="RemoveContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\"" Apr 24 00:01:55.242395 systemd[1]: Removed slice kubepods-burstable-pod4239b600_718c_4e02_9846_e3af0a0fc526.slice - libcontainer container kubepods-burstable-pod4239b600_718c_4e02_9846_e3af0a0fc526.slice. Apr 24 00:01:55.242658 systemd[1]: kubepods-burstable-pod4239b600_718c_4e02_9846_e3af0a0fc526.slice: Consumed 5.279s CPU time, 126M memory peak, 112K read from disk, 13.3M written to disk. Apr 24 00:01:55.249580 containerd[1639]: time="2026-04-24T00:01:55.249116066Z" level=info msg="RemoveContainer for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" returns successfully" Apr 24 00:01:55.249689 kubelet[2827]: I0424 00:01:55.249290 2827 scope.go:122] "RemoveContainer" containerID="f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d" Apr 24 00:01:55.251583 containerd[1639]: time="2026-04-24T00:01:55.251506191Z" level=info msg="RemoveContainer for \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\"" Apr 24 00:01:55.257668 containerd[1639]: time="2026-04-24T00:01:55.257631703Z" level=info msg="RemoveContainer for \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" returns successfully" Apr 24 00:01:55.257892 kubelet[2827]: I0424 00:01:55.257858 2827 scope.go:122] "RemoveContainer" containerID="12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b" Apr 24 00:01:55.260448 containerd[1639]: time="2026-04-24T00:01:55.260379055Z" level=info msg="RemoveContainer for \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\"" Apr 24 00:01:55.265844 containerd[1639]: time="2026-04-24T00:01:55.265803413Z" level=info msg="RemoveContainer for \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" returns successfully" Apr 24 00:01:55.266144 kubelet[2827]: I0424 00:01:55.266119 2827 scope.go:122] "RemoveContainer" containerID="fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428" Apr 24 00:01:55.267549 containerd[1639]: time="2026-04-24T00:01:55.267515097Z" level=info msg="RemoveContainer for \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\"" Apr 24 00:01:55.270501 containerd[1639]: time="2026-04-24T00:01:55.270476310Z" level=info msg="RemoveContainer for \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" returns successfully" Apr 24 00:01:55.270623 kubelet[2827]: I0424 00:01:55.270602 2827 scope.go:122] "RemoveContainer" containerID="227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9" Apr 24 00:01:55.271597 containerd[1639]: time="2026-04-24T00:01:55.271573831Z" level=info msg="RemoveContainer for \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\"" Apr 24 00:01:55.274514 containerd[1639]: time="2026-04-24T00:01:55.274496656Z" level=info msg="RemoveContainer for \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" returns successfully" Apr 24 00:01:55.274715 kubelet[2827]: I0424 00:01:55.274683 2827 scope.go:122] "RemoveContainer" containerID="289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52" Apr 24 00:01:55.274905 containerd[1639]: time="2026-04-24T00:01:55.274867043Z" level=error msg="ContainerStatus for \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\": not found" Apr 24 00:01:55.275005 kubelet[2827]: E0424 00:01:55.274982 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\": not found" containerID="289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52" Apr 24 00:01:55.275029 kubelet[2827]: I0424 00:01:55.275013 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52"} err="failed to get container status \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\": rpc error: code = NotFound desc = an error occurred when try to find container \"289f77608824825359cfd81c0fecb1b6f0c8afd4330e0e2257f2ed70a25d2d52\": not found" Apr 24 00:01:55.275029 kubelet[2827]: I0424 00:01:55.275028 2827 scope.go:122] "RemoveContainer" containerID="f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d" Apr 24 00:01:55.275195 containerd[1639]: time="2026-04-24T00:01:55.275175677Z" level=error msg="ContainerStatus for \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\": not found" Apr 24 00:01:55.275337 kubelet[2827]: E0424 00:01:55.275315 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\": not found" containerID="f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d" Apr 24 00:01:55.275367 kubelet[2827]: I0424 00:01:55.275332 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d"} err="failed to get container status \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\": rpc error: code = NotFound desc = an error occurred when try to find container \"f7eeee08649a781a4837c26ded0db5a54b9e300b09db9fe1a4b28d35f0d8d45d\": not found" Apr 24 00:01:55.275367 kubelet[2827]: I0424 00:01:55.275344 2827 scope.go:122] "RemoveContainer" containerID="12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b" Apr 24 00:01:55.275516 containerd[1639]: time="2026-04-24T00:01:55.275493986Z" level=error msg="ContainerStatus for \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\": not found" Apr 24 00:01:55.275606 kubelet[2827]: E0424 00:01:55.275583 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\": not found" containerID="12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b" Apr 24 00:01:55.275606 kubelet[2827]: I0424 00:01:55.275600 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b"} err="failed to get container status \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\": rpc error: code = NotFound desc = an error occurred when try to find container \"12b310f256960c180b466926a3868ee6474d269efb185c32ee61196864135e8b\": not found" Apr 24 00:01:55.275722 kubelet[2827]: I0424 00:01:55.275609 2827 scope.go:122] "RemoveContainer" containerID="fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428" Apr 24 00:01:55.275882 containerd[1639]: time="2026-04-24T00:01:55.275815079Z" level=error msg="ContainerStatus for \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\": not found" Apr 24 00:01:55.275997 kubelet[2827]: E0424 00:01:55.275983 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\": not found" containerID="fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428" Apr 24 00:01:55.276070 kubelet[2827]: I0424 00:01:55.276041 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428"} err="failed to get container status \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\": rpc error: code = NotFound desc = an error occurred when try to find container \"fa35cbb7fb8b3f7edb2c18a6f019a6f937d5a0b1a8ce321d169e67ebc73a0428\": not found" Apr 24 00:01:55.276070 kubelet[2827]: I0424 00:01:55.276060 2827 scope.go:122] "RemoveContainer" containerID="227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9" Apr 24 00:01:55.276229 containerd[1639]: time="2026-04-24T00:01:55.276192957Z" level=error msg="ContainerStatus for \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\": not found" Apr 24 00:01:55.276335 kubelet[2827]: E0424 00:01:55.276313 2827 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\": not found" containerID="227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9" Apr 24 00:01:55.276387 kubelet[2827]: I0424 00:01:55.276363 2827 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9"} err="failed to get container status \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\": rpc error: code = NotFound desc = an error occurred when try to find container \"227f64f0b1a5ee47ac454943a4c92717aa7d7ea13b99ac51b99831c4bd217fe9\": not found" Apr 24 00:01:55.649868 systemd[1]: var-lib-kubelet-pods-9d2a4eeb\x2d2587\x2d473e\x2db4d1\x2d64cdde2220e0-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dl8p72.mount: Deactivated successfully. Apr 24 00:01:55.650033 systemd[1]: var-lib-kubelet-pods-4239b600\x2d718c\x2d4e02\x2d9846\x2de3af0a0fc526-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djj57b.mount: Deactivated successfully. Apr 24 00:01:55.650154 systemd[1]: var-lib-kubelet-pods-4239b600\x2d718c\x2d4e02\x2d9846\x2de3af0a0fc526-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 24 00:01:55.650259 systemd[1]: var-lib-kubelet-pods-4239b600\x2d718c\x2d4e02\x2d9846\x2de3af0a0fc526-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 24 00:01:56.529525 sshd[4423]: Connection closed by 20.229.252.112 port 48772 Apr 24 00:01:56.531556 sshd-session[4420]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:56.539262 systemd-logind[1616]: Session 26 logged out. Waiting for processes to exit. Apr 24 00:01:56.540728 systemd[1]: sshd@25-135.181.109.116:22-20.229.252.112:48772.service: Deactivated successfully. Apr 24 00:01:56.545175 systemd[1]: session-26.scope: Deactivated successfully. Apr 24 00:01:56.549055 systemd-logind[1616]: Removed session 26. Apr 24 00:01:56.577078 systemd[1]: Started sshd@26-135.181.109.116:22-20.229.252.112:41584.service - OpenSSH per-connection server daemon (20.229.252.112:41584). Apr 24 00:01:56.605836 kubelet[2827]: I0424 00:01:56.605801 2827 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="4239b600-718c-4e02-9846-e3af0a0fc526" path="/var/lib/kubelet/pods/4239b600-718c-4e02-9846-e3af0a0fc526/volumes" Apr 24 00:01:56.608937 kubelet[2827]: I0424 00:01:56.608861 2827 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="9d2a4eeb-2587-473e-b4d1-64cdde2220e0" path="/var/lib/kubelet/pods/9d2a4eeb-2587-473e-b4d1-64cdde2220e0/volumes" Apr 24 00:01:56.793841 sshd[4574]: Accepted publickey for core from 20.229.252.112 port 41584 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:56.796432 sshd-session[4574]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:56.805235 systemd-logind[1616]: New session 27 of user core. Apr 24 00:01:56.811198 systemd[1]: Started session-27.scope - Session 27 of User core. Apr 24 00:01:57.217660 sshd[4577]: Connection closed by 20.229.252.112 port 41584 Apr 24 00:01:57.218134 sshd-session[4574]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:57.227678 systemd[1]: sshd@26-135.181.109.116:22-20.229.252.112:41584.service: Deactivated successfully. Apr 24 00:01:57.230207 systemd[1]: session-27.scope: Deactivated successfully. Apr 24 00:01:57.232467 systemd-logind[1616]: Session 27 logged out. Waiting for processes to exit. Apr 24 00:01:57.235177 systemd[1]: Created slice kubepods-burstable-pod17dabd81_36aa_4810_8eeb_e95842a9d44e.slice - libcontainer container kubepods-burstable-pod17dabd81_36aa_4810_8eeb_e95842a9d44e.slice. Apr 24 00:01:57.237665 systemd-logind[1616]: Removed session 27. Apr 24 00:01:57.257784 systemd[1]: Started sshd@27-135.181.109.116:22-20.229.252.112:41598.service - OpenSSH per-connection server daemon (20.229.252.112:41598). Apr 24 00:01:57.330801 kubelet[2827]: I0424 00:01:57.330733 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-bpf-maps\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.330801 kubelet[2827]: I0424 00:01:57.330778 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-hostproc\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.330801 kubelet[2827]: I0424 00:01:57.330800 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-etc-cni-netd\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330818 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-lib-modules\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330836 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-host-proc-sys-net\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330855 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-cni-path\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330873 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/17dabd81-36aa-4810-8eeb-e95842a9d44e-cilium-config-path\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330892 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/17dabd81-36aa-4810-8eeb-e95842a9d44e-clustermesh-secrets\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331458 kubelet[2827]: I0424 00:01:57.330910 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-cilium-run\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331699 kubelet[2827]: I0424 00:01:57.330926 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-xtables-lock\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331699 kubelet[2827]: I0424 00:01:57.331105 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/17dabd81-36aa-4810-8eeb-e95842a9d44e-cilium-ipsec-secrets\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331699 kubelet[2827]: I0424 00:01:57.331230 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/17dabd81-36aa-4810-8eeb-e95842a9d44e-hubble-tls\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331699 kubelet[2827]: I0424 00:01:57.331447 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-cilium-cgroup\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.331909 kubelet[2827]: I0424 00:01:57.331875 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ztsf\" (UniqueName: \"kubernetes.io/projected/17dabd81-36aa-4810-8eeb-e95842a9d44e-kube-api-access-4ztsf\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.332037 kubelet[2827]: I0424 00:01:57.331936 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/17dabd81-36aa-4810-8eeb-e95842a9d44e-host-proc-sys-kernel\") pod \"cilium-pgjm5\" (UID: \"17dabd81-36aa-4810-8eeb-e95842a9d44e\") " pod="kube-system/cilium-pgjm5" Apr 24 00:01:57.450181 sshd[4587]: Accepted publickey for core from 20.229.252.112 port 41598 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:57.453145 sshd-session[4587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:57.482695 systemd-logind[1616]: New session 28 of user core. Apr 24 00:01:57.489656 systemd[1]: Started session-28.scope - Session 28 of User core. Apr 24 00:01:57.541815 containerd[1639]: time="2026-04-24T00:01:57.541759487Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-pgjm5,Uid:17dabd81-36aa-4810-8eeb-e95842a9d44e,Namespace:kube-system,Attempt:0,}" Apr 24 00:01:57.563663 sshd[4594]: Connection closed by 20.229.252.112 port 41598 Apr 24 00:01:57.564651 containerd[1639]: time="2026-04-24T00:01:57.564344032Z" level=info msg="connecting to shim a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" namespace=k8s.io protocol=ttrpc version=3 Apr 24 00:01:57.565644 sshd-session[4587]: pam_unix(sshd:session): session closed for user core Apr 24 00:01:57.575738 systemd[1]: sshd@27-135.181.109.116:22-20.229.252.112:41598.service: Deactivated successfully. Apr 24 00:01:57.590777 systemd[1]: session-28.scope: Deactivated successfully. Apr 24 00:01:57.594062 systemd-logind[1616]: Session 28 logged out. Waiting for processes to exit. Apr 24 00:01:57.605957 systemd-logind[1616]: Removed session 28. Apr 24 00:01:57.609645 systemd[1]: Started cri-containerd-a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f.scope - libcontainer container a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f. Apr 24 00:01:57.611440 systemd[1]: Started sshd@28-135.181.109.116:22-20.229.252.112:41606.service - OpenSSH per-connection server daemon (20.229.252.112:41606). Apr 24 00:01:57.635851 containerd[1639]: time="2026-04-24T00:01:57.635805120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-pgjm5,Uid:17dabd81-36aa-4810-8eeb-e95842a9d44e,Namespace:kube-system,Attempt:0,} returns sandbox id \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\"" Apr 24 00:01:57.642315 containerd[1639]: time="2026-04-24T00:01:57.641445041Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 24 00:01:57.648829 containerd[1639]: time="2026-04-24T00:01:57.648810498Z" level=info msg="Container 248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:01:57.652961 containerd[1639]: time="2026-04-24T00:01:57.652933439Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d\"" Apr 24 00:01:57.654108 containerd[1639]: time="2026-04-24T00:01:57.653970438Z" level=info msg="StartContainer for \"248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d\"" Apr 24 00:01:57.654825 containerd[1639]: time="2026-04-24T00:01:57.654765875Z" level=info msg="connecting to shim 248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" protocol=ttrpc version=3 Apr 24 00:01:57.675648 systemd[1]: Started cri-containerd-248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d.scope - libcontainer container 248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d. Apr 24 00:01:57.703834 containerd[1639]: time="2026-04-24T00:01:57.703763280Z" level=info msg="StartContainer for \"248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d\" returns successfully" Apr 24 00:01:57.708756 systemd[1]: cri-containerd-248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d.scope: Deactivated successfully. Apr 24 00:01:57.710238 containerd[1639]: time="2026-04-24T00:01:57.710219941Z" level=info msg="received container exit event container_id:\"248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d\" id:\"248926c35cc0754644fe542a82c0e12166ca157c38a34481e044b4c77587879d\" pid:4662 exited_at:{seconds:1776988917 nanos:709909364}" Apr 24 00:01:57.801311 sshd[4633]: Accepted publickey for core from 20.229.252.112 port 41606 ssh2: RSA SHA256:CJukKVYD4j/mY5InY6hVIBK/YO021iJD6DQAJnGodf4 Apr 24 00:01:57.805368 sshd-session[4633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 24 00:01:57.818083 systemd-logind[1616]: New session 29 of user core. Apr 24 00:01:57.820812 systemd[1]: Started session-29.scope - Session 29 of User core. Apr 24 00:01:58.256718 containerd[1639]: time="2026-04-24T00:01:58.256605701Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 24 00:01:58.266730 containerd[1639]: time="2026-04-24T00:01:58.266675704Z" level=info msg="Container a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:01:58.276853 containerd[1639]: time="2026-04-24T00:01:58.276812328Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433\"" Apr 24 00:01:58.278711 containerd[1639]: time="2026-04-24T00:01:58.278677514Z" level=info msg="StartContainer for \"a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433\"" Apr 24 00:01:58.281110 containerd[1639]: time="2026-04-24T00:01:58.281069571Z" level=info msg="connecting to shim a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" protocol=ttrpc version=3 Apr 24 00:01:58.302627 systemd[1]: Started cri-containerd-a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433.scope - libcontainer container a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433. Apr 24 00:01:58.331413 containerd[1639]: time="2026-04-24T00:01:58.331376635Z" level=info msg="StartContainer for \"a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433\" returns successfully" Apr 24 00:01:58.336862 systemd[1]: cri-containerd-a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433.scope: Deactivated successfully. Apr 24 00:01:58.337561 containerd[1639]: time="2026-04-24T00:01:58.337299093Z" level=info msg="received container exit event container_id:\"a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433\" id:\"a8f9ed4855a5d09e5dbc388675058626ef4dfb097769f44f0380d25fb85cb433\" pid:4715 exited_at:{seconds:1776988918 nanos:337035366}" Apr 24 00:01:58.714460 kubelet[2827]: E0424 00:01:58.714377 2827 kubelet.go:3130] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 24 00:01:59.257690 containerd[1639]: time="2026-04-24T00:01:59.257637575Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 24 00:01:59.284001 containerd[1639]: time="2026-04-24T00:01:59.279843940Z" level=info msg="Container 4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:01:59.298056 containerd[1639]: time="2026-04-24T00:01:59.298005082Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af\"" Apr 24 00:01:59.298958 containerd[1639]: time="2026-04-24T00:01:59.298933077Z" level=info msg="StartContainer for \"4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af\"" Apr 24 00:01:59.301504 containerd[1639]: time="2026-04-24T00:01:59.301442210Z" level=info msg="connecting to shim 4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" protocol=ttrpc version=3 Apr 24 00:01:59.325693 systemd[1]: Started cri-containerd-4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af.scope - libcontainer container 4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af. Apr 24 00:01:59.392837 systemd[1]: cri-containerd-4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af.scope: Deactivated successfully. Apr 24 00:01:59.394448 containerd[1639]: time="2026-04-24T00:01:59.394350700Z" level=info msg="received container exit event container_id:\"4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af\" id:\"4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af\" pid:4757 exited_at:{seconds:1776988919 nanos:394245731}" Apr 24 00:01:59.395113 containerd[1639]: time="2026-04-24T00:01:59.395077953Z" level=info msg="StartContainer for \"4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af\" returns successfully" Apr 24 00:01:59.416417 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4b0c8b20946ccd323ac687446939010400ceaa935d46dea2d3d6bea3308857af-rootfs.mount: Deactivated successfully. Apr 24 00:02:00.272608 containerd[1639]: time="2026-04-24T00:02:00.272184075Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 24 00:02:00.286161 containerd[1639]: time="2026-04-24T00:02:00.286040033Z" level=info msg="Container c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:02:00.288472 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1592020279.mount: Deactivated successfully. Apr 24 00:02:00.294232 containerd[1639]: time="2026-04-24T00:02:00.294154845Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36\"" Apr 24 00:02:00.295585 containerd[1639]: time="2026-04-24T00:02:00.295143683Z" level=info msg="StartContainer for \"c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36\"" Apr 24 00:02:00.296012 containerd[1639]: time="2026-04-24T00:02:00.295996415Z" level=info msg="connecting to shim c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" protocol=ttrpc version=3 Apr 24 00:02:00.319637 systemd[1]: Started cri-containerd-c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36.scope - libcontainer container c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36. Apr 24 00:02:00.342575 systemd[1]: cri-containerd-c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36.scope: Deactivated successfully. Apr 24 00:02:00.344592 containerd[1639]: time="2026-04-24T00:02:00.344568308Z" level=info msg="received container exit event container_id:\"c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36\" id:\"c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36\" pid:4798 exited_at:{seconds:1776988920 nanos:343779613}" Apr 24 00:02:00.351088 containerd[1639]: time="2026-04-24T00:02:00.351045459Z" level=info msg="StartContainer for \"c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36\" returns successfully" Apr 24 00:02:00.362430 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c1e4ec100eadf5c8cd81071a1e4e756289f5355988db46c1db6574243b770d36-rootfs.mount: Deactivated successfully. Apr 24 00:02:01.279481 containerd[1639]: time="2026-04-24T00:02:01.279405943Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 24 00:02:01.301782 containerd[1639]: time="2026-04-24T00:02:01.300917262Z" level=info msg="Container 014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:02:01.305342 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3550336729.mount: Deactivated successfully. Apr 24 00:02:01.316061 containerd[1639]: time="2026-04-24T00:02:01.316018182Z" level=info msg="CreateContainer within sandbox \"a5e11059c5cce774ba0a9c18d4b1ee6bf9e5971ed3144475d92bda253fcfaa1f\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f\"" Apr 24 00:02:01.317277 containerd[1639]: time="2026-04-24T00:02:01.316661749Z" level=info msg="StartContainer for \"014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f\"" Apr 24 00:02:01.317743 containerd[1639]: time="2026-04-24T00:02:01.317708112Z" level=info msg="connecting to shim 014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f" address="unix:///run/containerd/s/5f22ce38ee8a86e4f885d1708b47f3bca55c98c350a3490ad2fe05183273a74f" protocol=ttrpc version=3 Apr 24 00:02:01.343659 systemd[1]: Started cri-containerd-014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f.scope - libcontainer container 014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f. Apr 24 00:02:01.381023 containerd[1639]: time="2026-04-24T00:02:01.380953181Z" level=info msg="StartContainer for \"014c7a403d532d2af4e20e62f3a83216b7bdf2c1e61713bc706924a874f17e0f\" returns successfully" Apr 24 00:02:01.688709 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-vaes-avx10_512)) Apr 24 00:02:02.302175 kubelet[2827]: I0424 00:02:02.301846 2827 pod_startup_latency_tracker.go:108] "Observed pod startup duration" pod="kube-system/cilium-pgjm5" podStartSLOduration=5.301826971 podStartE2EDuration="5.301826971s" podCreationTimestamp="2026-04-24 00:01:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-24 00:02:02.299775056 +0000 UTC m=+203.790942520" watchObservedRunningTime="2026-04-24 00:02:02.301826971 +0000 UTC m=+203.792994446" Apr 24 00:02:03.307891 kubelet[2827]: I0424 00:02:03.307850 2827 setters.go:546] "Node became not ready" node="ci-4459-2-4-n-09e3934cf7" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-04-24T00:02:03Z","lastTransitionTime":"2026-04-24T00:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Apr 24 00:02:04.259181 systemd-networkd[1503]: lxc_health: Link UP Apr 24 00:02:04.272645 systemd-networkd[1503]: lxc_health: Gained carrier Apr 24 00:02:06.119778 systemd-networkd[1503]: lxc_health: Gained IPv6LL Apr 24 00:02:10.549298 sshd[4696]: Connection closed by 20.229.252.112 port 41606 Apr 24 00:02:10.550912 sshd-session[4633]: pam_unix(sshd:session): session closed for user core Apr 24 00:02:10.557999 systemd[1]: sshd@28-135.181.109.116:22-20.229.252.112:41606.service: Deactivated successfully. Apr 24 00:02:10.562415 systemd[1]: session-29.scope: Deactivated successfully. Apr 24 00:02:10.564617 systemd-logind[1616]: Session 29 logged out. Waiting for processes to exit. Apr 24 00:02:10.568279 systemd-logind[1616]: Removed session 29. Apr 24 00:02:28.284445 systemd[1]: cri-containerd-00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076.scope: Deactivated successfully. Apr 24 00:02:28.285876 systemd[1]: cri-containerd-00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076.scope: Consumed 3.238s CPU time, 62M memory peak. Apr 24 00:02:28.289413 containerd[1639]: time="2026-04-24T00:02:28.289339200Z" level=info msg="received container exit event container_id:\"00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076\" id:\"00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076\" pid:2682 exit_status:1 exited_at:{seconds:1776988948 nanos:287996121}" Apr 24 00:02:28.338392 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076-rootfs.mount: Deactivated successfully. Apr 24 00:02:28.350571 kubelet[2827]: I0424 00:02:28.350242 2827 scope.go:122] "RemoveContainer" containerID="00c0f031f4ea228a1965d55ec64b2c766ec8a58020d1057c496fce63474c9076" Apr 24 00:02:28.353596 containerd[1639]: time="2026-04-24T00:02:28.353211250Z" level=info msg="CreateContainer within sandbox \"c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 24 00:02:28.370602 containerd[1639]: time="2026-04-24T00:02:28.368810677Z" level=info msg="Container cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:02:28.378626 containerd[1639]: time="2026-04-24T00:02:28.378574912Z" level=info msg="CreateContainer within sandbox \"c55a9b50376eaea1042110b919b8045ec6252997334a7b4388fc3157de3450ad\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636\"" Apr 24 00:02:28.379319 containerd[1639]: time="2026-04-24T00:02:28.379181743Z" level=info msg="StartContainer for \"cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636\"" Apr 24 00:02:28.380715 containerd[1639]: time="2026-04-24T00:02:28.380653087Z" level=info msg="connecting to shim cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636" address="unix:///run/containerd/s/ecfddf6f77c4f6e0c148e9c1ab5bb9e1113bf51c86d51e263e3714c31a455133" protocol=ttrpc version=3 Apr 24 00:02:28.408650 systemd[1]: Started cri-containerd-cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636.scope - libcontainer container cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636. Apr 24 00:02:28.454910 containerd[1639]: time="2026-04-24T00:02:28.454829050Z" level=info msg="StartContainer for \"cd3bf72e488882ebe46a63d30b13b30847ad39324564cd0d757a0261a264c636\" returns successfully" Apr 24 00:02:28.521862 kubelet[2827]: E0424 00:02:28.521774 2827 controller.go:251] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:42844->10.0.0.2:2379: read: connection timed out" Apr 24 00:02:30.377061 kubelet[2827]: E0424 00:02:30.376838 2827 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:42472->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4459-2-4-n-09e3934cf7.18a92206a4e82d94 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4459-2-4-n-09e3934cf7,UID:53c5f6c6d24b0d84b374cb52cf7e654c,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4459-2-4-n-09e3934cf7,},FirstTimestamp:2026-04-24 00:02:19.906198932 +0000 UTC m=+221.397366396,LastTimestamp:2026-04-24 00:02:19.906198932 +0000 UTC m=+221.397366396,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459-2-4-n-09e3934cf7,}" Apr 24 00:02:33.539299 systemd[1]: cri-containerd-3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c.scope: Deactivated successfully. Apr 24 00:02:33.540521 systemd[1]: cri-containerd-3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c.scope: Consumed 1.894s CPU time, 21.5M memory peak. Apr 24 00:02:33.543499 containerd[1639]: time="2026-04-24T00:02:33.543106299Z" level=info msg="received container exit event container_id:\"3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c\" id:\"3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c\" pid:2656 exit_status:1 exited_at:{seconds:1776988953 nanos:542115202}" Apr 24 00:02:33.584991 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c-rootfs.mount: Deactivated successfully. Apr 24 00:02:34.371207 kubelet[2827]: I0424 00:02:34.371132 2827 scope.go:122] "RemoveContainer" containerID="3e0cff33f093f403cc14c8c95393e8322fb3ddc2af4a66e104d7048474ee054c" Apr 24 00:02:34.373645 containerd[1639]: time="2026-04-24T00:02:34.373580452Z" level=info msg="CreateContainer within sandbox \"754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 24 00:02:34.388261 containerd[1639]: time="2026-04-24T00:02:34.388194847Z" level=info msg="Container a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3: CDI devices from CRI Config.CDIDevices: []" Apr 24 00:02:34.398219 containerd[1639]: time="2026-04-24T00:02:34.398144643Z" level=info msg="CreateContainer within sandbox \"754cf9ec8ec742d4ec7378145c592d5b8032e927e2ab0535cfe558d9ebf7ef4f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3\"" Apr 24 00:02:34.398749 containerd[1639]: time="2026-04-24T00:02:34.398715877Z" level=info msg="StartContainer for \"a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3\"" Apr 24 00:02:34.399742 containerd[1639]: time="2026-04-24T00:02:34.399705812Z" level=info msg="connecting to shim a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3" address="unix:///run/containerd/s/a89b511aed31b4b13dce274d6ebfa1286ada34a47b8a79bb77acf539f027b834" protocol=ttrpc version=3 Apr 24 00:02:34.424630 systemd[1]: Started cri-containerd-a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3.scope - libcontainer container a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3. Apr 24 00:02:34.489711 containerd[1639]: time="2026-04-24T00:02:34.489649708Z" level=info msg="StartContainer for \"a47897a56a7aada4f7928e05842122acdc5a43e0cc78dc4b10e5c4824a7933e3\" returns successfully" Apr 24 00:02:38.523215 kubelet[2827]: E0424 00:02:38.522978 2827 controller.go:251] "Failed to update lease" err="Put \"https://135.181.109.116:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-4-n-09e3934cf7?timeout=10s\": context deadline exceeded" Apr 24 00:02:38.611086 containerd[1639]: time="2026-04-24T00:02:38.610999174Z" level=info msg="StopPodSandbox for \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\"" Apr 24 00:02:38.612284 containerd[1639]: time="2026-04-24T00:02:38.611186761Z" level=info msg="TearDown network for sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" successfully" Apr 24 00:02:38.612284 containerd[1639]: time="2026-04-24T00:02:38.611203386Z" level=info msg="StopPodSandbox for \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" returns successfully" Apr 24 00:02:38.612284 containerd[1639]: time="2026-04-24T00:02:38.612235455Z" level=info msg="RemovePodSandbox for \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\"" Apr 24 00:02:38.612429 containerd[1639]: time="2026-04-24T00:02:38.612266021Z" level=info msg="Forcibly stopping sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\"" Apr 24 00:02:38.612482 containerd[1639]: time="2026-04-24T00:02:38.612438495Z" level=info msg="TearDown network for sandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" successfully" Apr 24 00:02:38.615186 containerd[1639]: time="2026-04-24T00:02:38.615116920Z" level=info msg="Ensure that sandbox ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0 in task-service has been cleanup successfully" Apr 24 00:02:38.619919 containerd[1639]: time="2026-04-24T00:02:38.619872731Z" level=info msg="RemovePodSandbox \"ee78405af22d79513ee4a479bf9643a78e9b6271daa592d2d608ee10b3e5afc0\" returns successfully" Apr 24 00:02:38.620567 containerd[1639]: time="2026-04-24T00:02:38.620410513Z" level=info msg="StopPodSandbox for \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\"" Apr 24 00:02:38.620716 containerd[1639]: time="2026-04-24T00:02:38.620527202Z" level=info msg="TearDown network for sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" successfully" Apr 24 00:02:38.620894 containerd[1639]: time="2026-04-24T00:02:38.620804026Z" level=info msg="StopPodSandbox for \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" returns successfully" Apr 24 00:02:38.621390 containerd[1639]: time="2026-04-24T00:02:38.621350280Z" level=info msg="RemovePodSandbox for \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\"" Apr 24 00:02:38.621390 containerd[1639]: time="2026-04-24T00:02:38.621384543Z" level=info msg="Forcibly stopping sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\"" Apr 24 00:02:38.621508 containerd[1639]: time="2026-04-24T00:02:38.621465818Z" level=info msg="TearDown network for sandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" successfully" Apr 24 00:02:38.623898 containerd[1639]: time="2026-04-24T00:02:38.623851886Z" level=info msg="Ensure that sandbox 5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5 in task-service has been cleanup successfully" Apr 24 00:02:38.627625 containerd[1639]: time="2026-04-24T00:02:38.627586346Z" level=info msg="RemovePodSandbox \"5fdf550899dd80b8078d97dbe6d18956671feb89a8f61c5d5818ee90b9c9a8f5\" returns successfully" Apr 24 00:02:48.525648 kubelet[2827]: E0424 00:02:48.524146 2827 controller.go:251] "Failed to update lease" err="Put \"https://135.181.109.116:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459-2-4-n-09e3934cf7?timeout=10s\": context deadline exceeded"