May 8 07:54:54.081060 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu May 8 02:00:00 -00 2025 May 8 07:54:54.081086 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=c6e5fdf442cda427c9e51c476b5f39b01a7265d89660ab7c7d9178b52b2cc04b May 8 07:54:54.081098 kernel: BIOS-provided physical RAM map: May 8 07:54:54.081106 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable May 8 07:54:54.081114 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved May 8 07:54:54.081121 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved May 8 07:54:54.081130 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000bffdcfff] usable May 8 07:54:54.081138 kernel: BIOS-e820: [mem 0x00000000bffdd000-0x00000000bfffffff] reserved May 8 07:54:54.081145 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 8 07:54:54.081153 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved May 8 07:54:54.081160 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000013fffffff] usable May 8 07:54:54.081169 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved May 8 07:54:54.081177 kernel: NX (Execute Disable) protection: active May 8 07:54:54.081185 kernel: APIC: Static calls initialized May 8 07:54:54.081194 kernel: SMBIOS 3.0.0 present. May 8 07:54:54.081202 kernel: DMI: OpenStack Foundation OpenStack Nova, BIOS 1.16.3-debian-1.16.3-2 04/01/2014 May 8 07:54:54.081211 kernel: Hypervisor detected: KVM May 8 07:54:54.081219 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 8 07:54:54.081227 kernel: kvm-clock: using sched offset of 4649358148 cycles May 8 07:54:54.081235 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 8 07:54:54.081244 kernel: tsc: Detected 1996.249 MHz processor May 8 07:54:54.081252 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 8 07:54:54.081261 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 8 07:54:54.081269 kernel: last_pfn = 0x140000 max_arch_pfn = 0x400000000 May 8 07:54:54.081278 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs May 8 07:54:54.081286 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 8 07:54:54.081296 kernel: last_pfn = 0xbffdd max_arch_pfn = 0x400000000 May 8 07:54:54.081305 kernel: ACPI: Early table checksum verification disabled May 8 07:54:54.081313 kernel: ACPI: RSDP 0x00000000000F51E0 000014 (v00 BOCHS ) May 8 07:54:54.081321 kernel: ACPI: RSDT 0x00000000BFFE1B65 000030 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 07:54:54.081329 kernel: ACPI: FACP 0x00000000BFFE1A49 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 07:54:54.081337 kernel: ACPI: DSDT 0x00000000BFFE0040 001A09 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 07:54:54.081346 kernel: ACPI: FACS 0x00000000BFFE0000 000040 May 8 07:54:54.081354 kernel: ACPI: APIC 0x00000000BFFE1ABD 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 8 07:54:54.081362 kernel: ACPI: WAET 0x00000000BFFE1B3D 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 07:54:54.081372 kernel: ACPI: Reserving FACP table memory at [mem 0xbffe1a49-0xbffe1abc] May 8 07:54:54.081380 kernel: ACPI: Reserving DSDT table memory at [mem 0xbffe0040-0xbffe1a48] May 8 07:54:54.081388 kernel: ACPI: Reserving FACS table memory at [mem 0xbffe0000-0xbffe003f] May 8 07:54:54.081396 kernel: ACPI: Reserving APIC table memory at [mem 0xbffe1abd-0xbffe1b3c] May 8 07:54:54.081405 kernel: ACPI: Reserving WAET table memory at [mem 0xbffe1b3d-0xbffe1b64] May 8 07:54:54.081416 kernel: No NUMA configuration found May 8 07:54:54.081426 kernel: Faking a node at [mem 0x0000000000000000-0x000000013fffffff] May 8 07:54:54.081435 kernel: NODE_DATA(0) allocated [mem 0x13fff5000-0x13fffcfff] May 8 07:54:54.081444 kernel: Zone ranges: May 8 07:54:54.081452 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 8 07:54:54.081461 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] May 8 07:54:54.081469 kernel: Normal [mem 0x0000000100000000-0x000000013fffffff] May 8 07:54:54.081477 kernel: Device empty May 8 07:54:54.081486 kernel: Movable zone start for each node May 8 07:54:54.081496 kernel: Early memory node ranges May 8 07:54:54.081505 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] May 8 07:54:54.081513 kernel: node 0: [mem 0x0000000000100000-0x00000000bffdcfff] May 8 07:54:54.081521 kernel: node 0: [mem 0x0000000100000000-0x000000013fffffff] May 8 07:54:54.081530 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000013fffffff] May 8 07:54:54.081539 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 8 07:54:54.081547 kernel: On node 0, zone DMA: 97 pages in unavailable ranges May 8 07:54:54.081556 kernel: On node 0, zone Normal: 35 pages in unavailable ranges May 8 07:54:54.081564 kernel: ACPI: PM-Timer IO Port: 0x608 May 8 07:54:54.081573 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 8 07:54:54.081583 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 8 07:54:54.081592 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 8 07:54:54.081600 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 8 07:54:54.081609 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 8 07:54:54.081617 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 8 07:54:54.081626 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 8 07:54:54.081634 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 8 07:54:54.081643 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs May 8 07:54:54.081651 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 8 07:54:54.081661 kernel: [mem 0xc0000000-0xfeffbfff] available for PCI devices May 8 07:54:54.081670 kernel: Booting paravirtualized kernel on KVM May 8 07:54:54.081679 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 8 07:54:54.081687 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 8 07:54:54.081696 kernel: percpu: Embedded 58 pages/cpu s197480 r8192 d31896 u1048576 May 8 07:54:54.081705 kernel: pcpu-alloc: s197480 r8192 d31896 u1048576 alloc=1*2097152 May 8 07:54:54.081713 kernel: pcpu-alloc: [0] 0 1 May 8 07:54:54.081721 kernel: kvm-guest: PV spinlocks disabled, no host support May 8 07:54:54.081733 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=c6e5fdf442cda427c9e51c476b5f39b01a7265d89660ab7c7d9178b52b2cc04b May 8 07:54:54.081742 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 8 07:54:54.081751 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 8 07:54:54.081760 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 8 07:54:54.081768 kernel: Fallback order for Node 0: 0 May 8 07:54:54.081777 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1031901 May 8 07:54:54.081785 kernel: Policy zone: Normal May 8 07:54:54.081794 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 8 07:54:54.081804 kernel: software IO TLB: area num 2. May 8 07:54:54.081813 kernel: Memory: 3968244K/4193772K available (14336K kernel code, 2309K rwdata, 9040K rodata, 53684K init, 1592K bss, 225268K reserved, 0K cma-reserved) May 8 07:54:54.081822 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 8 07:54:54.081830 kernel: ftrace: allocating 38190 entries in 150 pages May 8 07:54:54.081839 kernel: ftrace: allocated 150 pages with 4 groups May 8 07:54:54.081847 kernel: Dynamic Preempt: voluntary May 8 07:54:54.081856 kernel: rcu: Preemptible hierarchical RCU implementation. May 8 07:54:54.081865 kernel: rcu: RCU event tracing is enabled. May 8 07:54:54.081874 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 8 07:54:54.081884 kernel: Trampoline variant of Tasks RCU enabled. May 8 07:54:54.081893 kernel: Rude variant of Tasks RCU enabled. May 8 07:54:54.081901 kernel: Tracing variant of Tasks RCU enabled. May 8 07:54:54.081910 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 8 07:54:54.081934 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 8 07:54:54.081942 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 8 07:54:54.081951 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 8 07:54:54.082980 kernel: Console: colour VGA+ 80x25 May 8 07:54:54.082992 kernel: printk: console [tty0] enabled May 8 07:54:54.083001 kernel: printk: console [ttyS0] enabled May 8 07:54:54.083014 kernel: ACPI: Core revision 20230628 May 8 07:54:54.083024 kernel: APIC: Switch to symmetric I/O mode setup May 8 07:54:54.083034 kernel: x2apic enabled May 8 07:54:54.083043 kernel: APIC: Switched APIC routing to: physical x2apic May 8 07:54:54.083052 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 8 07:54:54.083062 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized May 8 07:54:54.083071 kernel: Calibrating delay loop (skipped) preset value.. 3992.49 BogoMIPS (lpj=1996249) May 8 07:54:54.083081 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 May 8 07:54:54.083090 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 May 8 07:54:54.083101 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 8 07:54:54.083111 kernel: Spectre V2 : Mitigation: Retpolines May 8 07:54:54.083120 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch May 8 07:54:54.083129 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT May 8 07:54:54.083139 kernel: Speculative Store Bypass: Vulnerable May 8 07:54:54.083148 kernel: x86/fpu: x87 FPU will use FXSAVE May 8 07:54:54.083164 kernel: Freeing SMP alternatives memory: 32K May 8 07:54:54.083175 kernel: pid_max: default: 32768 minimum: 301 May 8 07:54:54.083185 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 8 07:54:54.083194 kernel: landlock: Up and running. May 8 07:54:54.083204 kernel: SELinux: Initializing. May 8 07:54:54.083214 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 07:54:54.083225 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 07:54:54.083236 kernel: smpboot: CPU0: AMD Intel Core i7 9xx (Nehalem Class Core i7) (family: 0x6, model: 0x1a, stepping: 0x3) May 8 07:54:54.083246 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 8 07:54:54.083255 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 8 07:54:54.083267 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 8 07:54:54.083276 kernel: Performance Events: AMD PMU driver. May 8 07:54:54.083286 kernel: ... version: 0 May 8 07:54:54.083296 kernel: ... bit width: 48 May 8 07:54:54.083305 kernel: ... generic registers: 4 May 8 07:54:54.083315 kernel: ... value mask: 0000ffffffffffff May 8 07:54:54.083324 kernel: ... max period: 00007fffffffffff May 8 07:54:54.083334 kernel: ... fixed-purpose events: 0 May 8 07:54:54.083343 kernel: ... event mask: 000000000000000f May 8 07:54:54.083355 kernel: signal: max sigframe size: 1440 May 8 07:54:54.083364 kernel: rcu: Hierarchical SRCU implementation. May 8 07:54:54.083374 kernel: rcu: Max phase no-delay instances is 400. May 8 07:54:54.083384 kernel: smp: Bringing up secondary CPUs ... May 8 07:54:54.083393 kernel: smpboot: x86: Booting SMP configuration: May 8 07:54:54.083403 kernel: .... node #0, CPUs: #1 May 8 07:54:54.083412 kernel: smp: Brought up 1 node, 2 CPUs May 8 07:54:54.083422 kernel: smpboot: Max logical packages: 2 May 8 07:54:54.083432 kernel: smpboot: Total of 2 processors activated (7984.99 BogoMIPS) May 8 07:54:54.083441 kernel: devtmpfs: initialized May 8 07:54:54.083453 kernel: x86/mm: Memory block size: 128MB May 8 07:54:54.083462 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 8 07:54:54.083472 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 8 07:54:54.083482 kernel: pinctrl core: initialized pinctrl subsystem May 8 07:54:54.083492 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 8 07:54:54.083501 kernel: audit: initializing netlink subsys (disabled) May 8 07:54:54.083512 kernel: audit: type=2000 audit(1746690890.278:1): state=initialized audit_enabled=0 res=1 May 8 07:54:54.083522 kernel: thermal_sys: Registered thermal governor 'step_wise' May 8 07:54:54.083531 kernel: thermal_sys: Registered thermal governor 'user_space' May 8 07:54:54.083542 kernel: cpuidle: using governor menu May 8 07:54:54.083551 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 8 07:54:54.083560 kernel: dca service started, version 1.12.1 May 8 07:54:54.083569 kernel: PCI: Using configuration type 1 for base access May 8 07:54:54.083578 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 8 07:54:54.083587 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 8 07:54:54.083596 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 8 07:54:54.083605 kernel: ACPI: Added _OSI(Module Device) May 8 07:54:54.083614 kernel: ACPI: Added _OSI(Processor Device) May 8 07:54:54.083625 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 8 07:54:54.083633 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 8 07:54:54.083643 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 8 07:54:54.083652 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC May 8 07:54:54.083661 kernel: ACPI: Interpreter enabled May 8 07:54:54.083670 kernel: ACPI: PM: (supports S0 S3 S5) May 8 07:54:54.083678 kernel: ACPI: Using IOAPIC for interrupt routing May 8 07:54:54.083688 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 8 07:54:54.083697 kernel: PCI: Using E820 reservations for host bridge windows May 8 07:54:54.083707 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 8 07:54:54.083716 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 8 07:54:54.083866 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 8 07:54:54.084996 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 8 07:54:54.085096 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 8 07:54:54.085111 kernel: acpiphp: Slot [3] registered May 8 07:54:54.085121 kernel: acpiphp: Slot [4] registered May 8 07:54:54.085134 kernel: acpiphp: Slot [5] registered May 8 07:54:54.085143 kernel: acpiphp: Slot [6] registered May 8 07:54:54.085152 kernel: acpiphp: Slot [7] registered May 8 07:54:54.085161 kernel: acpiphp: Slot [8] registered May 8 07:54:54.085170 kernel: acpiphp: Slot [9] registered May 8 07:54:54.085178 kernel: acpiphp: Slot [10] registered May 8 07:54:54.085187 kernel: acpiphp: Slot [11] registered May 8 07:54:54.085196 kernel: acpiphp: Slot [12] registered May 8 07:54:54.085205 kernel: acpiphp: Slot [13] registered May 8 07:54:54.085216 kernel: acpiphp: Slot [14] registered May 8 07:54:54.085225 kernel: acpiphp: Slot [15] registered May 8 07:54:54.085234 kernel: acpiphp: Slot [16] registered May 8 07:54:54.085243 kernel: acpiphp: Slot [17] registered May 8 07:54:54.085251 kernel: acpiphp: Slot [18] registered May 8 07:54:54.085260 kernel: acpiphp: Slot [19] registered May 8 07:54:54.085269 kernel: acpiphp: Slot [20] registered May 8 07:54:54.085278 kernel: acpiphp: Slot [21] registered May 8 07:54:54.085287 kernel: acpiphp: Slot [22] registered May 8 07:54:54.085295 kernel: acpiphp: Slot [23] registered May 8 07:54:54.085306 kernel: acpiphp: Slot [24] registered May 8 07:54:54.085316 kernel: acpiphp: Slot [25] registered May 8 07:54:54.085324 kernel: acpiphp: Slot [26] registered May 8 07:54:54.085333 kernel: acpiphp: Slot [27] registered May 8 07:54:54.085342 kernel: acpiphp: Slot [28] registered May 8 07:54:54.085351 kernel: acpiphp: Slot [29] registered May 8 07:54:54.085360 kernel: acpiphp: Slot [30] registered May 8 07:54:54.085369 kernel: acpiphp: Slot [31] registered May 8 07:54:54.085377 kernel: PCI host bridge to bus 0000:00 May 8 07:54:54.085475 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 8 07:54:54.085562 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 8 07:54:54.085645 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 8 07:54:54.085728 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] May 8 07:54:54.085810 kernel: pci_bus 0000:00: root bus resource [mem 0xc000000000-0xc07fffffff window] May 8 07:54:54.085891 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 8 07:54:54.087039 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 May 8 07:54:54.087164 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 May 8 07:54:54.087274 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 May 8 07:54:54.087377 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc120-0xc12f] May 8 07:54:54.087476 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] May 8 07:54:54.087575 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] May 8 07:54:54.087675 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] May 8 07:54:54.087780 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] May 8 07:54:54.087888 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 May 8 07:54:54.088387 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI May 8 07:54:54.088492 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB May 8 07:54:54.088598 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 May 8 07:54:54.088693 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] May 8 07:54:54.088787 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xc000000000-0xc000003fff 64bit pref] May 8 07:54:54.088885 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfeb90000-0xfeb90fff] May 8 07:54:54.089057 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfeb80000-0xfeb8ffff pref] May 8 07:54:54.089152 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 8 07:54:54.089255 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 May 8 07:54:54.089347 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc080-0xc0bf] May 8 07:54:54.089439 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfeb91000-0xfeb91fff] May 8 07:54:54.089537 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xc000004000-0xc000007fff 64bit pref] May 8 07:54:54.089629 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfeb00000-0xfeb7ffff pref] May 8 07:54:54.089727 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 May 8 07:54:54.089821 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] May 8 07:54:54.089928 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfeb92000-0xfeb92fff] May 8 07:54:54.090042 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xc000008000-0xc00000bfff 64bit pref] May 8 07:54:54.090142 kernel: pci 0000:00:05.0: [1af4:1002] type 00 class 0x00ff00 May 8 07:54:54.090242 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc0c0-0xc0ff] May 8 07:54:54.090347 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xc00000c000-0xc00000ffff 64bit pref] May 8 07:54:54.090454 kernel: pci 0000:00:06.0: [1af4:1005] type 00 class 0x00ff00 May 8 07:54:54.090553 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc100-0xc11f] May 8 07:54:54.090652 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfeb93000-0xfeb93fff] May 8 07:54:54.090752 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xc000010000-0xc000013fff 64bit pref] May 8 07:54:54.090767 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 8 07:54:54.090777 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 8 07:54:54.090791 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 8 07:54:54.090801 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 8 07:54:54.090811 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 8 07:54:54.090820 kernel: iommu: Default domain type: Translated May 8 07:54:54.090830 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 8 07:54:54.090840 kernel: PCI: Using ACPI for IRQ routing May 8 07:54:54.090850 kernel: PCI: pci_cache_line_size set to 64 bytes May 8 07:54:54.090859 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] May 8 07:54:54.090869 kernel: e820: reserve RAM buffer [mem 0xbffdd000-0xbfffffff] May 8 07:54:54.091018 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device May 8 07:54:54.091117 kernel: pci 0000:00:02.0: vgaarb: bridge control possible May 8 07:54:54.091214 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 8 07:54:54.091229 kernel: vgaarb: loaded May 8 07:54:54.091239 kernel: clocksource: Switched to clocksource kvm-clock May 8 07:54:54.091249 kernel: VFS: Disk quotas dquot_6.6.0 May 8 07:54:54.091259 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 8 07:54:54.091268 kernel: pnp: PnP ACPI init May 8 07:54:54.091361 kernel: pnp 00:03: [dma 2] May 8 07:54:54.091382 kernel: pnp: PnP ACPI: found 5 devices May 8 07:54:54.091392 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 8 07:54:54.091402 kernel: NET: Registered PF_INET protocol family May 8 07:54:54.091412 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 8 07:54:54.091422 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 8 07:54:54.091431 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 8 07:54:54.091441 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 8 07:54:54.091451 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 8 07:54:54.091463 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 8 07:54:54.091473 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 07:54:54.091483 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 07:54:54.091493 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 8 07:54:54.091502 kernel: NET: Registered PF_XDP protocol family May 8 07:54:54.091586 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 8 07:54:54.091666 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 8 07:54:54.091745 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 8 07:54:54.091829 kernel: pci_bus 0000:00: resource 7 [mem 0xc0000000-0xfebfffff window] May 8 07:54:54.091907 kernel: pci_bus 0000:00: resource 8 [mem 0xc000000000-0xc07fffffff window] May 8 07:54:54.092036 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release May 8 07:54:54.092130 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 8 07:54:54.092144 kernel: PCI: CLS 0 bytes, default 64 May 8 07:54:54.092153 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) May 8 07:54:54.092163 kernel: software IO TLB: mapped [mem 0x00000000bbfdd000-0x00000000bffdd000] (64MB) May 8 07:54:54.092172 kernel: Initialise system trusted keyrings May 8 07:54:54.092185 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 8 07:54:54.092194 kernel: Key type asymmetric registered May 8 07:54:54.092203 kernel: Asymmetric key parser 'x509' registered May 8 07:54:54.092212 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 8 07:54:54.092221 kernel: io scheduler mq-deadline registered May 8 07:54:54.092230 kernel: io scheduler kyber registered May 8 07:54:54.092239 kernel: io scheduler bfq registered May 8 07:54:54.092248 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 8 07:54:54.092257 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 May 8 07:54:54.092269 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 8 07:54:54.092278 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 8 07:54:54.092287 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 8 07:54:54.092296 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 8 07:54:54.092305 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 8 07:54:54.092314 kernel: random: crng init done May 8 07:54:54.092323 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 8 07:54:54.092332 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 8 07:54:54.092341 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 8 07:54:54.092432 kernel: rtc_cmos 00:04: RTC can wake from S4 May 8 07:54:54.092447 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 8 07:54:54.092527 kernel: rtc_cmos 00:04: registered as rtc0 May 8 07:54:54.092611 kernel: rtc_cmos 00:04: setting system clock to 2025-05-08T07:54:53 UTC (1746690893) May 8 07:54:54.092696 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram May 8 07:54:54.092709 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled May 8 07:54:54.092718 kernel: NET: Registered PF_INET6 protocol family May 8 07:54:54.092727 kernel: Segment Routing with IPv6 May 8 07:54:54.092740 kernel: In-situ OAM (IOAM) with IPv6 May 8 07:54:54.092749 kernel: NET: Registered PF_PACKET protocol family May 8 07:54:54.092758 kernel: Key type dns_resolver registered May 8 07:54:54.092767 kernel: IPI shorthand broadcast: enabled May 8 07:54:54.092776 kernel: sched_clock: Marking stable (3537007092, 171650068)->(3741526021, -32868861) May 8 07:54:54.092785 kernel: registered taskstats version 1 May 8 07:54:54.092794 kernel: Loading compiled-in X.509 certificates May 8 07:54:54.092803 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: 1d65a3ec8f410b8df3c709e9a079b9ae6c6b7ac3' May 8 07:54:54.092812 kernel: Key type .fscrypt registered May 8 07:54:54.092823 kernel: Key type fscrypt-provisioning registered May 8 07:54:54.092832 kernel: ima: No TPM chip found, activating TPM-bypass! May 8 07:54:54.092841 kernel: ima: Allocated hash algorithm: sha1 May 8 07:54:54.092850 kernel: ima: No architecture policies found May 8 07:54:54.092859 kernel: clk: Disabling unused clocks May 8 07:54:54.092868 kernel: Warning: unable to open an initial console. May 8 07:54:54.092877 kernel: Freeing unused kernel image (initmem) memory: 53684K May 8 07:54:54.092887 kernel: Write protecting the kernel read-only data: 24576k May 8 07:54:54.092896 kernel: Freeing unused kernel image (rodata/data gap) memory: 1200K May 8 07:54:54.092906 kernel: Run /init as init process May 8 07:54:54.097601 kernel: with arguments: May 8 07:54:54.097623 kernel: /init May 8 07:54:54.097633 kernel: with environment: May 8 07:54:54.097643 kernel: HOME=/ May 8 07:54:54.097652 kernel: TERM=linux May 8 07:54:54.097661 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 8 07:54:54.097672 systemd[1]: Successfully made /usr/ read-only. May 8 07:54:54.097690 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 8 07:54:54.097701 systemd[1]: Detected virtualization kvm. May 8 07:54:54.097710 systemd[1]: Detected architecture x86-64. May 8 07:54:54.097720 systemd[1]: Running in initrd. May 8 07:54:54.097729 systemd[1]: No hostname configured, using default hostname. May 8 07:54:54.097739 systemd[1]: Hostname set to . May 8 07:54:54.097749 systemd[1]: Initializing machine ID from VM UUID. May 8 07:54:54.097761 systemd[1]: Queued start job for default target initrd.target. May 8 07:54:54.097779 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 07:54:54.097791 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 07:54:54.097801 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 8 07:54:54.097812 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 8 07:54:54.097822 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 8 07:54:54.097835 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 8 07:54:54.097846 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 8 07:54:54.097857 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 8 07:54:54.097867 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 07:54:54.097877 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 8 07:54:54.097887 systemd[1]: Reached target paths.target - Path Units. May 8 07:54:54.097897 systemd[1]: Reached target slices.target - Slice Units. May 8 07:54:54.097909 systemd[1]: Reached target swap.target - Swaps. May 8 07:54:54.097944 systemd[1]: Reached target timers.target - Timer Units. May 8 07:54:54.097955 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 8 07:54:54.097965 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 8 07:54:54.097975 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 8 07:54:54.097985 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 8 07:54:54.097995 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 8 07:54:54.098005 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 8 07:54:54.098016 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 8 07:54:54.098029 systemd[1]: Reached target sockets.target - Socket Units. May 8 07:54:54.098039 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 8 07:54:54.098049 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 8 07:54:54.098059 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 8 07:54:54.098070 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 8 07:54:54.098080 systemd[1]: Starting systemd-fsck-usr.service... May 8 07:54:54.098090 systemd[1]: Starting systemd-journald.service - Journal Service... May 8 07:54:54.098100 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 8 07:54:54.098111 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:54:54.098122 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 8 07:54:54.098134 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 8 07:54:54.098146 systemd[1]: Finished systemd-fsck-usr.service. May 8 07:54:54.098184 systemd-journald[187]: Collecting audit messages is disabled. May 8 07:54:54.098210 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 8 07:54:54.098223 systemd-journald[187]: Journal started May 8 07:54:54.098246 systemd-journald[187]: Runtime Journal (/run/log/journal/56acdddfb02e4db59b117a80d1f2085b) is 8M, max 78.5M, 70.5M free. May 8 07:54:54.098496 systemd-modules-load[189]: Inserted module 'overlay' May 8 07:54:54.101006 systemd[1]: Started systemd-journald.service - Journal Service. May 8 07:54:54.103294 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 8 07:54:54.150198 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 8 07:54:54.150221 kernel: Bridge firewalling registered May 8 07:54:54.128830 systemd-modules-load[189]: Inserted module 'br_netfilter' May 8 07:54:54.148256 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 8 07:54:54.150234 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:54:54.153200 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 8 07:54:54.155037 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 07:54:54.158062 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 8 07:54:54.169498 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 8 07:54:54.180228 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 07:54:54.182176 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 07:54:54.185963 systemd-tmpfiles[208]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 8 07:54:54.190125 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 07:54:54.193143 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 8 07:54:54.196089 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 8 07:54:54.199050 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 8 07:54:54.216407 dracut-cmdline[225]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=c6e5fdf442cda427c9e51c476b5f39b01a7265d89660ab7c7d9178b52b2cc04b May 8 07:54:54.237867 systemd-resolved[226]: Positive Trust Anchors: May 8 07:54:54.237882 systemd-resolved[226]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 07:54:54.237941 systemd-resolved[226]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 8 07:54:54.240807 systemd-resolved[226]: Defaulting to hostname 'linux'. May 8 07:54:54.241650 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 8 07:54:54.243027 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 8 07:54:54.299971 kernel: SCSI subsystem initialized May 8 07:54:54.309965 kernel: Loading iSCSI transport class v2.0-870. May 8 07:54:54.322218 kernel: iscsi: registered transport (tcp) May 8 07:54:54.344065 kernel: iscsi: registered transport (qla4xxx) May 8 07:54:54.344136 kernel: QLogic iSCSI HBA Driver May 8 07:54:54.367473 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 8 07:54:54.391657 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 8 07:54:54.393973 systemd[1]: Reached target network-pre.target - Preparation for Network. May 8 07:54:54.475524 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 8 07:54:54.480196 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 8 07:54:54.565021 kernel: raid6: sse2x4 gen() 7376 MB/s May 8 07:54:54.583011 kernel: raid6: sse2x2 gen() 15136 MB/s May 8 07:54:54.601326 kernel: raid6: sse2x1 gen() 10009 MB/s May 8 07:54:54.601437 kernel: raid6: using algorithm sse2x2 gen() 15136 MB/s May 8 07:54:54.620396 kernel: raid6: .... xor() 9463 MB/s, rmw enabled May 8 07:54:54.620467 kernel: raid6: using ssse3x2 recovery algorithm May 8 07:54:54.643280 kernel: xor: measuring software checksum speed May 8 07:54:54.643350 kernel: prefetch64-sse : 18485 MB/sec May 8 07:54:54.644568 kernel: generic_sse : 16858 MB/sec May 8 07:54:54.644629 kernel: xor: using function: prefetch64-sse (18485 MB/sec) May 8 07:54:54.817970 kernel: Btrfs loaded, zoned=no, fsverity=no May 8 07:54:54.827076 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 8 07:54:54.833620 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 07:54:54.858295 systemd-udevd[436]: Using default interface naming scheme 'v255'. May 8 07:54:54.872459 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 07:54:54.881295 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 8 07:54:54.922348 dracut-pre-trigger[447]: rd.md=0: removing MD RAID activation May 8 07:54:54.973836 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 8 07:54:54.979102 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 8 07:54:55.070190 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 8 07:54:55.079141 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 8 07:54:55.161936 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 May 8 07:54:55.179940 kernel: virtio_blk virtio2: 2/0/0 default/read/poll queues May 8 07:54:55.193105 kernel: virtio_blk virtio2: [vda] 20971520 512-byte logical blocks (10.7 GB/10.0 GiB) May 8 07:54:55.193221 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 8 07:54:55.193235 kernel: GPT:17805311 != 20971519 May 8 07:54:55.193247 kernel: GPT:Alternate GPT header not at the end of the disk. May 8 07:54:55.193258 kernel: GPT:17805311 != 20971519 May 8 07:54:55.193269 kernel: GPT: Use GNU Parted to correct GPT errors. May 8 07:54:55.193280 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 07:54:55.188792 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 07:54:55.188942 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:54:55.189581 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:54:55.207333 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:54:55.208185 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 8 07:54:55.226933 kernel: libata version 3.00 loaded. May 8 07:54:55.229978 kernel: ata_piix 0000:00:01.1: version 2.13 May 8 07:54:55.255095 kernel: scsi host0: ata_piix May 8 07:54:55.255239 kernel: scsi host1: ata_piix May 8 07:54:55.255350 kernel: BTRFS: device fsid 4222d142-b650-43a9-9f25-c45bb16dfbfe devid 1 transid 43 /dev/vda3 scanned by (udev-worker) (490) May 8 07:54:55.255364 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc120 irq 14 May 8 07:54:55.255376 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc128 irq 15 May 8 07:54:55.256934 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (489) May 8 07:54:55.278148 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 8 07:54:55.312287 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 8 07:54:55.313072 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:54:55.322742 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 8 07:54:55.323366 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 8 07:54:55.334863 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 8 07:54:55.336354 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 8 07:54:55.356269 disk-uuid[535]: Primary Header is updated. May 8 07:54:55.356269 disk-uuid[535]: Secondary Entries is updated. May 8 07:54:55.356269 disk-uuid[535]: Secondary Header is updated. May 8 07:54:55.366976 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 07:54:55.514754 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 8 07:54:55.527550 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 8 07:54:55.528140 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 07:54:55.529350 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 8 07:54:55.531266 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 8 07:54:55.547130 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 8 07:54:56.383112 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 07:54:56.384875 disk-uuid[536]: The operation has completed successfully. May 8 07:54:56.483012 systemd[1]: disk-uuid.service: Deactivated successfully. May 8 07:54:56.483875 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 8 07:54:56.529028 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 8 07:54:56.547014 sh[560]: Success May 8 07:54:56.572708 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 8 07:54:56.572743 kernel: device-mapper: uevent: version 1.0.3 May 8 07:54:56.573836 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 8 07:54:56.587008 kernel: device-mapper: verity: sha256 using implementation "sha256-ssse3" May 8 07:54:56.659458 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 8 07:54:56.661978 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 8 07:54:56.681949 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 8 07:54:56.693328 kernel: BTRFS info (device dm-0): first mount of filesystem 4222d142-b650-43a9-9f25-c45bb16dfbfe May 8 07:54:56.693417 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 8 07:54:56.695349 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 8 07:54:56.697303 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 8 07:54:56.699739 kernel: BTRFS info (device dm-0): using free space tree May 8 07:54:56.712321 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 8 07:54:56.713224 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 8 07:54:56.714717 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 8 07:54:56.718025 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 8 07:54:56.721051 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 8 07:54:56.755490 kernel: BTRFS info (device vda6): first mount of filesystem cf994bdc-a9fa-440e-8b4d-4cf650650e1d May 8 07:54:56.755590 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 8 07:54:56.755623 kernel: BTRFS info (device vda6): using free space tree May 8 07:54:56.763005 kernel: BTRFS info (device vda6): auto enabling async discard May 8 07:54:56.769998 kernel: BTRFS info (device vda6): last unmount of filesystem cf994bdc-a9fa-440e-8b4d-4cf650650e1d May 8 07:54:56.778669 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 8 07:54:56.781903 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 8 07:54:56.874301 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 8 07:54:56.878598 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 8 07:54:56.923281 systemd-networkd[746]: lo: Link UP May 8 07:54:56.923292 systemd-networkd[746]: lo: Gained carrier May 8 07:54:56.924426 systemd-networkd[746]: Enumeration completed May 8 07:54:56.925308 systemd-networkd[746]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 07:54:56.925312 systemd-networkd[746]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 07:54:56.926092 systemd[1]: Started systemd-networkd.service - Network Configuration. May 8 07:54:56.926293 systemd-networkd[746]: eth0: Link UP May 8 07:54:56.926296 systemd-networkd[746]: eth0: Gained carrier May 8 07:54:56.926305 systemd-networkd[746]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 07:54:56.931578 systemd[1]: Reached target network.target - Network. May 8 07:54:56.941673 systemd-networkd[746]: eth0: DHCPv4 address 172.24.4.249/24, gateway 172.24.4.1 acquired from 172.24.4.1 May 8 07:54:56.996582 ignition[668]: Ignition 2.21.0 May 8 07:54:56.996601 ignition[668]: Stage: fetch-offline May 8 07:54:56.996636 ignition[668]: no configs at "/usr/lib/ignition/base.d" May 8 07:54:56.996646 ignition[668]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:54:56.999149 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 8 07:54:56.996747 ignition[668]: parsed url from cmdline: "" May 8 07:54:56.996751 ignition[668]: no config URL provided May 8 07:54:56.996756 ignition[668]: reading system config file "/usr/lib/ignition/user.ign" May 8 07:54:57.001029 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 8 07:54:56.996765 ignition[668]: no config at "/usr/lib/ignition/user.ign" May 8 07:54:56.996770 ignition[668]: failed to fetch config: resource requires networking May 8 07:54:56.997465 ignition[668]: Ignition finished successfully May 8 07:54:57.023395 ignition[756]: Ignition 2.21.0 May 8 07:54:57.023407 ignition[756]: Stage: fetch May 8 07:54:57.025489 ignition[756]: no configs at "/usr/lib/ignition/base.d" May 8 07:54:57.025510 ignition[756]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:54:57.025638 ignition[756]: parsed url from cmdline: "" May 8 07:54:57.025647 ignition[756]: no config URL provided May 8 07:54:57.025652 ignition[756]: reading system config file "/usr/lib/ignition/user.ign" May 8 07:54:57.025661 ignition[756]: no config at "/usr/lib/ignition/user.ign" May 8 07:54:57.025831 ignition[756]: GET http://169.254.169.254/openstack/latest/user_data: attempt #1 May 8 07:54:57.027356 ignition[756]: config drive ("/dev/disk/by-label/config-2") not found. Waiting... May 8 07:54:57.027379 ignition[756]: config drive ("/dev/disk/by-label/CONFIG-2") not found. Waiting... May 8 07:54:57.293897 ignition[756]: GET result: OK May 8 07:54:57.294124 ignition[756]: parsing config with SHA512: 3a892b7ca63e70ab0168d9c0c5ce952a9af5d459b1ea15f32dfe7d0b9165bf4df9bdf57e5dcdca21ffc953f26d118dc44fdc6906129001de480414bb8540b593 May 8 07:54:57.304684 unknown[756]: fetched base config from "system" May 8 07:54:57.304702 unknown[756]: fetched base config from "system" May 8 07:54:57.305522 ignition[756]: fetch: fetch complete May 8 07:54:57.304716 unknown[756]: fetched user config from "openstack" May 8 07:54:57.305534 ignition[756]: fetch: fetch passed May 8 07:54:57.311111 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 8 07:54:57.305614 ignition[756]: Ignition finished successfully May 8 07:54:57.315244 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 8 07:54:57.369743 ignition[762]: Ignition 2.21.0 May 8 07:54:57.369775 ignition[762]: Stage: kargs May 8 07:54:57.370277 ignition[762]: no configs at "/usr/lib/ignition/base.d" May 8 07:54:57.370347 ignition[762]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:54:57.376414 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 8 07:54:57.373126 ignition[762]: kargs: kargs passed May 8 07:54:57.380167 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 8 07:54:57.373235 ignition[762]: Ignition finished successfully May 8 07:54:57.423564 ignition[768]: Ignition 2.21.0 May 8 07:54:57.423597 ignition[768]: Stage: disks May 8 07:54:57.423972 ignition[768]: no configs at "/usr/lib/ignition/base.d" May 8 07:54:57.424002 ignition[768]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:54:57.426183 ignition[768]: disks: disks passed May 8 07:54:57.428650 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 8 07:54:57.426275 ignition[768]: Ignition finished successfully May 8 07:54:57.431672 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 8 07:54:57.433575 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 8 07:54:57.436190 systemd[1]: Reached target local-fs.target - Local File Systems. May 8 07:54:57.438711 systemd[1]: Reached target sysinit.target - System Initialization. May 8 07:54:57.441596 systemd[1]: Reached target basic.target - Basic System. May 8 07:54:57.448136 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 8 07:54:57.496804 systemd-fsck[776]: ROOT: clean, 15/1628000 files, 120826/1617920 blocks May 8 07:54:57.508672 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 8 07:54:57.513199 systemd[1]: Mounting sysroot.mount - /sysroot... May 8 07:54:57.716978 kernel: EXT4-fs (vda9): mounted filesystem a930b30e-9e26-4ff0-8cba-060d78a760a1 r/w with ordered data mode. Quota mode: none. May 8 07:54:57.717351 systemd[1]: Mounted sysroot.mount - /sysroot. May 8 07:54:57.719166 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 8 07:54:57.724011 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 8 07:54:57.730081 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 8 07:54:57.737797 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 8 07:54:57.744140 systemd[1]: Starting flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent... May 8 07:54:57.747611 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 8 07:54:57.749682 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 8 07:54:57.757315 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 8 07:54:57.763166 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 8 07:54:57.794973 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (784) May 8 07:54:57.803835 kernel: BTRFS info (device vda6): first mount of filesystem cf994bdc-a9fa-440e-8b4d-4cf650650e1d May 8 07:54:57.803863 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 8 07:54:57.803882 kernel: BTRFS info (device vda6): using free space tree May 8 07:54:57.813937 kernel: BTRFS info (device vda6): auto enabling async discard May 8 07:54:57.823700 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 8 07:54:57.894434 initrd-setup-root[812]: cut: /sysroot/etc/passwd: No such file or directory May 8 07:54:57.902876 initrd-setup-root[819]: cut: /sysroot/etc/group: No such file or directory May 8 07:54:57.912488 initrd-setup-root[826]: cut: /sysroot/etc/shadow: No such file or directory May 8 07:54:57.920141 initrd-setup-root[833]: cut: /sysroot/etc/gshadow: No such file or directory May 8 07:54:58.013617 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 8 07:54:58.015472 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 8 07:54:58.018063 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 8 07:54:58.029147 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 8 07:54:58.032644 kernel: BTRFS info (device vda6): last unmount of filesystem cf994bdc-a9fa-440e-8b4d-4cf650650e1d May 8 07:54:58.053188 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 8 07:54:58.072732 ignition[900]: INFO : Ignition 2.21.0 May 8 07:54:58.072732 ignition[900]: INFO : Stage: mount May 8 07:54:58.073854 ignition[900]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 07:54:58.073854 ignition[900]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:54:58.075304 ignition[900]: INFO : mount: mount passed May 8 07:54:58.075304 ignition[900]: INFO : Ignition finished successfully May 8 07:54:58.077113 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 8 07:54:58.724107 systemd-networkd[746]: eth0: Gained IPv6LL May 8 07:55:04.958894 coreos-metadata[786]: May 08 07:55:04.958 WARN failed to locate config-drive, using the metadata service API instead May 8 07:55:05.001671 coreos-metadata[786]: May 08 07:55:05.001 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 May 8 07:55:05.016974 coreos-metadata[786]: May 08 07:55:05.016 INFO Fetch successful May 8 07:55:05.017684 coreos-metadata[786]: May 08 07:55:05.017 INFO wrote hostname ci-4327-0-0-w-45c5ab2731.novalocal to /sysroot/etc/hostname May 8 07:55:05.021159 systemd[1]: flatcar-openstack-hostname.service: Deactivated successfully. May 8 07:55:05.021274 systemd[1]: Finished flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent. May 8 07:55:05.025012 systemd[1]: Starting ignition-files.service - Ignition (files)... May 8 07:55:05.048508 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 8 07:55:05.079021 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (918) May 8 07:55:05.085650 kernel: BTRFS info (device vda6): first mount of filesystem cf994bdc-a9fa-440e-8b4d-4cf650650e1d May 8 07:55:05.085722 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 8 07:55:05.089639 kernel: BTRFS info (device vda6): using free space tree May 8 07:55:05.100041 kernel: BTRFS info (device vda6): auto enabling async discard May 8 07:55:05.105464 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 8 07:55:05.154596 ignition[935]: INFO : Ignition 2.21.0 May 8 07:55:05.154596 ignition[935]: INFO : Stage: files May 8 07:55:05.157147 ignition[935]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 07:55:05.157147 ignition[935]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:55:05.157147 ignition[935]: DEBUG : files: compiled without relabeling support, skipping May 8 07:55:05.162218 ignition[935]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 8 07:55:05.162218 ignition[935]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 8 07:55:05.162218 ignition[935]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 8 07:55:05.170001 ignition[935]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 8 07:55:05.170001 ignition[935]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 8 07:55:05.170001 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 8 07:55:05.170001 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 8 07:55:05.163273 unknown[935]: wrote ssh authorized keys file for user: core May 8 07:55:05.256294 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 8 07:55:05.564350 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 8 07:55:05.564350 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 8 07:55:05.564350 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 May 8 07:55:06.133027 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 8 07:55:06.835890 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 8 07:55:06.835890 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 07:55:06.840788 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 May 8 07:55:07.405877 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 8 07:55:09.516981 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 07:55:09.518327 ignition[935]: INFO : files: op(c): [started] processing unit "prepare-helm.service" May 8 07:55:09.519771 ignition[935]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 07:55:09.520748 ignition[935]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 07:55:09.520748 ignition[935]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" May 8 07:55:09.520748 ignition[935]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" May 8 07:55:09.523148 ignition[935]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" May 8 07:55:09.523148 ignition[935]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" May 8 07:55:09.523148 ignition[935]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" May 8 07:55:09.523148 ignition[935]: INFO : files: files passed May 8 07:55:09.523148 ignition[935]: INFO : Ignition finished successfully May 8 07:55:09.524104 systemd[1]: Finished ignition-files.service - Ignition (files). May 8 07:55:09.529647 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 8 07:55:09.534068 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 8 07:55:09.541011 systemd[1]: ignition-quench.service: Deactivated successfully. May 8 07:55:09.544068 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 8 07:55:09.557835 initrd-setup-root-after-ignition[969]: grep: May 8 07:55:09.557835 initrd-setup-root-after-ignition[965]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 8 07:55:09.563342 initrd-setup-root-after-ignition[965]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 8 07:55:09.565174 initrd-setup-root-after-ignition[969]: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 8 07:55:09.564069 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 8 07:55:09.565240 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 8 07:55:09.569024 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 8 07:55:09.617702 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 8 07:55:09.617902 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 8 07:55:09.621404 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 8 07:55:09.623230 systemd[1]: Reached target initrd.target - Initrd Default Target. May 8 07:55:09.625350 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 8 07:55:09.628019 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 8 07:55:09.654414 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 8 07:55:09.658327 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 8 07:55:09.689472 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 8 07:55:09.691188 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 07:55:09.693844 systemd[1]: Stopped target timers.target - Timer Units. May 8 07:55:09.696732 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 8 07:55:09.697123 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 8 07:55:09.700114 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 8 07:55:09.701876 systemd[1]: Stopped target basic.target - Basic System. May 8 07:55:09.704699 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 8 07:55:09.707286 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 8 07:55:09.709822 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 8 07:55:09.712744 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 8 07:55:09.715726 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 8 07:55:09.718611 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 8 07:55:09.721586 systemd[1]: Stopped target sysinit.target - System Initialization. May 8 07:55:09.724436 systemd[1]: Stopped target local-fs.target - Local File Systems. May 8 07:55:09.727360 systemd[1]: Stopped target swap.target - Swaps. May 8 07:55:09.729967 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 8 07:55:09.730252 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 8 07:55:09.733307 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 8 07:55:09.735282 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 07:55:09.737746 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 8 07:55:09.739042 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 07:55:09.740600 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 8 07:55:09.740880 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 8 07:55:09.744823 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 8 07:55:09.745206 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 8 07:55:09.748153 systemd[1]: ignition-files.service: Deactivated successfully. May 8 07:55:09.748423 systemd[1]: Stopped ignition-files.service - Ignition (files). May 8 07:55:09.753439 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 8 07:55:09.757613 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 8 07:55:09.758033 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 8 07:55:09.765504 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 8 07:55:09.768182 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 8 07:55:09.768525 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 8 07:55:09.771905 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 8 07:55:09.772184 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 8 07:55:09.783994 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 8 07:55:09.784078 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 8 07:55:09.801570 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 8 07:55:09.805955 systemd[1]: sysroot-boot.service: Deactivated successfully. May 8 07:55:09.806068 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 8 07:55:09.812395 ignition[989]: INFO : Ignition 2.21.0 May 8 07:55:09.812395 ignition[989]: INFO : Stage: umount May 8 07:55:09.812395 ignition[989]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 07:55:09.812395 ignition[989]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" May 8 07:55:09.812395 ignition[989]: INFO : umount: umount passed May 8 07:55:09.812395 ignition[989]: INFO : Ignition finished successfully May 8 07:55:09.815358 systemd[1]: ignition-mount.service: Deactivated successfully. May 8 07:55:09.815602 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 8 07:55:09.816423 systemd[1]: ignition-disks.service: Deactivated successfully. May 8 07:55:09.816468 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 8 07:55:09.817229 systemd[1]: ignition-kargs.service: Deactivated successfully. May 8 07:55:09.817270 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 8 07:55:09.818205 systemd[1]: ignition-fetch.service: Deactivated successfully. May 8 07:55:09.818245 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 8 07:55:09.819239 systemd[1]: Stopped target network.target - Network. May 8 07:55:09.820207 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 8 07:55:09.820251 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 8 07:55:09.821246 systemd[1]: Stopped target paths.target - Path Units. May 8 07:55:09.822193 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 8 07:55:09.823992 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 07:55:09.824784 systemd[1]: Stopped target slices.target - Slice Units. May 8 07:55:09.825725 systemd[1]: Stopped target sockets.target - Socket Units. May 8 07:55:09.826854 systemd[1]: iscsid.socket: Deactivated successfully. May 8 07:55:09.826889 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 8 07:55:09.828005 systemd[1]: iscsiuio.socket: Deactivated successfully. May 8 07:55:09.828035 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 8 07:55:09.828973 systemd[1]: ignition-setup.service: Deactivated successfully. May 8 07:55:09.829014 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 8 07:55:09.830104 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 8 07:55:09.830144 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 8 07:55:09.831302 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 8 07:55:09.831342 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 8 07:55:09.832324 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 8 07:55:09.833375 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 8 07:55:09.838957 systemd[1]: systemd-resolved.service: Deactivated successfully. May 8 07:55:09.839068 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 8 07:55:09.846565 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 8 07:55:09.846802 systemd[1]: systemd-networkd.service: Deactivated successfully. May 8 07:55:09.846900 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 8 07:55:09.848892 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 8 07:55:09.849338 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 8 07:55:09.850474 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 8 07:55:09.850528 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 8 07:55:09.853009 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 8 07:55:09.854034 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 8 07:55:09.854738 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 8 07:55:09.855352 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 8 07:55:09.855393 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 8 07:55:09.856464 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 8 07:55:09.856504 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 8 07:55:09.857199 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 8 07:55:09.857238 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 07:55:09.858664 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 07:55:09.860354 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 8 07:55:09.860413 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 8 07:55:09.872375 systemd[1]: systemd-udevd.service: Deactivated successfully. May 8 07:55:09.873107 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 07:55:09.874671 systemd[1]: network-cleanup.service: Deactivated successfully. May 8 07:55:09.874755 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 8 07:55:09.875873 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 8 07:55:09.876183 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 8 07:55:09.876904 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 8 07:55:09.876980 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 8 07:55:09.878164 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 8 07:55:09.878212 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 8 07:55:09.879751 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 8 07:55:09.879792 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 8 07:55:09.880889 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 8 07:55:09.880945 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 8 07:55:09.884012 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 8 07:55:09.884898 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 8 07:55:09.884968 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 8 07:55:09.886676 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 8 07:55:09.886719 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 07:55:09.889139 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 07:55:09.889183 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:55:09.891879 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. May 8 07:55:09.891950 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 8 07:55:09.892019 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 8 07:55:09.897761 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 8 07:55:09.897868 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 8 07:55:09.899232 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 8 07:55:09.901159 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 8 07:55:09.919086 systemd[1]: Switching root. May 8 07:55:09.958721 systemd-journald[187]: Journal stopped May 8 07:55:11.627264 systemd-journald[187]: Received SIGTERM from PID 1 (systemd). May 8 07:55:11.627331 kernel: SELinux: policy capability network_peer_controls=1 May 8 07:55:11.627350 kernel: SELinux: policy capability open_perms=1 May 8 07:55:11.627362 kernel: SELinux: policy capability extended_socket_class=1 May 8 07:55:11.627379 kernel: SELinux: policy capability always_check_network=0 May 8 07:55:11.627395 kernel: SELinux: policy capability cgroup_seclabel=1 May 8 07:55:11.627407 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 8 07:55:11.627419 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 8 07:55:11.627435 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 8 07:55:11.627447 kernel: audit: type=1403 audit(1746690910.570:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 8 07:55:11.627460 systemd[1]: Successfully loaded SELinux policy in 78.740ms. May 8 07:55:11.627480 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 26.797ms. May 8 07:55:11.627495 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 8 07:55:11.627509 systemd[1]: Detected virtualization kvm. May 8 07:55:11.627522 systemd[1]: Detected architecture x86-64. May 8 07:55:11.627535 systemd[1]: Detected first boot. May 8 07:55:11.627550 systemd[1]: Hostname set to . May 8 07:55:11.627564 systemd[1]: Initializing machine ID from VM UUID. May 8 07:55:11.627577 zram_generator::config[1033]: No configuration found. May 8 07:55:11.627592 kernel: Guest personality initialized and is inactive May 8 07:55:11.627604 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 8 07:55:11.627616 kernel: Initialized host personality May 8 07:55:11.627628 kernel: NET: Registered PF_VSOCK protocol family May 8 07:55:11.627640 systemd[1]: Populated /etc with preset unit settings. May 8 07:55:11.627655 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 8 07:55:11.627670 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 8 07:55:11.627682 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 8 07:55:11.627694 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 8 07:55:11.627707 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 8 07:55:11.627719 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 8 07:55:11.627731 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 8 07:55:11.627743 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 8 07:55:11.627755 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 8 07:55:11.627769 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 8 07:55:11.627782 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 8 07:55:11.627794 systemd[1]: Created slice user.slice - User and Session Slice. May 8 07:55:11.627806 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 8 07:55:11.627819 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 8 07:55:11.627831 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 8 07:55:11.627843 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 8 07:55:11.627856 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 8 07:55:11.627870 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 8 07:55:11.627883 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 8 07:55:11.627895 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 8 07:55:11.627907 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 8 07:55:11.627957 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 8 07:55:11.627972 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 8 07:55:11.627984 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 8 07:55:11.628000 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 8 07:55:11.628013 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 8 07:55:11.628025 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 8 07:55:11.628037 systemd[1]: Reached target slices.target - Slice Units. May 8 07:55:11.628049 systemd[1]: Reached target swap.target - Swaps. May 8 07:55:11.628065 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 8 07:55:11.628077 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 8 07:55:11.628090 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 8 07:55:11.628102 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 8 07:55:11.628114 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 8 07:55:11.628128 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 8 07:55:11.628141 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 8 07:55:11.628152 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 8 07:55:11.628164 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 8 07:55:11.628176 systemd[1]: Mounting media.mount - External Media Directory... May 8 07:55:11.628189 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:11.628201 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 8 07:55:11.628214 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 8 07:55:11.628226 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 8 07:55:11.628240 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 8 07:55:11.628253 systemd[1]: Reached target machines.target - Containers. May 8 07:55:11.628265 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 8 07:55:11.628278 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 07:55:11.628290 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 8 07:55:11.628302 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 8 07:55:11.628314 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 07:55:11.628326 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 8 07:55:11.628341 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 07:55:11.628353 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 8 07:55:11.628364 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 07:55:11.628377 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 8 07:55:11.628389 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 8 07:55:11.628401 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 8 07:55:11.628413 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 8 07:55:11.628425 systemd[1]: Stopped systemd-fsck-usr.service. May 8 07:55:11.628440 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 07:55:11.628453 systemd[1]: Starting systemd-journald.service - Journal Service... May 8 07:55:11.628465 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 8 07:55:11.628477 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 8 07:55:11.628489 kernel: loop: module loaded May 8 07:55:11.628501 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 8 07:55:11.628513 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 8 07:55:11.628528 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 8 07:55:11.628540 systemd[1]: verity-setup.service: Deactivated successfully. May 8 07:55:11.628552 systemd[1]: Stopped verity-setup.service. May 8 07:55:11.628565 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:11.628580 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 8 07:55:11.628593 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 8 07:55:11.628605 systemd[1]: Mounted media.mount - External Media Directory. May 8 07:55:11.628620 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 8 07:55:11.628633 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 8 07:55:11.628645 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 8 07:55:11.628659 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 8 07:55:11.628672 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 8 07:55:11.628687 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 8 07:55:11.628700 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 07:55:11.628714 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 07:55:11.628727 kernel: fuse: init (API version 7.39) May 8 07:55:11.628739 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 07:55:11.628752 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 07:55:11.628765 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 8 07:55:11.628778 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 8 07:55:11.628791 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 07:55:11.628806 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 07:55:11.628842 systemd-journald[1116]: Collecting audit messages is disabled. May 8 07:55:11.628869 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 8 07:55:11.628884 systemd-journald[1116]: Journal started May 8 07:55:11.628910 systemd-journald[1116]: Runtime Journal (/run/log/journal/56acdddfb02e4db59b117a80d1f2085b) is 8M, max 78.5M, 70.5M free. May 8 07:55:11.630053 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 8 07:55:11.276105 systemd[1]: Queued start job for default target multi-user.target. May 8 07:55:11.281870 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 8 07:55:11.282322 systemd[1]: systemd-journald.service: Deactivated successfully. May 8 07:55:11.635943 systemd[1]: Started systemd-journald.service - Journal Service. May 8 07:55:11.638164 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 8 07:55:11.655184 systemd[1]: Reached target network-pre.target - Preparation for Network. May 8 07:55:11.662972 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 8 07:55:11.668964 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 8 07:55:11.669566 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 8 07:55:11.669601 systemd[1]: Reached target local-fs.target - Local File Systems. May 8 07:55:11.673068 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 8 07:55:11.693015 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 8 07:55:11.693683 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 07:55:11.696829 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 8 07:55:11.696948 kernel: ACPI: bus type drm_connector registered May 8 07:55:11.698529 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 8 07:55:11.699117 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 07:55:11.703787 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 8 07:55:11.705000 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 8 07:55:11.706638 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 07:55:11.718012 systemd-journald[1116]: Time spent on flushing to /var/log/journal/56acdddfb02e4db59b117a80d1f2085b is 54.672ms for 962 entries. May 8 07:55:11.718012 systemd-journald[1116]: System Journal (/var/log/journal/56acdddfb02e4db59b117a80d1f2085b) is 8M, max 584.8M, 576.8M free. May 8 07:55:11.780758 systemd-journald[1116]: Received client request to flush runtime journal. May 8 07:55:11.717453 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 8 07:55:11.720838 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 8 07:55:11.721734 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 07:55:11.721883 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 8 07:55:11.786213 kernel: loop0: detected capacity change from 0 to 8 May 8 07:55:11.722730 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 8 07:55:11.727234 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 8 07:55:11.728772 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 8 07:55:11.729418 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 8 07:55:11.740063 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 8 07:55:11.781961 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 8 07:55:11.788967 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 07:55:11.794127 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 8 07:55:11.796340 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 8 07:55:11.801095 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 8 07:55:11.802034 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 8 07:55:11.830010 kernel: loop1: detected capacity change from 0 to 113872 May 8 07:55:11.856445 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 8 07:55:11.858198 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 8 07:55:11.876194 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 8 07:55:11.897291 systemd-tmpfiles[1187]: ACLs are not supported, ignoring. May 8 07:55:11.897309 systemd-tmpfiles[1187]: ACLs are not supported, ignoring. May 8 07:55:11.901716 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 8 07:55:11.904942 kernel: loop2: detected capacity change from 0 to 210664 May 8 07:55:11.961476 kernel: loop3: detected capacity change from 0 to 146240 May 8 07:55:12.023948 kernel: loop4: detected capacity change from 0 to 8 May 8 07:55:12.032147 kernel: loop5: detected capacity change from 0 to 113872 May 8 07:55:12.068933 kernel: loop6: detected capacity change from 0 to 210664 May 8 07:55:12.116941 kernel: loop7: detected capacity change from 0 to 146240 May 8 07:55:12.183445 (sd-merge)[1194]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-openstack'. May 8 07:55:12.184065 (sd-merge)[1194]: Merged extensions into '/usr'. May 8 07:55:12.189276 systemd[1]: Reload requested from client PID 1168 ('systemd-sysext') (unit systemd-sysext.service)... May 8 07:55:12.189292 systemd[1]: Reloading... May 8 07:55:12.268966 zram_generator::config[1217]: No configuration found. May 8 07:55:12.481555 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 07:55:12.605389 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 8 07:55:12.605685 systemd[1]: Reloading finished in 415 ms. May 8 07:55:12.625702 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 8 07:55:12.626590 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 8 07:55:12.635193 systemd[1]: Starting ensure-sysext.service... May 8 07:55:12.637532 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 8 07:55:12.643073 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 8 07:55:12.663146 ldconfig[1156]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 8 07:55:12.667010 systemd[1]: Reload requested from client PID 1276 ('systemctl') (unit ensure-sysext.service)... May 8 07:55:12.667026 systemd[1]: Reloading... May 8 07:55:12.690657 systemd-udevd[1278]: Using default interface naming scheme 'v255'. May 8 07:55:12.690683 systemd-tmpfiles[1277]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 8 07:55:12.690718 systemd-tmpfiles[1277]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 8 07:55:12.691506 systemd-tmpfiles[1277]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 8 07:55:12.691762 systemd-tmpfiles[1277]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 8 07:55:12.692688 systemd-tmpfiles[1277]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 8 07:55:12.693005 systemd-tmpfiles[1277]: ACLs are not supported, ignoring. May 8 07:55:12.693072 systemd-tmpfiles[1277]: ACLs are not supported, ignoring. May 8 07:55:12.701180 systemd-tmpfiles[1277]: Detected autofs mount point /boot during canonicalization of boot. May 8 07:55:12.701190 systemd-tmpfiles[1277]: Skipping /boot May 8 07:55:12.719656 systemd-tmpfiles[1277]: Detected autofs mount point /boot during canonicalization of boot. May 8 07:55:12.719670 systemd-tmpfiles[1277]: Skipping /boot May 8 07:55:12.746068 zram_generator::config[1302]: No configuration found. May 8 07:55:12.927940 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 43 scanned by (udev-worker) (1318) May 8 07:55:12.955057 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 07:55:13.006078 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 May 8 07:55:13.012585 kernel: mousedev: PS/2 mouse device common for all mice May 8 07:55:13.054939 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 May 8 07:55:13.066998 kernel: ACPI: button: Power Button [PWRF] May 8 07:55:13.081809 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 8 07:55:13.082995 systemd[1]: Reloading finished in 415 ms. May 8 07:55:13.094610 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 8 07:55:13.095972 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 8 07:55:13.105379 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 8 07:55:13.144100 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 8 07:55:13.147258 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.148480 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 8 07:55:13.152243 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 8 07:55:13.153724 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 07:55:13.157254 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 8 07:55:13.159180 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 07:55:13.164237 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 07:55:13.164899 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 07:55:13.169139 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 8 07:55:13.169681 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 07:55:13.170804 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 8 07:55:13.174510 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 8 07:55:13.180613 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 8 07:55:13.193238 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 8 07:55:13.194133 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.197547 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 07:55:13.198973 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 07:55:13.203469 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.203742 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 07:55:13.210596 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 8 07:55:13.212246 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 07:55:13.212380 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 07:55:13.217579 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 8 07:55:13.218215 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.220477 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 07:55:13.220970 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 07:55:13.231678 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.231989 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 8 07:55:13.246399 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 8 07:55:13.250172 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 8 07:55:13.250860 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 8 07:55:13.251016 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 8 07:55:13.251197 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 07:55:13.253293 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 07:55:13.253838 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 8 07:55:13.255363 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 8 07:55:13.269598 systemd[1]: Finished ensure-sysext.service. May 8 07:55:13.277538 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 8 07:55:13.279017 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 8 07:55:13.281058 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 07:55:13.281231 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 8 07:55:13.289982 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 07:55:13.291073 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 8 07:55:13.303886 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 8 07:55:13.312523 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 8 07:55:13.314355 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 07:55:13.314985 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 8 07:55:13.317293 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 07:55:13.317457 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 8 07:55:13.322121 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 07:55:13.322188 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 8 07:55:13.335132 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:55:13.339969 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 8 07:55:13.353965 augenrules[1462]: No rules May 8 07:55:13.354211 systemd[1]: audit-rules.service: Deactivated successfully. May 8 07:55:13.354458 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 8 07:55:13.359937 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 May 8 07:55:13.361939 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console May 8 07:55:13.365941 kernel: Console: switching to colour dummy device 80x25 May 8 07:55:13.367147 kernel: [drm] features: -virgl +edid -resource_blob -host_visible May 8 07:55:13.367182 kernel: [drm] features: -context_init May 8 07:55:13.371936 kernel: [drm] number of scanouts: 1 May 8 07:55:13.381941 kernel: [drm] number of cap sets: 0 May 8 07:55:13.384619 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 07:55:13.385181 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:55:13.387632 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 8 07:55:13.389945 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 May 8 07:55:13.393108 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:55:13.395954 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device May 8 07:55:13.397016 kernel: Console: switching to colour frame buffer device 160x50 May 8 07:55:13.407454 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device May 8 07:55:13.410745 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 07:55:13.411218 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:55:13.418030 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 8 07:55:13.442168 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 8 07:55:13.525448 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 8 07:55:13.535581 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 8 07:55:13.539531 systemd[1]: Reached target time-set.target - System Time Set. May 8 07:55:13.552997 systemd-networkd[1419]: lo: Link UP May 8 07:55:13.553005 systemd-networkd[1419]: lo: Gained carrier May 8 07:55:13.554274 systemd-networkd[1419]: Enumeration completed May 8 07:55:13.554350 systemd[1]: Started systemd-networkd.service - Network Configuration. May 8 07:55:13.556694 systemd-resolved[1421]: Positive Trust Anchors: May 8 07:55:13.556943 systemd-networkd[1419]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 07:55:13.556952 systemd-networkd[1419]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 07:55:13.557504 systemd-resolved[1421]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 07:55:13.557611 systemd-resolved[1421]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 8 07:55:13.557871 systemd-networkd[1419]: eth0: Link UP May 8 07:55:13.557878 systemd-networkd[1419]: eth0: Gained carrier May 8 07:55:13.557894 systemd-networkd[1419]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 8 07:55:13.558191 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 8 07:55:13.561031 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 8 07:55:13.566979 systemd-resolved[1421]: Using system hostname 'ci-4327-0-0-w-45c5ab2731.novalocal'. May 8 07:55:13.575180 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 8 07:55:13.575455 systemd[1]: Reached target network.target - Network. May 8 07:55:13.575516 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 8 07:55:13.575582 systemd[1]: Reached target sysinit.target - System Initialization. May 8 07:55:13.575708 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 8 07:55:13.575792 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 8 07:55:13.575864 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 8 07:55:13.576182 systemd-networkd[1419]: eth0: DHCPv4 address 172.24.4.249/24, gateway 172.24.4.1 acquired from 172.24.4.1 May 8 07:55:13.578286 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 8 07:55:13.578466 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 8 07:55:13.578541 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 8 07:55:13.578608 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 8 07:55:13.578632 systemd[1]: Reached target paths.target - Path Units. May 8 07:55:13.578691 systemd[1]: Reached target timers.target - Timer Units. May 8 07:55:13.578964 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. May 8 07:55:13.581191 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 8 07:55:13.582744 systemd[1]: Starting docker.socket - Docker Socket for the API... May 8 07:55:13.587284 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 8 07:55:13.588737 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 8 07:55:13.590188 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 8 07:55:13.595743 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 8 07:55:13.597593 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 8 07:55:13.601167 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 8 07:55:13.602705 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 8 07:55:13.605344 systemd[1]: Reached target sockets.target - Socket Units. May 8 07:55:13.607107 systemd[1]: Reached target basic.target - Basic System. May 8 07:55:13.608450 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 8 07:55:13.608476 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 8 07:55:13.611318 systemd[1]: Starting containerd.service - containerd container runtime... May 8 07:55:13.619245 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 8 07:55:13.625041 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 8 07:55:13.629955 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 8 07:55:13.636063 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 8 07:55:13.640170 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 8 07:55:13.640847 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 8 07:55:13.644852 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 8 07:55:13.650229 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 8 07:55:13.655961 jq[1501]: false May 8 07:55:13.658049 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 8 07:55:13.662280 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 8 07:55:13.665999 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Refreshing passwd entry cache May 8 07:55:13.665727 oslogin_cache_refresh[1503]: Refreshing passwd entry cache May 8 07:55:13.669018 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 8 07:55:13.675971 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Failure getting users, quitting May 8 07:55:13.675971 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 8 07:55:13.675971 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Refreshing group entry cache May 8 07:55:13.675087 oslogin_cache_refresh[1503]: Failure getting users, quitting May 8 07:55:13.675105 oslogin_cache_refresh[1503]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 8 07:55:13.675148 oslogin_cache_refresh[1503]: Refreshing group entry cache May 8 07:55:13.676232 systemd[1]: Starting systemd-logind.service - User Login Management... May 8 07:55:13.679302 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 8 07:55:13.679799 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 8 07:55:13.684827 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Failure getting groups, quitting May 8 07:55:13.684827 google_oslogin_nss_cache[1503]: oslogin_cache_refresh[1503]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 8 07:55:13.684080 oslogin_cache_refresh[1503]: Failure getting groups, quitting May 8 07:55:13.684091 oslogin_cache_refresh[1503]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 8 07:55:13.687696 extend-filesystems[1502]: Found loop4 May 8 07:55:13.687696 extend-filesystems[1502]: Found loop5 May 8 07:55:13.687696 extend-filesystems[1502]: Found loop6 May 8 07:55:13.687696 extend-filesystems[1502]: Found loop7 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda May 8 07:55:13.687696 extend-filesystems[1502]: Found vda1 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda2 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda3 May 8 07:55:13.687696 extend-filesystems[1502]: Found usr May 8 07:55:13.687696 extend-filesystems[1502]: Found vda4 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda6 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda7 May 8 07:55:13.687696 extend-filesystems[1502]: Found vda9 May 8 07:55:13.687696 extend-filesystems[1502]: Checking size of /dev/vda9 May 8 07:55:13.830791 kernel: EXT4-fs (vda9): resizing filesystem from 1617920 to 2014203 blocks May 8 07:55:13.830830 kernel: EXT4-fs (vda9): resized filesystem to 2014203 May 8 07:55:13.831373 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 43 scanned by (udev-worker) (1315) May 8 07:55:13.688122 systemd[1]: Starting update-engine.service - Update Engine... May 8 07:55:13.831550 extend-filesystems[1502]: Resized partition /dev/vda9 May 8 07:55:13.701428 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 8 07:55:13.833497 extend-filesystems[1527]: resize2fs 1.47.2 (1-Jan-2025) May 8 07:55:13.725437 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 8 07:55:13.859228 extend-filesystems[1527]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 8 07:55:13.859228 extend-filesystems[1527]: old_desc_blocks = 1, new_desc_blocks = 1 May 8 07:55:13.859228 extend-filesystems[1527]: The filesystem on /dev/vda9 is now 2014203 (4k) blocks long. May 8 07:55:13.860675 update_engine[1510]: I20250508 07:55:13.737761 1510 main.cc:92] Flatcar Update Engine starting May 8 07:55:13.732226 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 8 07:55:13.880696 extend-filesystems[1502]: Resized filesystem in /dev/vda9 May 8 07:55:13.732427 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 8 07:55:13.884248 jq[1515]: true May 8 07:55:13.732669 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 8 07:55:13.732818 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 8 07:55:13.753540 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 8 07:55:13.884679 tar[1528]: linux-amd64/helm May 8 07:55:13.753738 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 8 07:55:13.885014 jq[1534]: true May 8 07:55:13.784738 systemd[1]: motdgen.service: Deactivated successfully. May 8 07:55:13.785962 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 8 07:55:13.831872 (ntainerd)[1535]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 8 07:55:13.838704 systemd-logind[1509]: New seat seat0. May 8 07:55:13.850378 systemd[1]: extend-filesystems.service: Deactivated successfully. May 8 07:55:13.850608 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 8 07:55:13.872620 systemd-logind[1509]: Watching system buttons on /dev/input/event2 (Power Button) May 8 07:55:13.872637 systemd-logind[1509]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 8 07:55:13.873049 systemd[1]: Started systemd-logind.service - User Login Management. May 8 07:55:13.895812 dbus-daemon[1497]: [system] SELinux support is enabled May 8 07:55:13.896722 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 8 07:55:13.904729 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 8 07:55:13.904767 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 8 07:55:13.909002 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 8 07:55:13.909025 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 8 07:55:13.929897 dbus-daemon[1497]: [system] Successfully activated service 'org.freedesktop.systemd1' May 8 07:55:13.939420 update_engine[1510]: I20250508 07:55:13.937987 1510 update_check_scheduler.cc:74] Next update check in 3m13s May 8 07:55:13.941486 systemd[1]: Started update-engine.service - Update Engine. May 8 07:55:13.951271 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 8 07:55:14.017940 bash[1561]: Updated "/home/core/.ssh/authorized_keys" May 8 07:55:14.026248 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 8 07:55:14.039283 systemd[1]: Starting sshkeys.service... May 8 07:55:14.075036 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 8 07:55:14.080082 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 8 07:55:14.123267 locksmithd[1557]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 8 07:55:14.278936 containerd[1535]: time="2025-05-08T07:55:14Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 8 07:55:14.280312 containerd[1535]: time="2025-05-08T07:55:14.280280090Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 8 07:55:14.303647 containerd[1535]: time="2025-05-08T07:55:14.303609875Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="10.52µs" May 8 07:55:14.304929 containerd[1535]: time="2025-05-08T07:55:14.304873415Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 8 07:55:14.304929 containerd[1535]: time="2025-05-08T07:55:14.304903091Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 8 07:55:14.305109 containerd[1535]: time="2025-05-08T07:55:14.305084701Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 8 07:55:14.305151 containerd[1535]: time="2025-05-08T07:55:14.305112614Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 8 07:55:14.305151 containerd[1535]: time="2025-05-08T07:55:14.305140666Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 8 07:55:14.305221 containerd[1535]: time="2025-05-08T07:55:14.305200178Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 8 07:55:14.305221 containerd[1535]: time="2025-05-08T07:55:14.305215867Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 8 07:55:14.305454 containerd[1535]: time="2025-05-08T07:55:14.305427023Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 8 07:55:14.305454 containerd[1535]: time="2025-05-08T07:55:14.305449435Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 8 07:55:14.305518 containerd[1535]: time="2025-05-08T07:55:14.305462510Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 8 07:55:14.305518 containerd[1535]: time="2025-05-08T07:55:14.305472439Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 8 07:55:14.305593 containerd[1535]: time="2025-05-08T07:55:14.305547119Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 8 07:55:14.305761 containerd[1535]: time="2025-05-08T07:55:14.305734300Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 8 07:55:14.305790 containerd[1535]: time="2025-05-08T07:55:14.305771399Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 8 07:55:14.305790 containerd[1535]: time="2025-05-08T07:55:14.305784354Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 8 07:55:14.305843 containerd[1535]: time="2025-05-08T07:55:14.305817365Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 8 07:55:14.306134 containerd[1535]: time="2025-05-08T07:55:14.306111016Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 8 07:55:14.306205 containerd[1535]: time="2025-05-08T07:55:14.306175467Z" level=info msg="metadata content store policy set" policy=shared May 8 07:55:14.316670 containerd[1535]: time="2025-05-08T07:55:14.316616629Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 8 07:55:14.316670 containerd[1535]: time="2025-05-08T07:55:14.316659129Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316674137Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316687703Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316700246Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316712269Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316724712Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316738047Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316750531Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316763775Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316774425Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 8 07:55:14.316853 containerd[1535]: time="2025-05-08T07:55:14.316787921Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316879112Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316899089Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316933544Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316947260Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316957819Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316968269Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316981023Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.316991182Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317003225Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317014015Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317025617Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317082483Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317096990Z" level=info msg="Start snapshots syncer" May 8 07:55:14.317395 containerd[1535]: time="2025-05-08T07:55:14.317119312Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 8 07:55:14.318814 containerd[1535]: time="2025-05-08T07:55:14.317703919Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 8 07:55:14.318814 containerd[1535]: time="2025-05-08T07:55:14.317767979Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318161126Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318277224Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318314073Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318328951Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318347846Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318366572Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318384655Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318399183Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318432625Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318445580Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318463994Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318506364Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318522594Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 8 07:55:14.319197 containerd[1535]: time="2025-05-08T07:55:14.318536951Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318555897Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318571165Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318583438Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318600180Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318621319Z" level=info msg="runtime interface created" May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318627671Z" level=info msg="created NRI interface" May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318637259Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318653820Z" level=info msg="Connect containerd service" May 8 07:55:14.319498 containerd[1535]: time="2025-05-08T07:55:14.318683506Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 8 07:55:14.325510 containerd[1535]: time="2025-05-08T07:55:14.324609159Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 8 07:55:14.384098 sshd_keygen[1526]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 8 07:55:14.435269 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 8 07:55:14.443702 systemd[1]: Starting issuegen.service - Generate /run/issue... May 8 07:55:14.477124 systemd[1]: issuegen.service: Deactivated successfully. May 8 07:55:14.477979 systemd[1]: Finished issuegen.service - Generate /run/issue. May 8 07:55:14.486016 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 8 07:55:14.515382 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 8 07:55:14.524302 systemd[1]: Started getty@tty1.service - Getty on tty1. May 8 07:55:14.530050 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 8 07:55:14.531819 systemd[1]: Reached target getty.target - Login Prompts. May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553521495Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553582659Z" level=info msg=serving... address=/run/containerd/containerd.sock May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553612926Z" level=info msg="Start subscribing containerd event" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553636260Z" level=info msg="Start recovering state" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553722151Z" level=info msg="Start event monitor" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553737760Z" level=info msg="Start cni network conf syncer for default" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553748560Z" level=info msg="Start streaming server" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553756686Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553763879Z" level=info msg="runtime interface starting up..." May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553770622Z" level=info msg="starting plugins..." May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553783446Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 8 07:55:14.559108 containerd[1535]: time="2025-05-08T07:55:14.553879486Z" level=info msg="containerd successfully booted in 0.277074s" May 8 07:55:14.556024 systemd[1]: Started containerd.service - containerd container runtime. May 8 07:55:14.653061 tar[1528]: linux-amd64/LICENSE May 8 07:55:14.653061 tar[1528]: linux-amd64/README.md May 8 07:55:14.669007 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 8 07:55:15.491205 systemd-networkd[1419]: eth0: Gained IPv6LL May 8 07:55:15.492367 systemd-timesyncd[1441]: Network configuration changed, trying to establish connection. May 8 07:55:15.495734 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 8 07:55:15.503430 systemd[1]: Reached target network-online.target - Network is Online. May 8 07:55:15.513654 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:55:15.535615 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 8 07:55:15.593964 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 8 07:55:17.448130 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:55:17.463515 (kubelet)[1629]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:55:18.851324 kubelet[1629]: E0508 07:55:18.851190 1629 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:55:18.855357 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:55:18.855605 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:55:18.856253 systemd[1]: kubelet.service: Consumed 2.319s CPU time, 244.5M memory peak. May 8 07:55:19.635540 login[1600]: pam_lastlog(login:session): file /var/log/lastlog is locked/read, retrying May 8 07:55:19.637369 login[1599]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) May 8 07:55:19.674183 systemd-logind[1509]: New session 1 of user core. May 8 07:55:19.675398 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 8 07:55:19.680558 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 8 07:55:19.721163 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 8 07:55:19.727124 systemd[1]: Starting user@500.service - User Manager for UID 500... May 8 07:55:19.747267 (systemd)[1643]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 8 07:55:19.753840 systemd-logind[1509]: New session c1 of user core. May 8 07:55:19.954315 systemd[1643]: Queued start job for default target default.target. May 8 07:55:19.963884 systemd[1643]: Created slice app.slice - User Application Slice. May 8 07:55:19.963936 systemd[1643]: Reached target paths.target - Paths. May 8 07:55:19.963981 systemd[1643]: Reached target timers.target - Timers. May 8 07:55:19.966194 systemd[1643]: Starting dbus.socket - D-Bus User Message Bus Socket... May 8 07:55:19.980677 systemd[1643]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 8 07:55:19.981000 systemd[1643]: Reached target sockets.target - Sockets. May 8 07:55:19.981116 systemd[1643]: Reached target basic.target - Basic System. May 8 07:55:19.981221 systemd[1643]: Reached target default.target - Main User Target. May 8 07:55:19.981280 systemd[1]: Started user@500.service - User Manager for UID 500. May 8 07:55:19.981291 systemd[1643]: Startup finished in 213ms. May 8 07:55:19.988093 systemd[1]: Started session-1.scope - Session 1 of User core. May 8 07:55:20.640651 login[1600]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) May 8 07:55:20.651149 systemd-logind[1509]: New session 2 of user core. May 8 07:55:20.664461 systemd[1]: Started session-2.scope - Session 2 of User core. May 8 07:55:20.718384 coreos-metadata[1496]: May 08 07:55:20.718 WARN failed to locate config-drive, using the metadata service API instead May 8 07:55:20.787339 coreos-metadata[1496]: May 08 07:55:20.787 INFO Fetching http://169.254.169.254/openstack/2012-08-10/meta_data.json: Attempt #1 May 8 07:55:20.971280 coreos-metadata[1496]: May 08 07:55:20.971 INFO Fetch successful May 8 07:55:20.971280 coreos-metadata[1496]: May 08 07:55:20.971 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 May 8 07:55:20.982487 coreos-metadata[1496]: May 08 07:55:20.982 INFO Fetch successful May 8 07:55:20.982487 coreos-metadata[1496]: May 08 07:55:20.982 INFO Fetching http://169.254.169.254/latest/meta-data/instance-id: Attempt #1 May 8 07:55:20.994723 coreos-metadata[1496]: May 08 07:55:20.994 INFO Fetch successful May 8 07:55:20.994723 coreos-metadata[1496]: May 08 07:55:20.994 INFO Fetching http://169.254.169.254/latest/meta-data/instance-type: Attempt #1 May 8 07:55:21.009265 coreos-metadata[1496]: May 08 07:55:21.009 INFO Fetch successful May 8 07:55:21.009265 coreos-metadata[1496]: May 08 07:55:21.009 INFO Fetching http://169.254.169.254/latest/meta-data/local-ipv4: Attempt #1 May 8 07:55:21.022435 coreos-metadata[1496]: May 08 07:55:21.022 INFO Fetch successful May 8 07:55:21.022435 coreos-metadata[1496]: May 08 07:55:21.022 INFO Fetching http://169.254.169.254/latest/meta-data/public-ipv4: Attempt #1 May 8 07:55:21.035731 coreos-metadata[1496]: May 08 07:55:21.035 INFO Fetch successful May 8 07:55:21.082983 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 8 07:55:21.084360 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 8 07:55:21.203798 coreos-metadata[1568]: May 08 07:55:21.203 WARN failed to locate config-drive, using the metadata service API instead May 8 07:55:21.247195 coreos-metadata[1568]: May 08 07:55:21.247 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys: Attempt #1 May 8 07:55:21.262749 coreos-metadata[1568]: May 08 07:55:21.262 INFO Fetch successful May 8 07:55:21.262749 coreos-metadata[1568]: May 08 07:55:21.262 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys/0/openssh-key: Attempt #1 May 8 07:55:21.275827 coreos-metadata[1568]: May 08 07:55:21.275 INFO Fetch successful May 8 07:55:21.281477 unknown[1568]: wrote ssh authorized keys file for user: core May 8 07:55:21.325002 update-ssh-keys[1680]: Updated "/home/core/.ssh/authorized_keys" May 8 07:55:21.326645 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 8 07:55:21.330147 systemd[1]: Finished sshkeys.service. May 8 07:55:21.335720 systemd[1]: Reached target multi-user.target - Multi-User System. May 8 07:55:21.336171 systemd[1]: Startup finished in 3.757s (kernel) + 16.749s (initrd) + 10.843s (userspace) = 31.350s. May 8 07:55:28.862362 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 8 07:55:28.866000 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:55:29.200649 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:55:29.220448 (kubelet)[1692]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:55:29.330348 kubelet[1692]: E0508 07:55:29.330193 1692 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:55:29.337531 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:55:29.337874 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:55:29.338637 systemd[1]: kubelet.service: Consumed 326ms CPU time, 96.1M memory peak. May 8 07:55:39.362491 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 8 07:55:39.365706 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:55:39.682341 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:55:39.697779 (kubelet)[1708]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:55:39.762876 kubelet[1708]: E0508 07:55:39.762776 1708 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:55:39.765400 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:55:39.765663 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:55:39.766477 systemd[1]: kubelet.service: Consumed 266ms CPU time, 96.1M memory peak. May 8 07:55:45.709544 systemd-timesyncd[1441]: Contacted time server 5.161.111.190:123 (2.flatcar.pool.ntp.org). May 8 07:55:45.709664 systemd-timesyncd[1441]: Initial clock synchronization to Thu 2025-05-08 07:55:45.944936 UTC. May 8 07:55:49.863727 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. May 8 07:55:49.867004 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:55:50.182864 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:55:50.202504 (kubelet)[1724]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:55:50.294371 kubelet[1724]: E0508 07:55:50.294310 1724 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:55:50.297515 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:55:50.297770 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:55:50.298896 systemd[1]: kubelet.service: Consumed 292ms CPU time, 98.1M memory peak. May 8 07:55:59.025042 update_engine[1510]: I20250508 07:55:59.024821 1510 update_attempter.cc:509] Updating boot flags... May 8 07:55:59.092025 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 43 scanned by (udev-worker) (1740) May 8 07:55:59.148976 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 43 scanned by (udev-worker) (1740) May 8 07:55:59.240997 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 43 scanned by (udev-worker) (1740) May 8 07:56:00.362482 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. May 8 07:56:00.366550 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:00.687471 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:00.702784 (kubelet)[1756]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:00.783734 kubelet[1756]: E0508 07:56:00.783597 1756 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:00.786205 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:00.786444 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:00.787222 systemd[1]: kubelet.service: Consumed 286ms CPU time, 98.2M memory peak. May 8 07:56:10.862478 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. May 8 07:56:10.865654 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:11.202825 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:11.218796 (kubelet)[1772]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:11.292171 kubelet[1772]: E0508 07:56:11.292084 1772 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:11.295137 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:11.295269 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:11.295535 systemd[1]: kubelet.service: Consumed 280ms CPU time, 97.5M memory peak. May 8 07:56:21.362526 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. May 8 07:56:21.365789 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:21.671642 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:21.680156 (kubelet)[1787]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:21.729655 kubelet[1787]: E0508 07:56:21.729544 1787 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:21.732690 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:21.733020 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:21.733507 systemd[1]: kubelet.service: Consumed 239ms CPU time, 97.3M memory peak. May 8 07:56:31.862423 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. May 8 07:56:31.866038 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:32.184124 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:32.201739 (kubelet)[1804]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:32.269902 kubelet[1804]: E0508 07:56:32.269799 1804 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:32.272523 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:32.272853 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:32.273626 systemd[1]: kubelet.service: Consumed 270ms CPU time, 95.8M memory peak. May 8 07:56:42.362287 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. May 8 07:56:42.365480 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:42.654141 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:42.671044 (kubelet)[1821]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:42.740904 kubelet[1821]: E0508 07:56:42.740806 1821 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:42.744470 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:42.744813 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:42.745700 systemd[1]: kubelet.service: Consumed 272ms CPU time, 96M memory peak. May 8 07:56:52.862484 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. May 8 07:56:52.867180 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:56:53.163308 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:56:53.172139 (kubelet)[1836]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:56:53.276420 kubelet[1836]: E0508 07:56:53.276296 1836 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:56:53.280783 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:56:53.281297 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:56:53.281989 systemd[1]: kubelet.service: Consumed 266ms CPU time, 98.3M memory peak. May 8 07:57:03.363217 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. May 8 07:57:03.366485 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:03.697770 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:03.715746 (kubelet)[1854]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:03.807540 kubelet[1854]: E0508 07:57:03.807488 1854 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:03.810907 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:03.811064 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:03.811346 systemd[1]: kubelet.service: Consumed 297ms CPU time, 96M memory peak. May 8 07:57:13.862420 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. May 8 07:57:13.865729 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:14.170544 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:14.183295 (kubelet)[1870]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:14.234062 kubelet[1870]: E0508 07:57:14.234024 1870 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:14.237459 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:14.237771 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:14.238451 systemd[1]: kubelet.service: Consumed 239ms CPU time, 97.7M memory peak. May 8 07:57:24.362497 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. May 8 07:57:24.365633 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:24.644384 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:24.657178 (kubelet)[1886]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:24.705515 kubelet[1886]: E0508 07:57:24.705424 1886 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:24.709598 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:24.709877 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:24.710828 systemd[1]: kubelet.service: Consumed 229ms CPU time, 99.5M memory peak. May 8 07:57:34.862030 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. May 8 07:57:34.864385 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:35.172245 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:35.178272 (kubelet)[1901]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:35.222964 kubelet[1901]: E0508 07:57:35.222927 1901 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:35.226874 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:35.227212 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:35.228060 systemd[1]: kubelet.service: Consumed 227ms CPU time, 99.3M memory peak. May 8 07:57:45.362369 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 14. May 8 07:57:45.365844 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:45.635396 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:45.641145 (kubelet)[1917]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:45.685532 kubelet[1917]: E0508 07:57:45.685477 1917 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:45.689347 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:45.689622 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:45.690223 systemd[1]: kubelet.service: Consumed 225ms CPU time, 97.7M memory peak. May 8 07:57:55.862388 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 15. May 8 07:57:55.865531 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:57:56.234348 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:57:56.253523 (kubelet)[1934]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:57:56.332507 kubelet[1934]: E0508 07:57:56.332425 1934 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:57:56.334289 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:57:56.334556 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:57:56.335335 systemd[1]: kubelet.service: Consumed 280ms CPU time, 99.7M memory peak. May 8 07:58:06.362290 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 16. May 8 07:58:06.366314 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:06.671371 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:06.680305 (kubelet)[1949]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:06.730551 kubelet[1949]: E0508 07:58:06.730506 1949 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:06.732383 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:06.732676 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:06.733276 systemd[1]: kubelet.service: Consumed 240ms CPU time, 97.2M memory peak. May 8 07:58:16.862294 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 17. May 8 07:58:16.865025 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:17.160707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:17.176442 (kubelet)[1966]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:17.255360 kubelet[1966]: E0508 07:58:17.255267 1966 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:17.258485 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:17.258738 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:17.259692 systemd[1]: kubelet.service: Consumed 265ms CPU time, 97.6M memory peak. May 8 07:58:25.621670 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 8 07:58:25.625790 systemd[1]: Started sshd@0-172.24.4.249:22-172.24.4.1:56050.service - OpenSSH per-connection server daemon (172.24.4.1:56050). May 8 07:58:26.607888 sshd[1975]: Accepted publickey for core from 172.24.4.1 port 56050 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:26.610774 sshd-session[1975]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:26.623034 systemd-logind[1509]: New session 3 of user core. May 8 07:58:26.633283 systemd[1]: Started session-3.scope - Session 3 of User core. May 8 07:58:26.986648 update_engine[1510]: I20250508 07:58:26.986488 1510 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs May 8 07:58:26.986648 update_engine[1510]: I20250508 07:58:26.986593 1510 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs May 8 07:58:26.987766 update_engine[1510]: I20250508 07:58:26.987031 1510 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs May 8 07:58:26.988326 update_engine[1510]: I20250508 07:58:26.988246 1510 omaha_request_params.cc:62] Current group set to developer May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988499 1510 update_attempter.cc:499] Already updated boot flags. Skipping. May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988545 1510 update_attempter.cc:643] Scheduling an action processor start. May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988588 1510 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988663 1510 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988817 1510 omaha_request_action.cc:271] Posting an Omaha request to disabled May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988847 1510 omaha_request_action.cc:272] Request: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: May 8 07:58:26.988999 update_engine[1510]: I20250508 07:58:26.988869 1510 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 8 07:58:26.990691 locksmithd[1557]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 May 8 07:58:26.992991 update_engine[1510]: I20250508 07:58:26.992539 1510 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 8 07:58:26.993583 update_engine[1510]: I20250508 07:58:26.993468 1510 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 8 07:58:27.000601 update_engine[1510]: E20250508 07:58:27.000509 1510 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 8 07:58:27.000750 update_engine[1510]: I20250508 07:58:27.000674 1510 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 May 8 07:58:27.230277 systemd[1]: Started sshd@1-172.24.4.249:22-172.24.4.1:56052.service - OpenSSH per-connection server daemon (172.24.4.1:56052). May 8 07:58:27.362360 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 18. May 8 07:58:27.366161 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:27.624343 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:27.637450 (kubelet)[1988]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:27.688016 kubelet[1988]: E0508 07:58:27.687979 1988 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:27.691491 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:27.691633 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:27.691939 systemd[1]: kubelet.service: Consumed 234ms CPU time, 97.2M memory peak. May 8 07:58:28.915589 sshd[1980]: Accepted publickey for core from 172.24.4.1 port 56052 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:28.917809 sshd-session[1980]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:28.929019 systemd-logind[1509]: New session 4 of user core. May 8 07:58:28.937374 systemd[1]: Started session-4.scope - Session 4 of User core. May 8 07:58:29.655446 sshd[1999]: Connection closed by 172.24.4.1 port 56052 May 8 07:58:29.656399 sshd-session[1980]: pam_unix(sshd:session): session closed for user core May 8 07:58:29.673220 systemd[1]: sshd@1-172.24.4.249:22-172.24.4.1:56052.service: Deactivated successfully. May 8 07:58:29.676459 systemd[1]: session-4.scope: Deactivated successfully. May 8 07:58:29.678686 systemd-logind[1509]: Session 4 logged out. Waiting for processes to exit. May 8 07:58:29.683093 systemd[1]: Started sshd@2-172.24.4.249:22-172.24.4.1:56066.service - OpenSSH per-connection server daemon (172.24.4.1:56066). May 8 07:58:29.686699 systemd-logind[1509]: Removed session 4. May 8 07:58:30.870161 sshd[2004]: Accepted publickey for core from 172.24.4.1 port 56066 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:30.873023 sshd-session[2004]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:30.885042 systemd-logind[1509]: New session 5 of user core. May 8 07:58:30.892236 systemd[1]: Started session-5.scope - Session 5 of User core. May 8 07:58:31.503310 sshd[2007]: Connection closed by 172.24.4.1 port 56066 May 8 07:58:31.504212 sshd-session[2004]: pam_unix(sshd:session): session closed for user core May 8 07:58:31.520122 systemd[1]: sshd@2-172.24.4.249:22-172.24.4.1:56066.service: Deactivated successfully. May 8 07:58:31.523402 systemd[1]: session-5.scope: Deactivated successfully. May 8 07:58:31.527271 systemd-logind[1509]: Session 5 logged out. Waiting for processes to exit. May 8 07:58:31.530533 systemd[1]: Started sshd@3-172.24.4.249:22-172.24.4.1:56074.service - OpenSSH per-connection server daemon (172.24.4.1:56074). May 8 07:58:31.534782 systemd-logind[1509]: Removed session 5. May 8 07:58:32.689311 sshd[2012]: Accepted publickey for core from 172.24.4.1 port 56074 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:32.691820 sshd-session[2012]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:32.704019 systemd-logind[1509]: New session 6 of user core. May 8 07:58:32.711243 systemd[1]: Started session-6.scope - Session 6 of User core. May 8 07:58:33.429965 sshd[2015]: Connection closed by 172.24.4.1 port 56074 May 8 07:58:33.430429 sshd-session[2012]: pam_unix(sshd:session): session closed for user core May 8 07:58:33.446665 systemd[1]: sshd@3-172.24.4.249:22-172.24.4.1:56074.service: Deactivated successfully. May 8 07:58:33.449830 systemd[1]: session-6.scope: Deactivated successfully. May 8 07:58:33.451827 systemd-logind[1509]: Session 6 logged out. Waiting for processes to exit. May 8 07:58:33.456498 systemd[1]: Started sshd@4-172.24.4.249:22-172.24.4.1:56078.service - OpenSSH per-connection server daemon (172.24.4.1:56078). May 8 07:58:33.460156 systemd-logind[1509]: Removed session 6. May 8 07:58:34.660875 sshd[2020]: Accepted publickey for core from 172.24.4.1 port 56078 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:34.663488 sshd-session[2020]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:34.675594 systemd-logind[1509]: New session 7 of user core. May 8 07:58:34.684203 systemd[1]: Started session-7.scope - Session 7 of User core. May 8 07:58:35.157432 sudo[2024]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 8 07:58:35.158781 sudo[2024]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 07:58:35.176973 sudo[2024]: pam_unix(sudo:session): session closed for user root May 8 07:58:35.403488 sshd[2023]: Connection closed by 172.24.4.1 port 56078 May 8 07:58:35.401032 sshd-session[2020]: pam_unix(sshd:session): session closed for user core May 8 07:58:35.417712 systemd[1]: sshd@4-172.24.4.249:22-172.24.4.1:56078.service: Deactivated successfully. May 8 07:58:35.421636 systemd[1]: session-7.scope: Deactivated successfully. May 8 07:58:35.425609 systemd-logind[1509]: Session 7 logged out. Waiting for processes to exit. May 8 07:58:35.430229 systemd[1]: Started sshd@5-172.24.4.249:22-172.24.4.1:41072.service - OpenSSH per-connection server daemon (172.24.4.1:41072). May 8 07:58:35.433356 systemd-logind[1509]: Removed session 7. May 8 07:58:36.566466 sshd[2029]: Accepted publickey for core from 172.24.4.1 port 41072 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:36.569210 sshd-session[2029]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:36.582023 systemd-logind[1509]: New session 8 of user core. May 8 07:58:36.589223 systemd[1]: Started session-8.scope - Session 8 of User core. May 8 07:58:36.986737 update_engine[1510]: I20250508 07:58:36.986478 1510 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 8 07:58:36.987460 update_engine[1510]: I20250508 07:58:36.987104 1510 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 8 07:58:36.987709 update_engine[1510]: I20250508 07:58:36.987547 1510 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 8 07:58:36.992675 update_engine[1510]: E20250508 07:58:36.992561 1510 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 8 07:58:36.992785 update_engine[1510]: I20250508 07:58:36.992728 1510 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 May 8 07:58:37.016668 sudo[2034]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 8 07:58:37.017365 sudo[2034]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 07:58:37.025024 sudo[2034]: pam_unix(sudo:session): session closed for user root May 8 07:58:37.037147 sudo[2033]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 8 07:58:37.037776 sudo[2033]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 07:58:37.059753 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 8 07:58:37.137994 augenrules[2056]: No rules May 8 07:58:37.140818 systemd[1]: audit-rules.service: Deactivated successfully. May 8 07:58:37.141429 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 8 07:58:37.143182 sudo[2033]: pam_unix(sudo:session): session closed for user root May 8 07:58:37.349714 sshd[2032]: Connection closed by 172.24.4.1 port 41072 May 8 07:58:37.350664 sshd-session[2029]: pam_unix(sshd:session): session closed for user core May 8 07:58:37.371003 systemd[1]: sshd@5-172.24.4.249:22-172.24.4.1:41072.service: Deactivated successfully. May 8 07:58:37.374611 systemd[1]: session-8.scope: Deactivated successfully. May 8 07:58:37.380268 systemd-logind[1509]: Session 8 logged out. Waiting for processes to exit. May 8 07:58:37.382495 systemd[1]: Started sshd@6-172.24.4.249:22-172.24.4.1:41080.service - OpenSSH per-connection server daemon (172.24.4.1:41080). May 8 07:58:37.386117 systemd-logind[1509]: Removed session 8. May 8 07:58:37.862357 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 19. May 8 07:58:37.866392 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:38.035271 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:38.044178 (kubelet)[2075]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:38.086055 kubelet[2075]: E0508 07:58:38.085969 2075 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:38.088321 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:38.088553 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:38.089028 systemd[1]: kubelet.service: Consumed 183ms CPU time, 95.7M memory peak. May 8 07:58:38.384725 sshd[2064]: Accepted publickey for core from 172.24.4.1 port 41080 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 07:58:38.387366 sshd-session[2064]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 07:58:38.400382 systemd-logind[1509]: New session 9 of user core. May 8 07:58:38.408254 systemd[1]: Started session-9.scope - Session 9 of User core. May 8 07:58:38.812047 sudo[2083]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 8 07:58:38.812668 sudo[2083]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 8 07:58:39.582845 systemd[1]: Starting docker.service - Docker Application Container Engine... May 8 07:58:39.599220 (dockerd)[2100]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 8 07:58:40.004852 dockerd[2100]: time="2025-05-08T07:58:40.004780683Z" level=info msg="Starting up" May 8 07:58:40.007330 dockerd[2100]: time="2025-05-08T07:58:40.006854957Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 8 07:58:40.107309 systemd[1]: var-lib-docker-metacopy\x2dcheck931728746-merged.mount: Deactivated successfully. May 8 07:58:40.146969 dockerd[2100]: time="2025-05-08T07:58:40.146850408Z" level=info msg="Loading containers: start." May 8 07:58:40.166049 kernel: Initializing XFRM netlink socket May 8 07:58:40.471229 systemd-networkd[1419]: docker0: Link UP May 8 07:58:40.486849 dockerd[2100]: time="2025-05-08T07:58:40.486788962Z" level=info msg="Loading containers: done." May 8 07:58:40.508616 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2396554086-merged.mount: Deactivated successfully. May 8 07:58:40.520029 dockerd[2100]: time="2025-05-08T07:58:40.519765995Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 8 07:58:40.520029 dockerd[2100]: time="2025-05-08T07:58:40.519855325Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 8 07:58:40.520029 dockerd[2100]: time="2025-05-08T07:58:40.519989039Z" level=info msg="Initializing buildkit" May 8 07:58:40.591376 dockerd[2100]: time="2025-05-08T07:58:40.591205726Z" level=info msg="Completed buildkit initialization" May 8 07:58:40.611014 dockerd[2100]: time="2025-05-08T07:58:40.610495230Z" level=info msg="Daemon has completed initialization" May 8 07:58:40.611014 dockerd[2100]: time="2025-05-08T07:58:40.610703855Z" level=info msg="API listen on /run/docker.sock" May 8 07:58:40.611336 systemd[1]: Started docker.service - Docker Application Container Engine. May 8 07:58:42.380654 containerd[1535]: time="2025-05-08T07:58:42.380605138Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 8 07:58:43.138434 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount933336176.mount: Deactivated successfully. May 8 07:58:45.043942 containerd[1535]: time="2025-05-08T07:58:45.043875213Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:45.045367 containerd[1535]: time="2025-05-08T07:58:45.045148481Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.12: active requests=0, bytes read=32674881" May 8 07:58:45.046510 containerd[1535]: time="2025-05-08T07:58:45.046480009Z" level=info msg="ImageCreate event name:\"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:45.049859 containerd[1535]: time="2025-05-08T07:58:45.049810864Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:45.050810 containerd[1535]: time="2025-05-08T07:58:45.050778541Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.12\" with image id \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.12\", repo digest \"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\", size \"32671673\" in 2.670114471s" May 8 07:58:45.050861 containerd[1535]: time="2025-05-08T07:58:45.050814590Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\"" May 8 07:58:45.069825 containerd[1535]: time="2025-05-08T07:58:45.069781638Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 8 07:58:46.986454 update_engine[1510]: I20250508 07:58:46.986325 1510 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 8 07:58:46.986954 update_engine[1510]: I20250508 07:58:46.986636 1510 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 8 07:58:46.986954 update_engine[1510]: I20250508 07:58:46.986891 1510 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 8 07:58:46.992477 update_engine[1510]: E20250508 07:58:46.992383 1510 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 8 07:58:46.992477 update_engine[1510]: I20250508 07:58:46.992451 1510 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 May 8 07:58:47.327560 containerd[1535]: time="2025-05-08T07:58:47.326965482Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:47.329091 containerd[1535]: time="2025-05-08T07:58:47.329055162Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.12: active requests=0, bytes read=29617542" May 8 07:58:47.330471 containerd[1535]: time="2025-05-08T07:58:47.330424773Z" level=info msg="ImageCreate event name:\"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:47.333367 containerd[1535]: time="2025-05-08T07:58:47.333322669Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:47.334845 containerd[1535]: time="2025-05-08T07:58:47.334709324Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.12\" with image id \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.12\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\", size \"31105907\" in 2.264891255s" May 8 07:58:47.334845 containerd[1535]: time="2025-05-08T07:58:47.334765580Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\"" May 8 07:58:47.352579 containerd[1535]: time="2025-05-08T07:58:47.352260759Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 8 07:58:48.112863 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 20. May 8 07:58:48.118293 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:48.278047 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:48.286454 (kubelet)[2390]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:48.337861 kubelet[2390]: E0508 07:58:48.337795 2390 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:48.340296 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:48.340529 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:48.340900 systemd[1]: kubelet.service: Consumed 176ms CPU time, 95.2M memory peak. May 8 07:58:49.136517 containerd[1535]: time="2025-05-08T07:58:49.136430688Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:49.137981 containerd[1535]: time="2025-05-08T07:58:49.137943013Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.12: active requests=0, bytes read=17903690" May 8 07:58:49.139491 containerd[1535]: time="2025-05-08T07:58:49.139450027Z" level=info msg="ImageCreate event name:\"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:49.143252 containerd[1535]: time="2025-05-08T07:58:49.143192951Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:49.144385 containerd[1535]: time="2025-05-08T07:58:49.144217729Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.12\" with image id \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.12\", repo digest \"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\", size \"19392073\" in 1.791918508s" May 8 07:58:49.144385 containerd[1535]: time="2025-05-08T07:58:49.144252405Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\"" May 8 07:58:49.161026 containerd[1535]: time="2025-05-08T07:58:49.160987432Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 8 07:58:50.535644 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2343156430.mount: Deactivated successfully. May 8 07:58:51.498432 containerd[1535]: time="2025-05-08T07:58:51.498285834Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:51.500758 containerd[1535]: time="2025-05-08T07:58:51.500668394Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.12: active requests=0, bytes read=29185825" May 8 07:58:51.503501 containerd[1535]: time="2025-05-08T07:58:51.503401191Z" level=info msg="ImageCreate event name:\"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:51.508404 containerd[1535]: time="2025-05-08T07:58:51.508306128Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:51.510123 containerd[1535]: time="2025-05-08T07:58:51.509859230Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.12\" with image id \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\", repo tag \"registry.k8s.io/kube-proxy:v1.30.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\", size \"29184836\" in 2.348819868s" May 8 07:58:51.510123 containerd[1535]: time="2025-05-08T07:58:51.509970843Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\"" May 8 07:58:51.549562 containerd[1535]: time="2025-05-08T07:58:51.549445260Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 8 07:58:52.232026 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1926638567.mount: Deactivated successfully. May 8 07:58:53.403822 containerd[1535]: time="2025-05-08T07:58:53.403772687Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:53.404966 containerd[1535]: time="2025-05-08T07:58:53.404910281Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185769" May 8 07:58:53.406255 containerd[1535]: time="2025-05-08T07:58:53.406156581Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:53.409777 containerd[1535]: time="2025-05-08T07:58:53.409732183Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:53.412577 containerd[1535]: time="2025-05-08T07:58:53.412393515Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.862865116s" May 8 07:58:53.412577 containerd[1535]: time="2025-05-08T07:58:53.412454732Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" May 8 07:58:53.439099 containerd[1535]: time="2025-05-08T07:58:53.439049398Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 8 07:58:54.015880 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3125732298.mount: Deactivated successfully. May 8 07:58:54.026978 containerd[1535]: time="2025-05-08T07:58:54.026832399Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:54.028757 containerd[1535]: time="2025-05-08T07:58:54.028635590Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322298" May 8 07:58:54.030463 containerd[1535]: time="2025-05-08T07:58:54.030368317Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:54.035488 containerd[1535]: time="2025-05-08T07:58:54.035350365Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:54.038038 containerd[1535]: time="2025-05-08T07:58:54.037198280Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 598.052759ms" May 8 07:58:54.038038 containerd[1535]: time="2025-05-08T07:58:54.037264276Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" May 8 07:58:54.076255 containerd[1535]: time="2025-05-08T07:58:54.076155165Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 8 07:58:54.744807 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2785268476.mount: Deactivated successfully. May 8 07:58:56.987421 update_engine[1510]: I20250508 07:58:56.987036 1510 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 8 07:58:56.987421 update_engine[1510]: I20250508 07:58:56.987331 1510 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 8 07:58:56.988127 update_engine[1510]: I20250508 07:58:56.987573 1510 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 8 07:58:56.993053 update_engine[1510]: E20250508 07:58:56.992992 1510 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 8 07:58:56.993053 update_engine[1510]: I20250508 07:58:56.993033 1510 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 8 07:58:56.993053 update_engine[1510]: I20250508 07:58:56.993041 1510 omaha_request_action.cc:617] Omaha request response: May 8 07:58:56.993169 update_engine[1510]: E20250508 07:58:56.993103 1510 omaha_request_action.cc:636] Omaha request network transfer failed. May 8 07:58:56.993169 update_engine[1510]: I20250508 07:58:56.993134 1510 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. May 8 07:58:56.993169 update_engine[1510]: I20250508 07:58:56.993140 1510 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 8 07:58:56.993169 update_engine[1510]: I20250508 07:58:56.993149 1510 update_attempter.cc:306] Processing Done. May 8 07:58:56.993169 update_engine[1510]: E20250508 07:58:56.993161 1510 update_attempter.cc:619] Update failed. May 8 07:58:56.993169 update_engine[1510]: I20250508 07:58:56.993166 1510 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993172 1510 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993178 1510 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993250 1510 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993273 1510 omaha_request_action.cc:271] Posting an Omaha request to disabled May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993278 1510 omaha_request_action.cc:272] Request: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: May 8 07:58:56.993307 update_engine[1510]: I20250508 07:58:56.993283 1510 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 8 07:58:56.993568 update_engine[1510]: I20250508 07:58:56.993405 1510 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 8 07:58:56.993595 update_engine[1510]: I20250508 07:58:56.993561 1510 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 8 07:58:56.994834 locksmithd[1557]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 May 8 07:58:56.998985 update_engine[1510]: E20250508 07:58:56.998696 1510 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 8 07:58:56.999413 update_engine[1510]: I20250508 07:58:56.999375 1510 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 8 07:58:56.999413 update_engine[1510]: I20250508 07:58:56.999397 1510 omaha_request_action.cc:617] Omaha request response: May 8 07:58:56.999413 update_engine[1510]: I20250508 07:58:56.999404 1510 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 8 07:58:56.999413 update_engine[1510]: I20250508 07:58:56.999409 1510 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 8 07:58:56.999413 update_engine[1510]: I20250508 07:58:56.999414 1510 update_attempter.cc:306] Processing Done. May 8 07:58:56.999557 update_engine[1510]: I20250508 07:58:56.999420 1510 update_attempter.cc:310] Error event sent. May 8 07:58:56.999557 update_engine[1510]: I20250508 07:58:56.999429 1510 update_check_scheduler.cc:74] Next update check in 46m38s May 8 07:58:57.000469 locksmithd[1557]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 May 8 07:58:57.861756 containerd[1535]: time="2025-05-08T07:58:57.861689513Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:57.863274 containerd[1535]: time="2025-05-08T07:58:57.862947668Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238579" May 8 07:58:57.864584 containerd[1535]: time="2025-05-08T07:58:57.864529369Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:57.867861 containerd[1535]: time="2025-05-08T07:58:57.867793041Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:58:57.869045 containerd[1535]: time="2025-05-08T07:58:57.868834313Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 3.792643611s" May 8 07:58:57.869045 containerd[1535]: time="2025-05-08T07:58:57.868887073Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" May 8 07:58:58.362370 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 21. May 8 07:58:58.365625 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:58:58.538808 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:58:58.545167 (kubelet)[2552]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 8 07:58:58.594181 kubelet[2552]: E0508 07:58:58.594129 2552 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 07:58:58.595798 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 07:58:58.595950 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 07:58:58.596616 systemd[1]: kubelet.service: Consumed 184ms CPU time, 95.4M memory peak. May 8 07:59:01.830487 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:01.831550 systemd[1]: kubelet.service: Consumed 184ms CPU time, 95.4M memory peak. May 8 07:59:01.836318 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:59:01.876119 systemd[1]: Reload requested from client PID 2631 ('systemctl') (unit session-9.scope)... May 8 07:59:01.876178 systemd[1]: Reloading... May 8 07:59:01.970948 zram_generator::config[2682]: No configuration found. May 8 07:59:02.162247 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 07:59:02.301934 systemd[1]: Reloading finished in 425 ms. May 8 07:59:02.363371 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:02.375313 (kubelet)[2734]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 8 07:59:02.377891 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:59:02.378899 systemd[1]: kubelet.service: Deactivated successfully. May 8 07:59:02.379463 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:02.379555 systemd[1]: kubelet.service: Consumed 147ms CPU time, 84.5M memory peak. May 8 07:59:02.384625 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:59:02.717491 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:02.729224 (kubelet)[2745]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 8 07:59:02.793531 kubelet[2745]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 07:59:02.793531 kubelet[2745]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 07:59:02.793531 kubelet[2745]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 07:59:02.794173 kubelet[2745]: I0508 07:59:02.793548 2745 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 07:59:03.726995 kubelet[2745]: I0508 07:59:03.726958 2745 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 07:59:03.726995 kubelet[2745]: I0508 07:59:03.726983 2745 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 07:59:03.727216 kubelet[2745]: I0508 07:59:03.727177 2745 server.go:927] "Client rotation is on, will bootstrap in background" May 8 07:59:03.743090 kubelet[2745]: I0508 07:59:03.743046 2745 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 07:59:03.745539 kubelet[2745]: E0508 07:59:03.745412 2745 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://172.24.4.249:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.759109 kubelet[2745]: I0508 07:59:03.759073 2745 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 07:59:03.759948 kubelet[2745]: I0508 07:59:03.759264 2745 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 07:59:03.759948 kubelet[2745]: I0508 07:59:03.759299 2745 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4327-0-0-w-45c5ab2731.novalocal","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 07:59:03.760709 kubelet[2745]: I0508 07:59:03.760464 2745 topology_manager.go:138] "Creating topology manager with none policy" May 8 07:59:03.760709 kubelet[2745]: I0508 07:59:03.760485 2745 container_manager_linux.go:301] "Creating device plugin manager" May 8 07:59:03.760709 kubelet[2745]: I0508 07:59:03.760611 2745 state_mem.go:36] "Initialized new in-memory state store" May 8 07:59:03.762210 kubelet[2745]: I0508 07:59:03.762058 2745 kubelet.go:400] "Attempting to sync node with API server" May 8 07:59:03.762210 kubelet[2745]: I0508 07:59:03.762076 2745 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 07:59:03.762210 kubelet[2745]: I0508 07:59:03.762096 2745 kubelet.go:312] "Adding apiserver pod source" May 8 07:59:03.762210 kubelet[2745]: I0508 07:59:03.762110 2745 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 07:59:03.768783 kubelet[2745]: W0508 07:59:03.768377 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.24.4.249:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.768783 kubelet[2745]: E0508 07:59:03.768425 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://172.24.4.249:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.768783 kubelet[2745]: W0508 07:59:03.768482 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.24.4.249:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4327-0-0-w-45c5ab2731.novalocal&limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.768783 kubelet[2745]: E0508 07:59:03.768518 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://172.24.4.249:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4327-0-0-w-45c5ab2731.novalocal&limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.769256 kubelet[2745]: I0508 07:59:03.769105 2745 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 8 07:59:03.771962 kubelet[2745]: I0508 07:59:03.771104 2745 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 07:59:03.771962 kubelet[2745]: W0508 07:59:03.771150 2745 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 8 07:59:03.771962 kubelet[2745]: I0508 07:59:03.771825 2745 server.go:1264] "Started kubelet" May 8 07:59:03.782360 kubelet[2745]: I0508 07:59:03.782250 2745 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 07:59:03.785880 kubelet[2745]: E0508 07:59:03.785521 2745 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.24.4.249:6443/api/v1/namespaces/default/events\": dial tcp 172.24.4.249:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4327-0-0-w-45c5ab2731.novalocal.183d7e5744d6db55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4327-0-0-w-45c5ab2731.novalocal,UID:ci-4327-0-0-w-45c5ab2731.novalocal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4327-0-0-w-45c5ab2731.novalocal,},FirstTimestamp:2025-05-08 07:59:03.771806549 +0000 UTC m=+1.036664464,LastTimestamp:2025-05-08 07:59:03.771806549 +0000 UTC m=+1.036664464,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4327-0-0-w-45c5ab2731.novalocal,}" May 8 07:59:03.787959 kubelet[2745]: I0508 07:59:03.787476 2745 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 07:59:03.788650 kubelet[2745]: I0508 07:59:03.788594 2745 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 07:59:03.789792 kubelet[2745]: I0508 07:59:03.789780 2745 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 07:59:03.789900 kubelet[2745]: I0508 07:59:03.789890 2745 reconciler.go:26] "Reconciler: start to sync state" May 8 07:59:03.790423 kubelet[2745]: I0508 07:59:03.790392 2745 server.go:455] "Adding debug handlers to kubelet server" May 8 07:59:03.792095 kubelet[2745]: I0508 07:59:03.792023 2745 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 07:59:03.792433 kubelet[2745]: I0508 07:59:03.792396 2745 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 07:59:03.797607 kubelet[2745]: E0508 07:59:03.795886 2745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.24.4.249:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4327-0-0-w-45c5ab2731.novalocal?timeout=10s\": dial tcp 172.24.4.249:6443: connect: connection refused" interval="200ms" May 8 07:59:03.797607 kubelet[2745]: W0508 07:59:03.796477 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.24.4.249:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.797607 kubelet[2745]: E0508 07:59:03.796556 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://172.24.4.249:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.798386 kubelet[2745]: I0508 07:59:03.798353 2745 factory.go:221] Registration of the systemd container factory successfully May 8 07:59:03.798702 kubelet[2745]: I0508 07:59:03.798503 2745 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 07:59:03.801483 kubelet[2745]: I0508 07:59:03.801447 2745 factory.go:221] Registration of the containerd container factory successfully May 8 07:59:03.803840 kubelet[2745]: I0508 07:59:03.803777 2745 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 07:59:03.805779 kubelet[2745]: I0508 07:59:03.805740 2745 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 07:59:03.805837 kubelet[2745]: I0508 07:59:03.805798 2745 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 07:59:03.805837 kubelet[2745]: I0508 07:59:03.805830 2745 kubelet.go:2337] "Starting kubelet main sync loop" May 8 07:59:03.805943 kubelet[2745]: E0508 07:59:03.805891 2745 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 07:59:03.807873 kubelet[2745]: E0508 07:59:03.807841 2745 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 8 07:59:03.817607 kubelet[2745]: W0508 07:59:03.817567 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.24.4.249:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.817906 kubelet[2745]: E0508 07:59:03.817757 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://172.24.4.249:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:03.833786 kubelet[2745]: I0508 07:59:03.833753 2745 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 07:59:03.833786 kubelet[2745]: I0508 07:59:03.833786 2745 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 07:59:03.833902 kubelet[2745]: I0508 07:59:03.833811 2745 state_mem.go:36] "Initialized new in-memory state store" May 8 07:59:03.842227 kubelet[2745]: I0508 07:59:03.842193 2745 policy_none.go:49] "None policy: Start" May 8 07:59:03.842994 kubelet[2745]: I0508 07:59:03.842800 2745 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 07:59:03.842994 kubelet[2745]: I0508 07:59:03.842819 2745 state_mem.go:35] "Initializing new in-memory state store" May 8 07:59:03.852678 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 8 07:59:03.861771 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 8 07:59:03.866659 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 8 07:59:03.876728 kubelet[2745]: I0508 07:59:03.876668 2745 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 07:59:03.877974 kubelet[2745]: I0508 07:59:03.877006 2745 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 07:59:03.877974 kubelet[2745]: I0508 07:59:03.877374 2745 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 07:59:03.883008 kubelet[2745]: E0508 07:59:03.882958 2745 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4327-0-0-w-45c5ab2731.novalocal\" not found" May 8 07:59:03.890270 kubelet[2745]: I0508 07:59:03.890235 2745 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:03.890909 kubelet[2745]: E0508 07:59:03.890864 2745 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://172.24.4.249:6443/api/v1/nodes\": dial tcp 172.24.4.249:6443: connect: connection refused" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:03.906108 kubelet[2745]: I0508 07:59:03.906074 2745 topology_manager.go:215] "Topology Admit Handler" podUID="5ec3363a12d537b798fb5d9dc8f34553" podNamespace="kube-system" podName="kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:03.907327 kubelet[2745]: I0508 07:59:03.907271 2745 topology_manager.go:215] "Topology Admit Handler" podUID="7c63c3f6ccd5045ea86465ab405588f0" podNamespace="kube-system" podName="kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:03.909240 kubelet[2745]: I0508 07:59:03.908866 2745 topology_manager.go:215] "Topology Admit Handler" podUID="6fda4e6c7d4a9ba1cd1eb28f215feb8c" podNamespace="kube-system" podName="kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:03.915513 systemd[1]: Created slice kubepods-burstable-pod5ec3363a12d537b798fb5d9dc8f34553.slice - libcontainer container kubepods-burstable-pod5ec3363a12d537b798fb5d9dc8f34553.slice. May 8 07:59:03.930310 systemd[1]: Created slice kubepods-burstable-pod7c63c3f6ccd5045ea86465ab405588f0.slice - libcontainer container kubepods-burstable-pod7c63c3f6ccd5045ea86465ab405588f0.slice. May 8 07:59:03.936522 systemd[1]: Created slice kubepods-burstable-pod6fda4e6c7d4a9ba1cd1eb28f215feb8c.slice - libcontainer container kubepods-burstable-pod6fda4e6c7d4a9ba1cd1eb28f215feb8c.slice. May 8 07:59:03.997234 kubelet[2745]: E0508 07:59:03.997074 2745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.24.4.249:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4327-0-0-w-45c5ab2731.novalocal?timeout=10s\": dial tcp 172.24.4.249:6443: connect: connection refused" interval="400ms" May 8 07:59:04.090644 kubelet[2745]: I0508 07:59:04.090495 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-kubeconfig\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.090644 kubelet[2745]: I0508 07:59:04.090611 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-ca-certs\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.090644 kubelet[2745]: I0508 07:59:04.090663 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-k8s-certs\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.090644 kubelet[2745]: I0508 07:59:04.090713 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.091326 kubelet[2745]: I0508 07:59:04.090761 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-ca-certs\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.091326 kubelet[2745]: I0508 07:59:04.090805 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-flexvolume-dir\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.091326 kubelet[2745]: I0508 07:59:04.090848 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-k8s-certs\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.091326 kubelet[2745]: I0508 07:59:04.090893 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.092137 kubelet[2745]: I0508 07:59:04.090983 2745 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6fda4e6c7d4a9ba1cd1eb28f215feb8c-kubeconfig\") pod \"kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"6fda4e6c7d4a9ba1cd1eb28f215feb8c\") " pod="kube-system/kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.094282 kubelet[2745]: I0508 07:59:04.094184 2745 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.094849 kubelet[2745]: E0508 07:59:04.094730 2745 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://172.24.4.249:6443/api/v1/nodes\": dial tcp 172.24.4.249:6443: connect: connection refused" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.228687 containerd[1535]: time="2025-05-08T07:59:04.228612187Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:5ec3363a12d537b798fb5d9dc8f34553,Namespace:kube-system,Attempt:0,}" May 8 07:59:04.236030 containerd[1535]: time="2025-05-08T07:59:04.235584221Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:7c63c3f6ccd5045ea86465ab405588f0,Namespace:kube-system,Attempt:0,}" May 8 07:59:04.241260 containerd[1535]: time="2025-05-08T07:59:04.240974421Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:6fda4e6c7d4a9ba1cd1eb28f215feb8c,Namespace:kube-system,Attempt:0,}" May 8 07:59:04.397742 kubelet[2745]: E0508 07:59:04.397627 2745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.24.4.249:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4327-0-0-w-45c5ab2731.novalocal?timeout=10s\": dial tcp 172.24.4.249:6443: connect: connection refused" interval="800ms" May 8 07:59:04.497556 kubelet[2745]: I0508 07:59:04.497491 2745 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.498103 kubelet[2745]: E0508 07:59:04.498055 2745 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://172.24.4.249:6443/api/v1/nodes\": dial tcp 172.24.4.249:6443: connect: connection refused" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:04.863389 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1090915687.mount: Deactivated successfully. May 8 07:59:04.877061 containerd[1535]: time="2025-05-08T07:59:04.876133612Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 07:59:04.881303 containerd[1535]: time="2025-05-08T07:59:04.881176480Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321146" May 8 07:59:04.887416 containerd[1535]: time="2025-05-08T07:59:04.886836885Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 07:59:04.889526 containerd[1535]: time="2025-05-08T07:59:04.889024362Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 07:59:04.889526 containerd[1535]: time="2025-05-08T07:59:04.889306821Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 8 07:59:04.893832 containerd[1535]: time="2025-05-08T07:59:04.893778622Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 07:59:04.895347 containerd[1535]: time="2025-05-08T07:59:04.895042299Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 8 07:59:04.896718 containerd[1535]: time="2025-05-08T07:59:04.896663498Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 8 07:59:04.901512 containerd[1535]: time="2025-05-08T07:59:04.900340384Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 648.732211ms" May 8 07:59:04.904363 containerd[1535]: time="2025-05-08T07:59:04.904285641Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 670.594376ms" May 8 07:59:04.913116 containerd[1535]: time="2025-05-08T07:59:04.913029661Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 673.597549ms" May 8 07:59:04.980011 containerd[1535]: time="2025-05-08T07:59:04.979445016Z" level=info msg="connecting to shim 2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338" address="unix:///run/containerd/s/e9bcf83816bd31577438b92c5ba80580d79c857e432cf526e64443d05a5352c9" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:05.003655 containerd[1535]: time="2025-05-08T07:59:05.003459131Z" level=info msg="connecting to shim 6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420" address="unix:///run/containerd/s/af7e3976f6f381d2c01fd2d37c4435fa8e08b95b208b768948f0d70983266602" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:05.008738 kubelet[2745]: W0508 07:59:05.008607 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.24.4.249:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.008738 kubelet[2745]: E0508 07:59:05.008690 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://172.24.4.249:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.016559 containerd[1535]: time="2025-05-08T07:59:05.016314749Z" level=info msg="connecting to shim 4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66" address="unix:///run/containerd/s/229416f4819e9bc43a127944cbae438cf8ef6ecb92f8a247cd9fa4bf38e9f6ac" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:05.036873 kubelet[2745]: W0508 07:59:05.036705 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.24.4.249:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4327-0-0-w-45c5ab2731.novalocal&limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.036873 kubelet[2745]: E0508 07:59:05.036781 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://172.24.4.249:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4327-0-0-w-45c5ab2731.novalocal&limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.038679 systemd[1]: Started cri-containerd-2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338.scope - libcontainer container 2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338. May 8 07:59:05.044323 systemd[1]: Started cri-containerd-6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420.scope - libcontainer container 6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420. May 8 07:59:05.051217 systemd[1]: Started cri-containerd-4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66.scope - libcontainer container 4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66. May 8 07:59:05.113152 containerd[1535]: time="2025-05-08T07:59:05.112572958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:7c63c3f6ccd5045ea86465ab405588f0,Namespace:kube-system,Attempt:0,} returns sandbox id \"4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66\"" May 8 07:59:05.121003 containerd[1535]: time="2025-05-08T07:59:05.120884486Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:5ec3363a12d537b798fb5d9dc8f34553,Namespace:kube-system,Attempt:0,} returns sandbox id \"2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338\"" May 8 07:59:05.123674 containerd[1535]: time="2025-05-08T07:59:05.123461768Z" level=info msg="CreateContainer within sandbox \"4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 8 07:59:05.127200 containerd[1535]: time="2025-05-08T07:59:05.127161608Z" level=info msg="CreateContainer within sandbox \"2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 8 07:59:05.145556 containerd[1535]: time="2025-05-08T07:59:05.145520913Z" level=info msg="Container 53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:05.149021 containerd[1535]: time="2025-05-08T07:59:05.148896326Z" level=info msg="Container 684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:05.150551 containerd[1535]: time="2025-05-08T07:59:05.150435238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal,Uid:6fda4e6c7d4a9ba1cd1eb28f215feb8c,Namespace:kube-system,Attempt:0,} returns sandbox id \"6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420\"" May 8 07:59:05.153440 containerd[1535]: time="2025-05-08T07:59:05.153399076Z" level=info msg="CreateContainer within sandbox \"6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 8 07:59:05.164855 kubelet[2745]: W0508 07:59:05.164790 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.24.4.249:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.164855 kubelet[2745]: E0508 07:59:05.164856 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://172.24.4.249:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.168760 containerd[1535]: time="2025-05-08T07:59:05.168720302Z" level=info msg="CreateContainer within sandbox \"2da47e97fc65276358410f5ad5152084c692a7505d39b1f4bb30ee2c73d33338\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec\"" May 8 07:59:05.177682 containerd[1535]: time="2025-05-08T07:59:05.177617897Z" level=info msg="StartContainer for \"684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec\"" May 8 07:59:05.178836 containerd[1535]: time="2025-05-08T07:59:05.178797485Z" level=info msg="connecting to shim 684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec" address="unix:///run/containerd/s/e9bcf83816bd31577438b92c5ba80580d79c857e432cf526e64443d05a5352c9" protocol=ttrpc version=3 May 8 07:59:05.180364 containerd[1535]: time="2025-05-08T07:59:05.180135986Z" level=info msg="CreateContainer within sandbox \"4e3f925a7f618beeb2c0abee3c85daa09206077b0944863a30e73f310e7a6c66\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae\"" May 8 07:59:05.181361 containerd[1535]: time="2025-05-08T07:59:05.180876487Z" level=info msg="StartContainer for \"53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae\"" May 8 07:59:05.181551 containerd[1535]: time="2025-05-08T07:59:05.181332846Z" level=info msg="Container 5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:05.182402 kubelet[2745]: W0508 07:59:05.182356 2745 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.24.4.249:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.183166 containerd[1535]: time="2025-05-08T07:59:05.182893019Z" level=info msg="connecting to shim 53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae" address="unix:///run/containerd/s/229416f4819e9bc43a127944cbae438cf8ef6ecb92f8a247cd9fa4bf38e9f6ac" protocol=ttrpc version=3 May 8 07:59:05.183223 kubelet[2745]: E0508 07:59:05.183147 2745 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://172.24.4.249:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.24.4.249:6443: connect: connection refused May 8 07:59:05.198936 kubelet[2745]: E0508 07:59:05.198843 2745 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.24.4.249:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4327-0-0-w-45c5ab2731.novalocal?timeout=10s\": dial tcp 172.24.4.249:6443: connect: connection refused" interval="1.6s" May 8 07:59:05.205670 containerd[1535]: time="2025-05-08T07:59:05.205545195Z" level=info msg="CreateContainer within sandbox \"6bce5b73d899748f62713d7865098da77a9356c34f1d305250f8c33b2f401420\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772\"" May 8 07:59:05.206415 containerd[1535]: time="2025-05-08T07:59:05.206346622Z" level=info msg="StartContainer for \"5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772\"" May 8 07:59:05.207425 containerd[1535]: time="2025-05-08T07:59:05.207396162Z" level=info msg="connecting to shim 5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772" address="unix:///run/containerd/s/af7e3976f6f381d2c01fd2d37c4435fa8e08b95b208b768948f0d70983266602" protocol=ttrpc version=3 May 8 07:59:05.208088 systemd[1]: Started cri-containerd-684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec.scope - libcontainer container 684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec. May 8 07:59:05.212935 systemd[1]: Started cri-containerd-53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae.scope - libcontainer container 53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae. May 8 07:59:05.232368 systemd[1]: Started cri-containerd-5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772.scope - libcontainer container 5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772. May 8 07:59:05.296696 containerd[1535]: time="2025-05-08T07:59:05.296059087Z" level=info msg="StartContainer for \"684aa2baf825c2d2bd246b3e651fe14617a2d970d5af0356ccc6079f8ae8c2ec\" returns successfully" May 8 07:59:05.302336 kubelet[2745]: I0508 07:59:05.302029 2745 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:05.302598 kubelet[2745]: E0508 07:59:05.302555 2745 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://172.24.4.249:6443/api/v1/nodes\": dial tcp 172.24.4.249:6443: connect: connection refused" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:05.303978 containerd[1535]: time="2025-05-08T07:59:05.303908024Z" level=info msg="StartContainer for \"53f7239157636f289f5640aab49a156a8875df390b399fe22146d920c586faae\" returns successfully" May 8 07:59:05.336743 containerd[1535]: time="2025-05-08T07:59:05.336684903Z" level=info msg="StartContainer for \"5d620fddaad5f6d6469224469ed974695e9fd88d164bc0ed243632b841bb3772\" returns successfully" May 8 07:59:06.906051 kubelet[2745]: I0508 07:59:06.905651 2745 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:07.194330 kubelet[2745]: E0508 07:59:07.194199 2745 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4327-0-0-w-45c5ab2731.novalocal\" not found" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:07.276902 kubelet[2745]: I0508 07:59:07.276799 2745 kubelet_node_status.go:76] "Successfully registered node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:07.773886 kubelet[2745]: I0508 07:59:07.771712 2745 apiserver.go:52] "Watching apiserver" May 8 07:59:07.790501 kubelet[2745]: I0508 07:59:07.790421 2745 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 8 07:59:07.873768 kubelet[2745]: E0508 07:59:07.872670 2745 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:09.889336 systemd[1]: Reload requested from client PID 3018 ('systemctl') (unit session-9.scope)... May 8 07:59:09.889374 systemd[1]: Reloading... May 8 07:59:10.019984 zram_generator::config[3063]: No configuration found. May 8 07:59:10.143882 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 07:59:10.303492 systemd[1]: Reloading finished in 413 ms. May 8 07:59:10.332560 kubelet[2745]: I0508 07:59:10.332296 2745 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 07:59:10.332311 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:59:10.336624 systemd[1]: kubelet.service: Deactivated successfully. May 8 07:59:10.336821 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:10.336858 systemd[1]: kubelet.service: Consumed 1.544s CPU time, 115.1M memory peak. May 8 07:59:10.339307 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 8 07:59:10.455686 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 8 07:59:10.463270 (kubelet)[3127]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 8 07:59:10.533309 kubelet[3127]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 07:59:10.533309 kubelet[3127]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 07:59:10.533309 kubelet[3127]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 07:59:10.533724 kubelet[3127]: I0508 07:59:10.533429 3127 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 07:59:10.539086 kubelet[3127]: I0508 07:59:10.539059 3127 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 07:59:10.539086 kubelet[3127]: I0508 07:59:10.539079 3127 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 07:59:10.539245 kubelet[3127]: I0508 07:59:10.539230 3127 server.go:927] "Client rotation is on, will bootstrap in background" May 8 07:59:10.540515 kubelet[3127]: I0508 07:59:10.540500 3127 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 8 07:59:10.543938 kubelet[3127]: I0508 07:59:10.542761 3127 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 07:59:10.550113 kubelet[3127]: I0508 07:59:10.550094 3127 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 07:59:10.550288 kubelet[3127]: I0508 07:59:10.550263 3127 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 07:59:10.550461 kubelet[3127]: I0508 07:59:10.550290 3127 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4327-0-0-w-45c5ab2731.novalocal","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 07:59:10.550593 kubelet[3127]: I0508 07:59:10.550466 3127 topology_manager.go:138] "Creating topology manager with none policy" May 8 07:59:10.550593 kubelet[3127]: I0508 07:59:10.550478 3127 container_manager_linux.go:301] "Creating device plugin manager" May 8 07:59:10.550593 kubelet[3127]: I0508 07:59:10.550510 3127 state_mem.go:36] "Initialized new in-memory state store" May 8 07:59:10.550593 kubelet[3127]: I0508 07:59:10.550584 3127 kubelet.go:400] "Attempting to sync node with API server" May 8 07:59:10.551269 kubelet[3127]: I0508 07:59:10.550602 3127 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 07:59:10.551269 kubelet[3127]: I0508 07:59:10.550620 3127 kubelet.go:312] "Adding apiserver pod source" May 8 07:59:10.551269 kubelet[3127]: I0508 07:59:10.550634 3127 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 07:59:10.551615 kubelet[3127]: I0508 07:59:10.551597 3127 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 8 07:59:10.551873 kubelet[3127]: I0508 07:59:10.551860 3127 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 07:59:10.552328 kubelet[3127]: I0508 07:59:10.552314 3127 server.go:1264] "Started kubelet" May 8 07:59:10.557962 kubelet[3127]: I0508 07:59:10.556717 3127 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 07:59:10.564452 kubelet[3127]: E0508 07:59:10.564353 3127 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 8 07:59:10.566292 kubelet[3127]: I0508 07:59:10.566268 3127 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 07:59:10.569015 kubelet[3127]: I0508 07:59:10.568234 3127 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 07:59:10.573064 kubelet[3127]: I0508 07:59:10.573040 3127 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 07:59:10.573064 kubelet[3127]: I0508 07:59:10.570145 3127 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 07:59:10.573969 kubelet[3127]: I0508 07:59:10.569274 3127 server.go:455] "Adding debug handlers to kubelet server" May 8 07:59:10.574861 kubelet[3127]: I0508 07:59:10.570157 3127 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 07:59:10.574943 kubelet[3127]: I0508 07:59:10.574934 3127 reconciler.go:26] "Reconciler: start to sync state" May 8 07:59:10.576334 kubelet[3127]: I0508 07:59:10.576222 3127 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 07:59:10.579574 kubelet[3127]: I0508 07:59:10.579403 3127 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 07:59:10.588945 kubelet[3127]: I0508 07:59:10.582698 3127 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 07:59:10.588945 kubelet[3127]: I0508 07:59:10.582739 3127 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 07:59:10.588945 kubelet[3127]: I0508 07:59:10.582755 3127 kubelet.go:2337] "Starting kubelet main sync loop" May 8 07:59:10.588945 kubelet[3127]: E0508 07:59:10.582790 3127 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 07:59:10.588945 kubelet[3127]: I0508 07:59:10.583129 3127 factory.go:221] Registration of the containerd container factory successfully May 8 07:59:10.588945 kubelet[3127]: I0508 07:59:10.583141 3127 factory.go:221] Registration of the systemd container factory successfully May 8 07:59:10.625940 kubelet[3127]: I0508 07:59:10.625878 3127 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 07:59:10.625940 kubelet[3127]: I0508 07:59:10.625893 3127 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 07:59:10.625940 kubelet[3127]: I0508 07:59:10.625908 3127 state_mem.go:36] "Initialized new in-memory state store" May 8 07:59:10.626123 kubelet[3127]: I0508 07:59:10.626061 3127 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 8 07:59:10.626123 kubelet[3127]: I0508 07:59:10.626073 3127 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 8 07:59:10.626123 kubelet[3127]: I0508 07:59:10.626090 3127 policy_none.go:49] "None policy: Start" May 8 07:59:10.626782 kubelet[3127]: I0508 07:59:10.626768 3127 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 07:59:10.626827 kubelet[3127]: I0508 07:59:10.626787 3127 state_mem.go:35] "Initializing new in-memory state store" May 8 07:59:10.626911 kubelet[3127]: I0508 07:59:10.626898 3127 state_mem.go:75] "Updated machine memory state" May 8 07:59:10.631195 kubelet[3127]: I0508 07:59:10.631167 3127 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 07:59:10.631332 kubelet[3127]: I0508 07:59:10.631302 3127 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 07:59:10.631398 kubelet[3127]: I0508 07:59:10.631385 3127 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 07:59:10.673555 kubelet[3127]: I0508 07:59:10.673522 3127 kubelet_node_status.go:73] "Attempting to register node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.683929 kubelet[3127]: I0508 07:59:10.683881 3127 topology_manager.go:215] "Topology Admit Handler" podUID="6fda4e6c7d4a9ba1cd1eb28f215feb8c" podNamespace="kube-system" podName="kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.684671 kubelet[3127]: I0508 07:59:10.684012 3127 topology_manager.go:215] "Topology Admit Handler" podUID="5ec3363a12d537b798fb5d9dc8f34553" podNamespace="kube-system" podName="kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.684671 kubelet[3127]: I0508 07:59:10.684050 3127 topology_manager.go:215] "Topology Admit Handler" podUID="7c63c3f6ccd5045ea86465ab405588f0" podNamespace="kube-system" podName="kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.699484 kubelet[3127]: W0508 07:59:10.699449 3127 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 8 07:59:10.700099 kubelet[3127]: W0508 07:59:10.700005 3127 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 8 07:59:10.700266 kubelet[3127]: W0508 07:59:10.700254 3127 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 8 07:59:10.704949 kubelet[3127]: I0508 07:59:10.704236 3127 kubelet_node_status.go:112] "Node was previously registered" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.704949 kubelet[3127]: I0508 07:59:10.704358 3127 kubelet_node_status.go:76] "Successfully registered node" node="ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876448 kubelet[3127]: I0508 07:59:10.876364 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-k8s-certs\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876448 kubelet[3127]: I0508 07:59:10.876439 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876738 kubelet[3127]: I0508 07:59:10.876467 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-flexvolume-dir\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876738 kubelet[3127]: I0508 07:59:10.876498 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5ec3363a12d537b798fb5d9dc8f34553-ca-certs\") pod \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"5ec3363a12d537b798fb5d9dc8f34553\") " pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876738 kubelet[3127]: I0508 07:59:10.876520 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-ca-certs\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.876738 kubelet[3127]: I0508 07:59:10.876541 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-k8s-certs\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.877143 kubelet[3127]: I0508 07:59:10.876561 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-kubeconfig\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.877143 kubelet[3127]: I0508 07:59:10.876583 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7c63c3f6ccd5045ea86465ab405588f0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"7c63c3f6ccd5045ea86465ab405588f0\") " pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.877143 kubelet[3127]: I0508 07:59:10.876607 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6fda4e6c7d4a9ba1cd1eb28f215feb8c-kubeconfig\") pod \"kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal\" (UID: \"6fda4e6c7d4a9ba1cd1eb28f215feb8c\") " pod="kube-system/kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:10.879119 sudo[3160]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 8 07:59:10.880576 sudo[3160]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 8 07:59:11.484126 sudo[3160]: pam_unix(sudo:session): session closed for user root May 8 07:59:11.551744 kubelet[3127]: I0508 07:59:11.551691 3127 apiserver.go:52] "Watching apiserver" May 8 07:59:11.575639 kubelet[3127]: I0508 07:59:11.575595 3127 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 8 07:59:11.627939 kubelet[3127]: W0508 07:59:11.625257 3127 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 8 07:59:11.627939 kubelet[3127]: E0508 07:59:11.625325 3127 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal\" already exists" pod="kube-system/kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:11.627939 kubelet[3127]: W0508 07:59:11.626095 3127 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 8 07:59:11.627939 kubelet[3127]: E0508 07:59:11.626153 3127 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal\" already exists" pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" May 8 07:59:11.663045 kubelet[3127]: I0508 07:59:11.662391 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4327-0-0-w-45c5ab2731.novalocal" podStartSLOduration=1.6623742099999999 podStartE2EDuration="1.66237421s" podCreationTimestamp="2025-05-08 07:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:11.662214475 +0000 UTC m=+1.192877181" watchObservedRunningTime="2025-05-08 07:59:11.66237421 +0000 UTC m=+1.193036906" May 8 07:59:11.665899 kubelet[3127]: I0508 07:59:11.665525 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4327-0-0-w-45c5ab2731.novalocal" podStartSLOduration=1.6655096889999998 podStartE2EDuration="1.665509689s" podCreationTimestamp="2025-05-08 07:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:11.648750451 +0000 UTC m=+1.179413147" watchObservedRunningTime="2025-05-08 07:59:11.665509689 +0000 UTC m=+1.196172395" May 8 07:59:11.688753 kubelet[3127]: I0508 07:59:11.688471 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4327-0-0-w-45c5ab2731.novalocal" podStartSLOduration=1.688451105 podStartE2EDuration="1.688451105s" podCreationTimestamp="2025-05-08 07:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:11.676754752 +0000 UTC m=+1.207417458" watchObservedRunningTime="2025-05-08 07:59:11.688451105 +0000 UTC m=+1.219113801" May 8 07:59:13.974233 sudo[2083]: pam_unix(sudo:session): session closed for user root May 8 07:59:14.253973 sshd[2082]: Connection closed by 172.24.4.1 port 41080 May 8 07:59:14.254913 sshd-session[2064]: pam_unix(sshd:session): session closed for user core May 8 07:59:14.262259 systemd[1]: sshd@6-172.24.4.249:22-172.24.4.1:41080.service: Deactivated successfully. May 8 07:59:14.268463 systemd[1]: session-9.scope: Deactivated successfully. May 8 07:59:14.268968 systemd[1]: session-9.scope: Consumed 7.647s CPU time, 293.8M memory peak. May 8 07:59:14.274066 systemd-logind[1509]: Session 9 logged out. Waiting for processes to exit. May 8 07:59:14.276774 systemd-logind[1509]: Removed session 9. May 8 07:59:25.130303 kubelet[3127]: I0508 07:59:25.129794 3127 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 8 07:59:25.131176 containerd[1535]: time="2025-05-08T07:59:25.130626077Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 8 07:59:25.132452 kubelet[3127]: I0508 07:59:25.132107 3127 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 8 07:59:25.805010 kubelet[3127]: I0508 07:59:25.804904 3127 topology_manager.go:215] "Topology Admit Handler" podUID="3b37f782-66d0-415c-95f5-8435cd9f6fab" podNamespace="kube-system" podName="kube-proxy-mlztw" May 8 07:59:25.831346 systemd[1]: Created slice kubepods-besteffort-pod3b37f782_66d0_415c_95f5_8435cd9f6fab.slice - libcontainer container kubepods-besteffort-pod3b37f782_66d0_415c_95f5_8435cd9f6fab.slice. May 8 07:59:25.838223 kubelet[3127]: I0508 07:59:25.838186 3127 topology_manager.go:215] "Topology Admit Handler" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" podNamespace="kube-system" podName="cilium-xlz5j" May 8 07:59:25.850144 systemd[1]: Created slice kubepods-burstable-pod01da910d_c3da_436f_ae6b_18a93de1a297.slice - libcontainer container kubepods-burstable-pod01da910d_c3da_436f_ae6b_18a93de1a297.slice. May 8 07:59:25.878469 kubelet[3127]: I0508 07:59:25.878419 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-hubble-tls\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.879055 kubelet[3127]: I0508 07:59:25.878751 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-etc-cni-netd\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.879055 kubelet[3127]: I0508 07:59:25.878782 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-lib-modules\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.879055 kubelet[3127]: I0508 07:59:25.878824 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz6r4\" (UniqueName: \"kubernetes.io/projected/3b37f782-66d0-415c-95f5-8435cd9f6fab-kube-api-access-mz6r4\") pod \"kube-proxy-mlztw\" (UID: \"3b37f782-66d0-415c-95f5-8435cd9f6fab\") " pod="kube-system/kube-proxy-mlztw" May 8 07:59:25.879055 kubelet[3127]: I0508 07:59:25.878850 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-bpf-maps\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.880847 kubelet[3127]: I0508 07:59:25.878996 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-xtables-lock\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.880847 kubelet[3127]: I0508 07:59:25.879165 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3b37f782-66d0-415c-95f5-8435cd9f6fab-kube-proxy\") pod \"kube-proxy-mlztw\" (UID: \"3b37f782-66d0-415c-95f5-8435cd9f6fab\") " pod="kube-system/kube-proxy-mlztw" May 8 07:59:25.880847 kubelet[3127]: I0508 07:59:25.879209 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r98vr\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-kube-api-access-r98vr\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.880847 kubelet[3127]: I0508 07:59:25.879298 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cni-path\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.880847 kubelet[3127]: I0508 07:59:25.879336 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-kernel\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881017 kubelet[3127]: I0508 07:59:25.879377 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-config-path\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881017 kubelet[3127]: I0508 07:59:25.879418 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3b37f782-66d0-415c-95f5-8435cd9f6fab-xtables-lock\") pod \"kube-proxy-mlztw\" (UID: \"3b37f782-66d0-415c-95f5-8435cd9f6fab\") " pod="kube-system/kube-proxy-mlztw" May 8 07:59:25.881017 kubelet[3127]: I0508 07:59:25.879440 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-net\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881017 kubelet[3127]: I0508 07:59:25.879485 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b37f782-66d0-415c-95f5-8435cd9f6fab-lib-modules\") pod \"kube-proxy-mlztw\" (UID: \"3b37f782-66d0-415c-95f5-8435cd9f6fab\") " pod="kube-system/kube-proxy-mlztw" May 8 07:59:25.881017 kubelet[3127]: I0508 07:59:25.879510 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/01da910d-c3da-436f-ae6b-18a93de1a297-clustermesh-secrets\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881161 kubelet[3127]: I0508 07:59:25.879531 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-cgroup\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881161 kubelet[3127]: I0508 07:59:25.879586 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-run\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:25.881161 kubelet[3127]: I0508 07:59:25.879609 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-hostproc\") pod \"cilium-xlz5j\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " pod="kube-system/cilium-xlz5j" May 8 07:59:26.145416 containerd[1535]: time="2025-05-08T07:59:26.145358463Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-mlztw,Uid:3b37f782-66d0-415c-95f5-8435cd9f6fab,Namespace:kube-system,Attempt:0,}" May 8 07:59:26.159282 containerd[1535]: time="2025-05-08T07:59:26.159024694Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xlz5j,Uid:01da910d-c3da-436f-ae6b-18a93de1a297,Namespace:kube-system,Attempt:0,}" May 8 07:59:26.197061 kubelet[3127]: I0508 07:59:26.197010 3127 topology_manager.go:215] "Topology Admit Handler" podUID="60355b88-43ea-493c-a59c-b9a39cecea05" podNamespace="kube-system" podName="cilium-operator-599987898-hnrs9" May 8 07:59:26.213054 systemd[1]: Created slice kubepods-besteffort-pod60355b88_43ea_493c_a59c_b9a39cecea05.slice - libcontainer container kubepods-besteffort-pod60355b88_43ea_493c_a59c_b9a39cecea05.slice. May 8 07:59:26.227419 containerd[1535]: time="2025-05-08T07:59:26.227090126Z" level=info msg="connecting to shim 87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369" address="unix:///run/containerd/s/5ac238c8d733e89919a7faf6a55778021d6fabbd7450ce0068628dd4d7be3f91" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:26.234884 containerd[1535]: time="2025-05-08T07:59:26.234840322Z" level=info msg="connecting to shim 84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:26.274072 systemd[1]: Started cri-containerd-87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369.scope - libcontainer container 87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369. May 8 07:59:26.278048 systemd[1]: Started cri-containerd-84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b.scope - libcontainer container 84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b. May 8 07:59:26.285123 kubelet[3127]: I0508 07:59:26.285042 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/60355b88-43ea-493c-a59c-b9a39cecea05-cilium-config-path\") pod \"cilium-operator-599987898-hnrs9\" (UID: \"60355b88-43ea-493c-a59c-b9a39cecea05\") " pod="kube-system/cilium-operator-599987898-hnrs9" May 8 07:59:26.285123 kubelet[3127]: I0508 07:59:26.285078 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbtb\" (UniqueName: \"kubernetes.io/projected/60355b88-43ea-493c-a59c-b9a39cecea05-kube-api-access-qtbtb\") pod \"cilium-operator-599987898-hnrs9\" (UID: \"60355b88-43ea-493c-a59c-b9a39cecea05\") " pod="kube-system/cilium-operator-599987898-hnrs9" May 8 07:59:26.320583 containerd[1535]: time="2025-05-08T07:59:26.320363976Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-mlztw,Uid:3b37f782-66d0-415c-95f5-8435cd9f6fab,Namespace:kube-system,Attempt:0,} returns sandbox id \"87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369\"" May 8 07:59:26.324535 containerd[1535]: time="2025-05-08T07:59:26.323903574Z" level=info msg="CreateContainer within sandbox \"87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 8 07:59:26.327355 containerd[1535]: time="2025-05-08T07:59:26.327038399Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xlz5j,Uid:01da910d-c3da-436f-ae6b-18a93de1a297,Namespace:kube-system,Attempt:0,} returns sandbox id \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\"" May 8 07:59:26.330006 containerd[1535]: time="2025-05-08T07:59:26.329662639Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 8 07:59:26.342757 containerd[1535]: time="2025-05-08T07:59:26.342724397Z" level=info msg="Container 63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:26.359519 containerd[1535]: time="2025-05-08T07:59:26.359488443Z" level=info msg="CreateContainer within sandbox \"87c15d53aa9165e26c034aef7e39313c5107e02766a122956a303c608e460369\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806\"" May 8 07:59:26.360120 containerd[1535]: time="2025-05-08T07:59:26.360099059Z" level=info msg="StartContainer for \"63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806\"" May 8 07:59:26.367711 containerd[1535]: time="2025-05-08T07:59:26.367033829Z" level=info msg="connecting to shim 63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806" address="unix:///run/containerd/s/5ac238c8d733e89919a7faf6a55778021d6fabbd7450ce0068628dd4d7be3f91" protocol=ttrpc version=3 May 8 07:59:26.386070 systemd[1]: Started cri-containerd-63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806.scope - libcontainer container 63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806. May 8 07:59:26.445313 containerd[1535]: time="2025-05-08T07:59:26.443102289Z" level=info msg="StartContainer for \"63dd2f473cfc678b30c8eb209a040dd2da7784274480db60790f777254362806\" returns successfully" May 8 07:59:26.524130 containerd[1535]: time="2025-05-08T07:59:26.524063783Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-hnrs9,Uid:60355b88-43ea-493c-a59c-b9a39cecea05,Namespace:kube-system,Attempt:0,}" May 8 07:59:26.554994 containerd[1535]: time="2025-05-08T07:59:26.554459014Z" level=info msg="connecting to shim 7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb" address="unix:///run/containerd/s/a150d2af03cbc775cf951bda31cfd1da96435dfdbe30739349e82570a3705ccb" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:26.580096 systemd[1]: Started cri-containerd-7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb.scope - libcontainer container 7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb. May 8 07:59:26.628716 containerd[1535]: time="2025-05-08T07:59:26.628616116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-hnrs9,Uid:60355b88-43ea-493c-a59c-b9a39cecea05,Namespace:kube-system,Attempt:0,} returns sandbox id \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\"" May 8 07:59:26.670477 kubelet[3127]: I0508 07:59:26.670080 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-mlztw" podStartSLOduration=1.670061564 podStartE2EDuration="1.670061564s" podCreationTimestamp="2025-05-08 07:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:26.669542613 +0000 UTC m=+16.200205309" watchObservedRunningTime="2025-05-08 07:59:26.670061564 +0000 UTC m=+16.200724260" May 8 07:59:33.110784 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount566557022.mount: Deactivated successfully. May 8 07:59:35.336728 containerd[1535]: time="2025-05-08T07:59:35.336678540Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:59:35.338230 containerd[1535]: time="2025-05-08T07:59:35.338049981Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" May 8 07:59:35.339613 containerd[1535]: time="2025-05-08T07:59:35.339539970Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:59:35.341142 containerd[1535]: time="2025-05-08T07:59:35.341039017Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 9.011343236s" May 8 07:59:35.341142 containerd[1535]: time="2025-05-08T07:59:35.341068451Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 8 07:59:35.342454 containerd[1535]: time="2025-05-08T07:59:35.342433662Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 8 07:59:35.344087 containerd[1535]: time="2025-05-08T07:59:35.344033012Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 8 07:59:35.357935 containerd[1535]: time="2025-05-08T07:59:35.355727166Z" level=info msg="Container 5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:35.369322 containerd[1535]: time="2025-05-08T07:59:35.369250262Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\"" May 8 07:59:35.370035 containerd[1535]: time="2025-05-08T07:59:35.369802478Z" level=info msg="StartContainer for \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\"" May 8 07:59:35.370756 containerd[1535]: time="2025-05-08T07:59:35.370735373Z" level=info msg="connecting to shim 5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" protocol=ttrpc version=3 May 8 07:59:35.393069 systemd[1]: Started cri-containerd-5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c.scope - libcontainer container 5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c. May 8 07:59:35.429398 containerd[1535]: time="2025-05-08T07:59:35.429364030Z" level=info msg="StartContainer for \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" returns successfully" May 8 07:59:35.438150 systemd[1]: cri-containerd-5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c.scope: Deactivated successfully. May 8 07:59:35.439007 containerd[1535]: time="2025-05-08T07:59:35.438979561Z" level=info msg="received exit event container_id:\"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" id:\"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" pid:3532 exited_at:{seconds:1746691175 nanos:438032188}" May 8 07:59:35.439911 containerd[1535]: time="2025-05-08T07:59:35.439715695Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" id:\"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" pid:3532 exited_at:{seconds:1746691175 nanos:438032188}" May 8 07:59:35.459897 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c-rootfs.mount: Deactivated successfully. May 8 07:59:37.700989 containerd[1535]: time="2025-05-08T07:59:37.697465834Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 8 07:59:37.729973 containerd[1535]: time="2025-05-08T07:59:37.727246846Z" level=info msg="Container 03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:37.747985 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1917299893.mount: Deactivated successfully. May 8 07:59:37.758107 containerd[1535]: time="2025-05-08T07:59:37.758024883Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\"" May 8 07:59:37.759713 containerd[1535]: time="2025-05-08T07:59:37.759621221Z" level=info msg="StartContainer for \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\"" May 8 07:59:37.762740 containerd[1535]: time="2025-05-08T07:59:37.762661188Z" level=info msg="connecting to shim 03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" protocol=ttrpc version=3 May 8 07:59:37.794080 systemd[1]: Started cri-containerd-03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106.scope - libcontainer container 03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106. May 8 07:59:37.834364 containerd[1535]: time="2025-05-08T07:59:37.834234813Z" level=info msg="StartContainer for \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" returns successfully" May 8 07:59:37.840826 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 8 07:59:37.841239 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 8 07:59:37.841555 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 8 07:59:37.845251 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 8 07:59:37.847556 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 8 07:59:37.849795 systemd[1]: cri-containerd-03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106.scope: Deactivated successfully. May 8 07:59:37.852383 containerd[1535]: time="2025-05-08T07:59:37.852325665Z" level=info msg="TaskExit event in podsandbox handler container_id:\"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" id:\"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" pid:3576 exited_at:{seconds:1746691177 nanos:850544507}" May 8 07:59:37.852580 containerd[1535]: time="2025-05-08T07:59:37.852551481Z" level=info msg="received exit event container_id:\"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" id:\"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" pid:3576 exited_at:{seconds:1746691177 nanos:850544507}" May 8 07:59:37.875121 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 8 07:59:38.729534 containerd[1535]: time="2025-05-08T07:59:38.729469956Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 8 07:59:38.730367 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106-rootfs.mount: Deactivated successfully. May 8 07:59:38.763510 containerd[1535]: time="2025-05-08T07:59:38.763479907Z" level=info msg="Container a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:38.784689 containerd[1535]: time="2025-05-08T07:59:38.784643436Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\"" May 8 07:59:38.787289 containerd[1535]: time="2025-05-08T07:59:38.787140275Z" level=info msg="StartContainer for \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\"" May 8 07:59:38.789341 containerd[1535]: time="2025-05-08T07:59:38.789270278Z" level=info msg="connecting to shim a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" protocol=ttrpc version=3 May 8 07:59:38.820069 systemd[1]: Started cri-containerd-a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3.scope - libcontainer container a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3. May 8 07:59:38.872223 systemd[1]: cri-containerd-a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3.scope: Deactivated successfully. May 8 07:59:38.873342 containerd[1535]: time="2025-05-08T07:59:38.873307220Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" id:\"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" pid:3631 exited_at:{seconds:1746691178 nanos:872048691}" May 8 07:59:38.878691 containerd[1535]: time="2025-05-08T07:59:38.878512837Z" level=info msg="received exit event container_id:\"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" id:\"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" pid:3631 exited_at:{seconds:1746691178 nanos:872048691}" May 8 07:59:38.882606 containerd[1535]: time="2025-05-08T07:59:38.880293848Z" level=info msg="StartContainer for \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" returns successfully" May 8 07:59:38.918058 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3-rootfs.mount: Deactivated successfully. May 8 07:59:39.416030 containerd[1535]: time="2025-05-08T07:59:39.415981588Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:59:39.417268 containerd[1535]: time="2025-05-08T07:59:39.417064533Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" May 8 07:59:39.418342 containerd[1535]: time="2025-05-08T07:59:39.418286515Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 8 07:59:39.419833 containerd[1535]: time="2025-05-08T07:59:39.419598373Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 4.076816089s" May 8 07:59:39.419833 containerd[1535]: time="2025-05-08T07:59:39.419634650Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 8 07:59:39.423057 containerd[1535]: time="2025-05-08T07:59:39.423017273Z" level=info msg="CreateContainer within sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 8 07:59:39.430958 containerd[1535]: time="2025-05-08T07:59:39.430788464Z" level=info msg="Container 92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:39.447330 containerd[1535]: time="2025-05-08T07:59:39.447251632Z" level=info msg="CreateContainer within sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\"" May 8 07:59:39.448877 containerd[1535]: time="2025-05-08T07:59:39.448850128Z" level=info msg="StartContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\"" May 8 07:59:39.449890 containerd[1535]: time="2025-05-08T07:59:39.449862434Z" level=info msg="connecting to shim 92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f" address="unix:///run/containerd/s/a150d2af03cbc775cf951bda31cfd1da96435dfdbe30739349e82570a3705ccb" protocol=ttrpc version=3 May 8 07:59:39.468055 systemd[1]: Started cri-containerd-92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f.scope - libcontainer container 92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f. May 8 07:59:39.503948 containerd[1535]: time="2025-05-08T07:59:39.503726424Z" level=info msg="StartContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" returns successfully" May 8 07:59:39.730674 containerd[1535]: time="2025-05-08T07:59:39.730574321Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 8 07:59:39.749367 containerd[1535]: time="2025-05-08T07:59:39.748609396Z" level=info msg="Container d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:39.757694 containerd[1535]: time="2025-05-08T07:59:39.757608469Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\"" May 8 07:59:39.759611 containerd[1535]: time="2025-05-08T07:59:39.759585572Z" level=info msg="StartContainer for \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\"" May 8 07:59:39.760965 containerd[1535]: time="2025-05-08T07:59:39.760555901Z" level=info msg="connecting to shim d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" protocol=ttrpc version=3 May 8 07:59:39.797409 kubelet[3127]: I0508 07:59:39.797340 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-hnrs9" podStartSLOduration=1.007500724 podStartE2EDuration="13.797313008s" podCreationTimestamp="2025-05-08 07:59:26 +0000 UTC" firstStartedPulling="2025-05-08 07:59:26.63083379 +0000 UTC m=+16.161496496" lastFinishedPulling="2025-05-08 07:59:39.420646084 +0000 UTC m=+28.951308780" observedRunningTime="2025-05-08 07:59:39.796374087 +0000 UTC m=+29.327036803" watchObservedRunningTime="2025-05-08 07:59:39.797313008 +0000 UTC m=+29.327975714" May 8 07:59:39.807603 systemd[1]: Started cri-containerd-d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8.scope - libcontainer container d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8. May 8 07:59:39.858026 systemd[1]: cri-containerd-d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8.scope: Deactivated successfully. May 8 07:59:39.863282 containerd[1535]: time="2025-05-08T07:59:39.862698961Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" id:\"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" pid:3708 exited_at:{seconds:1746691179 nanos:860063093}" May 8 07:59:39.863282 containerd[1535]: time="2025-05-08T07:59:39.862980479Z" level=info msg="received exit event container_id:\"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" id:\"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" pid:3708 exited_at:{seconds:1746691179 nanos:860063093}" May 8 07:59:39.880047 containerd[1535]: time="2025-05-08T07:59:39.879935304Z" level=info msg="StartContainer for \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" returns successfully" May 8 07:59:39.904703 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8-rootfs.mount: Deactivated successfully. May 8 07:59:40.742883 containerd[1535]: time="2025-05-08T07:59:40.741836110Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 8 07:59:40.764538 containerd[1535]: time="2025-05-08T07:59:40.764504505Z" level=info msg="Container 81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:40.785030 containerd[1535]: time="2025-05-08T07:59:40.784992831Z" level=info msg="CreateContainer within sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\"" May 8 07:59:40.786014 containerd[1535]: time="2025-05-08T07:59:40.785972407Z" level=info msg="StartContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\"" May 8 07:59:40.788040 containerd[1535]: time="2025-05-08T07:59:40.787902556Z" level=info msg="connecting to shim 81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff" address="unix:///run/containerd/s/be4b88b0d0f7be207b14a4f747be0d2d1123171ef53a109263a81f0ab3e8fd99" protocol=ttrpc version=3 May 8 07:59:40.828318 systemd[1]: Started cri-containerd-81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff.scope - libcontainer container 81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff. May 8 07:59:40.869246 containerd[1535]: time="2025-05-08T07:59:40.869067365Z" level=info msg="StartContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" returns successfully" May 8 07:59:40.928835 containerd[1535]: time="2025-05-08T07:59:40.928599788Z" level=info msg="TaskExit event in podsandbox handler container_id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" id:\"3ddc18def721b0a52242a75f675ebd9c697d7441385b9b9312bbf16c5e0aef14\" pid:3773 exited_at:{seconds:1746691180 nanos:928255203}" May 8 07:59:40.967320 kubelet[3127]: I0508 07:59:40.967131 3127 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 8 07:59:41.010999 kubelet[3127]: I0508 07:59:41.008681 3127 topology_manager.go:215] "Topology Admit Handler" podUID="1ce5a23b-f17a-4855-93ef-b2330fa6da43" podNamespace="kube-system" podName="coredns-7db6d8ff4d-j6zs8" May 8 07:59:41.020124 kubelet[3127]: I0508 07:59:41.018103 3127 topology_manager.go:215] "Topology Admit Handler" podUID="8cf51126-03aa-4a46-85cc-8d942599d347" podNamespace="kube-system" podName="coredns-7db6d8ff4d-bqt4w" May 8 07:59:41.021868 systemd[1]: Created slice kubepods-burstable-pod1ce5a23b_f17a_4855_93ef_b2330fa6da43.slice - libcontainer container kubepods-burstable-pod1ce5a23b_f17a_4855_93ef_b2330fa6da43.slice. May 8 07:59:41.034789 systemd[1]: Created slice kubepods-burstable-pod8cf51126_03aa_4a46_85cc_8d942599d347.slice - libcontainer container kubepods-burstable-pod8cf51126_03aa_4a46_85cc_8d942599d347.slice. May 8 07:59:41.096221 kubelet[3127]: I0508 07:59:41.096173 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ce5a23b-f17a-4855-93ef-b2330fa6da43-config-volume\") pod \"coredns-7db6d8ff4d-j6zs8\" (UID: \"1ce5a23b-f17a-4855-93ef-b2330fa6da43\") " pod="kube-system/coredns-7db6d8ff4d-j6zs8" May 8 07:59:41.096221 kubelet[3127]: I0508 07:59:41.096215 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w69fz\" (UniqueName: \"kubernetes.io/projected/8cf51126-03aa-4a46-85cc-8d942599d347-kube-api-access-w69fz\") pod \"coredns-7db6d8ff4d-bqt4w\" (UID: \"8cf51126-03aa-4a46-85cc-8d942599d347\") " pod="kube-system/coredns-7db6d8ff4d-bqt4w" May 8 07:59:41.096389 kubelet[3127]: I0508 07:59:41.096242 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dltls\" (UniqueName: \"kubernetes.io/projected/1ce5a23b-f17a-4855-93ef-b2330fa6da43-kube-api-access-dltls\") pod \"coredns-7db6d8ff4d-j6zs8\" (UID: \"1ce5a23b-f17a-4855-93ef-b2330fa6da43\") " pod="kube-system/coredns-7db6d8ff4d-j6zs8" May 8 07:59:41.096389 kubelet[3127]: I0508 07:59:41.096260 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf51126-03aa-4a46-85cc-8d942599d347-config-volume\") pod \"coredns-7db6d8ff4d-bqt4w\" (UID: \"8cf51126-03aa-4a46-85cc-8d942599d347\") " pod="kube-system/coredns-7db6d8ff4d-bqt4w" May 8 07:59:41.329303 containerd[1535]: time="2025-05-08T07:59:41.328948969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-j6zs8,Uid:1ce5a23b-f17a-4855-93ef-b2330fa6da43,Namespace:kube-system,Attempt:0,}" May 8 07:59:41.342136 containerd[1535]: time="2025-05-08T07:59:41.342077337Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-bqt4w,Uid:8cf51126-03aa-4a46-85cc-8d942599d347,Namespace:kube-system,Attempt:0,}" May 8 07:59:41.807871 kubelet[3127]: I0508 07:59:41.807548 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-xlz5j" podStartSLOduration=7.793423453 podStartE2EDuration="16.807532277s" podCreationTimestamp="2025-05-08 07:59:25 +0000 UTC" firstStartedPulling="2025-05-08 07:59:26.328218101 +0000 UTC m=+15.858880798" lastFinishedPulling="2025-05-08 07:59:35.342326905 +0000 UTC m=+24.872989622" observedRunningTime="2025-05-08 07:59:41.805233637 +0000 UTC m=+31.335896343" watchObservedRunningTime="2025-05-08 07:59:41.807532277 +0000 UTC m=+31.338194983" May 8 07:59:43.067521 systemd-networkd[1419]: cilium_host: Link UP May 8 07:59:43.067959 systemd-networkd[1419]: cilium_net: Link UP May 8 07:59:43.075089 systemd-networkd[1419]: cilium_net: Gained carrier May 8 07:59:43.075448 systemd-networkd[1419]: cilium_host: Gained carrier May 8 07:59:43.166955 systemd-networkd[1419]: cilium_vxlan: Link UP May 8 07:59:43.167206 systemd-networkd[1419]: cilium_vxlan: Gained carrier May 8 07:59:43.426960 kernel: NET: Registered PF_ALG protocol family May 8 07:59:43.715023 systemd-networkd[1419]: cilium_net: Gained IPv6LL May 8 07:59:43.972067 systemd-networkd[1419]: cilium_host: Gained IPv6LL May 8 07:59:44.156953 systemd-networkd[1419]: lxc_health: Link UP May 8 07:59:44.159902 systemd-networkd[1419]: lxc_health: Gained carrier May 8 07:59:44.362126 systemd-networkd[1419]: lxcc9617a06d2fa: Link UP May 8 07:59:44.368115 kernel: eth0: renamed from tmpbab03 May 8 07:59:44.374711 systemd-networkd[1419]: lxcc9617a06d2fa: Gained carrier May 8 07:59:44.384692 systemd-networkd[1419]: lxc862f72c0fb3a: Link UP May 8 07:59:44.391956 kernel: eth0: renamed from tmp7c6a9 May 8 07:59:44.392777 systemd-networkd[1419]: lxc862f72c0fb3a: Gained carrier May 8 07:59:44.867070 systemd-networkd[1419]: cilium_vxlan: Gained IPv6LL May 8 07:59:45.443100 systemd-networkd[1419]: lxcc9617a06d2fa: Gained IPv6LL May 8 07:59:45.635082 systemd-networkd[1419]: lxc862f72c0fb3a: Gained IPv6LL May 8 07:59:46.019234 systemd-networkd[1419]: lxc_health: Gained IPv6LL May 8 07:59:48.618604 containerd[1535]: time="2025-05-08T07:59:48.617380548Z" level=info msg="connecting to shim bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd" address="unix:///run/containerd/s/d06cdcdbee903e47ed10f282b12b91830e7604303d99a7145e078d13354699de" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:48.654608 systemd[1]: Started cri-containerd-bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd.scope - libcontainer container bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd. May 8 07:59:48.674929 containerd[1535]: time="2025-05-08T07:59:48.674397132Z" level=info msg="connecting to shim 7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24" address="unix:///run/containerd/s/3426b05c2320dc83c40fd50e5d6bc140f54ce24246ce6fe0c1479ddc5799911a" namespace=k8s.io protocol=ttrpc version=3 May 8 07:59:48.727516 systemd[1]: Started cri-containerd-7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24.scope - libcontainer container 7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24. May 8 07:59:48.736942 containerd[1535]: time="2025-05-08T07:59:48.736879573Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-j6zs8,Uid:1ce5a23b-f17a-4855-93ef-b2330fa6da43,Namespace:kube-system,Attempt:0,} returns sandbox id \"bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd\"" May 8 07:59:48.742711 containerd[1535]: time="2025-05-08T07:59:48.742669931Z" level=info msg="CreateContainer within sandbox \"bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 07:59:48.769104 containerd[1535]: time="2025-05-08T07:59:48.769047954Z" level=info msg="Container 6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:48.786234 containerd[1535]: time="2025-05-08T07:59:48.786159707Z" level=info msg="CreateContainer within sandbox \"bab03819b18e35ce5d453916c4e0475c6e4c944b14950ab69df494f79deaabcd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085\"" May 8 07:59:48.787028 containerd[1535]: time="2025-05-08T07:59:48.786994072Z" level=info msg="StartContainer for \"6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085\"" May 8 07:59:48.787862 containerd[1535]: time="2025-05-08T07:59:48.787835619Z" level=info msg="connecting to shim 6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085" address="unix:///run/containerd/s/d06cdcdbee903e47ed10f282b12b91830e7604303d99a7145e078d13354699de" protocol=ttrpc version=3 May 8 07:59:48.806018 containerd[1535]: time="2025-05-08T07:59:48.805575815Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-bqt4w,Uid:8cf51126-03aa-4a46-85cc-8d942599d347,Namespace:kube-system,Attempt:0,} returns sandbox id \"7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24\"" May 8 07:59:48.810154 containerd[1535]: time="2025-05-08T07:59:48.810003471Z" level=info msg="CreateContainer within sandbox \"7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 07:59:48.816086 systemd[1]: Started cri-containerd-6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085.scope - libcontainer container 6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085. May 8 07:59:48.826068 containerd[1535]: time="2025-05-08T07:59:48.826039743Z" level=info msg="Container 8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0: CDI devices from CRI Config.CDIDevices: []" May 8 07:59:48.840385 containerd[1535]: time="2025-05-08T07:59:48.840186688Z" level=info msg="CreateContainer within sandbox \"7c6a976230ff54d986a36a2a1719f3e947d03783cf0a43ff6a6bb1d6498a4e24\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0\"" May 8 07:59:48.841895 containerd[1535]: time="2025-05-08T07:59:48.841187410Z" level=info msg="StartContainer for \"8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0\"" May 8 07:59:48.842001 containerd[1535]: time="2025-05-08T07:59:48.841940453Z" level=info msg="connecting to shim 8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0" address="unix:///run/containerd/s/3426b05c2320dc83c40fd50e5d6bc140f54ce24246ce6fe0c1479ddc5799911a" protocol=ttrpc version=3 May 8 07:59:48.861107 systemd[1]: Started cri-containerd-8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0.scope - libcontainer container 8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0. May 8 07:59:48.875345 containerd[1535]: time="2025-05-08T07:59:48.874441612Z" level=info msg="StartContainer for \"6bcac069d42541a2cf6dd1f40026b5c17c7d894e98c8336d0f15386938df2085\" returns successfully" May 8 07:59:48.913677 containerd[1535]: time="2025-05-08T07:59:48.913613436Z" level=info msg="StartContainer for \"8ff39faddd33f3cba1414e048db51698a8fe584a2b264b80766a6ce91faeb3e0\" returns successfully" May 8 07:59:49.610082 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount487365379.mount: Deactivated successfully. May 8 07:59:49.812515 kubelet[3127]: I0508 07:59:49.812387 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-j6zs8" podStartSLOduration=23.812357517 podStartE2EDuration="23.812357517s" podCreationTimestamp="2025-05-08 07:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:49.808429753 +0000 UTC m=+39.339092499" watchObservedRunningTime="2025-05-08 07:59:49.812357517 +0000 UTC m=+39.343020263" May 8 08:00:33.224133 systemd[1]: Started sshd@7-172.24.4.249:22-172.24.4.1:43392.service - OpenSSH per-connection server daemon (172.24.4.1:43392). May 8 08:00:34.514371 sshd[4431]: Accepted publickey for core from 172.24.4.1 port 43392 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:34.518462 sshd-session[4431]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:34.535052 systemd-logind[1509]: New session 10 of user core. May 8 08:00:34.544284 systemd[1]: Started session-10.scope - Session 10 of User core. May 8 08:00:35.255360 sshd[4433]: Connection closed by 172.24.4.1 port 43392 May 8 08:00:35.256436 sshd-session[4431]: pam_unix(sshd:session): session closed for user core May 8 08:00:35.262196 systemd[1]: sshd@7-172.24.4.249:22-172.24.4.1:43392.service: Deactivated successfully. May 8 08:00:35.265767 systemd[1]: session-10.scope: Deactivated successfully. May 8 08:00:35.270610 systemd-logind[1509]: Session 10 logged out. Waiting for processes to exit. May 8 08:00:35.273812 systemd-logind[1509]: Removed session 10. May 8 08:00:40.271150 systemd[1]: Started sshd@8-172.24.4.249:22-172.24.4.1:33066.service - OpenSSH per-connection server daemon (172.24.4.1:33066). May 8 08:00:41.466048 sshd[4446]: Accepted publickey for core from 172.24.4.1 port 33066 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:41.468070 sshd-session[4446]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:41.475583 systemd-logind[1509]: New session 11 of user core. May 8 08:00:41.482057 systemd[1]: Started session-11.scope - Session 11 of User core. May 8 08:00:42.246867 sshd[4448]: Connection closed by 172.24.4.1 port 33066 May 8 08:00:42.246460 sshd-session[4446]: pam_unix(sshd:session): session closed for user core May 8 08:00:42.252591 systemd[1]: sshd@8-172.24.4.249:22-172.24.4.1:33066.service: Deactivated successfully. May 8 08:00:42.255430 systemd[1]: session-11.scope: Deactivated successfully. May 8 08:00:42.257138 systemd-logind[1509]: Session 11 logged out. Waiting for processes to exit. May 8 08:00:42.259725 systemd-logind[1509]: Removed session 11. May 8 08:00:47.275280 systemd[1]: Started sshd@9-172.24.4.249:22-172.24.4.1:51552.service - OpenSSH per-connection server daemon (172.24.4.1:51552). May 8 08:00:48.229025 sshd[4462]: Accepted publickey for core from 172.24.4.1 port 51552 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:48.231616 sshd-session[4462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:48.244272 systemd-logind[1509]: New session 12 of user core. May 8 08:00:48.250254 systemd[1]: Started session-12.scope - Session 12 of User core. May 8 08:00:49.012668 sshd[4464]: Connection closed by 172.24.4.1 port 51552 May 8 08:00:49.013700 sshd-session[4462]: pam_unix(sshd:session): session closed for user core May 8 08:00:49.019096 systemd[1]: sshd@9-172.24.4.249:22-172.24.4.1:51552.service: Deactivated successfully. May 8 08:00:49.026555 systemd[1]: session-12.scope: Deactivated successfully. May 8 08:00:49.033364 systemd-logind[1509]: Session 12 logged out. Waiting for processes to exit. May 8 08:00:49.036853 systemd-logind[1509]: Removed session 12. May 8 08:00:54.032538 systemd[1]: Started sshd@10-172.24.4.249:22-172.24.4.1:51550.service - OpenSSH per-connection server daemon (172.24.4.1:51550). May 8 08:00:55.325241 sshd[4477]: Accepted publickey for core from 172.24.4.1 port 51550 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:55.328834 sshd-session[4477]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:55.345351 systemd-logind[1509]: New session 13 of user core. May 8 08:00:55.354236 systemd[1]: Started session-13.scope - Session 13 of User core. May 8 08:00:56.158708 sshd[4479]: Connection closed by 172.24.4.1 port 51550 May 8 08:00:56.158447 sshd-session[4477]: pam_unix(sshd:session): session closed for user core May 8 08:00:56.178779 systemd[1]: sshd@10-172.24.4.249:22-172.24.4.1:51550.service: Deactivated successfully. May 8 08:00:56.185271 systemd[1]: session-13.scope: Deactivated successfully. May 8 08:00:56.189375 systemd-logind[1509]: Session 13 logged out. Waiting for processes to exit. May 8 08:00:56.196599 systemd[1]: Started sshd@11-172.24.4.249:22-172.24.4.1:51554.service - OpenSSH per-connection server daemon (172.24.4.1:51554). May 8 08:00:56.202459 systemd-logind[1509]: Removed session 13. May 8 08:00:57.384441 sshd[4491]: Accepted publickey for core from 172.24.4.1 port 51554 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:57.388411 sshd-session[4491]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:57.402065 systemd-logind[1509]: New session 14 of user core. May 8 08:00:57.407173 systemd[1]: Started session-14.scope - Session 14 of User core. May 8 08:00:58.298218 sshd[4497]: Connection closed by 172.24.4.1 port 51554 May 8 08:00:58.301885 sshd-session[4491]: pam_unix(sshd:session): session closed for user core May 8 08:00:58.328815 systemd[1]: sshd@11-172.24.4.249:22-172.24.4.1:51554.service: Deactivated successfully. May 8 08:00:58.333974 systemd[1]: session-14.scope: Deactivated successfully. May 8 08:00:58.340276 systemd-logind[1509]: Session 14 logged out. Waiting for processes to exit. May 8 08:00:58.344351 systemd[1]: Started sshd@12-172.24.4.249:22-172.24.4.1:51566.service - OpenSSH per-connection server daemon (172.24.4.1:51566). May 8 08:00:58.349284 systemd-logind[1509]: Removed session 14. May 8 08:00:59.372201 sshd[4506]: Accepted publickey for core from 172.24.4.1 port 51566 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:00:59.376091 sshd-session[4506]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:00:59.390057 systemd-logind[1509]: New session 15 of user core. May 8 08:00:59.396236 systemd[1]: Started session-15.scope - Session 15 of User core. May 8 08:01:00.307531 sshd[4509]: Connection closed by 172.24.4.1 port 51566 May 8 08:01:00.307284 sshd-session[4506]: pam_unix(sshd:session): session closed for user core May 8 08:01:00.318678 systemd-logind[1509]: Session 15 logged out. Waiting for processes to exit. May 8 08:01:00.321071 systemd[1]: sshd@12-172.24.4.249:22-172.24.4.1:51566.service: Deactivated successfully. May 8 08:01:00.331432 systemd[1]: session-15.scope: Deactivated successfully. May 8 08:01:00.338492 systemd-logind[1509]: Removed session 15. May 8 08:01:05.332177 systemd[1]: Started sshd@13-172.24.4.249:22-172.24.4.1:54782.service - OpenSSH per-connection server daemon (172.24.4.1:54782). May 8 08:01:06.481776 sshd[4521]: Accepted publickey for core from 172.24.4.1 port 54782 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:06.485105 sshd-session[4521]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:06.498478 systemd-logind[1509]: New session 16 of user core. May 8 08:01:06.505232 systemd[1]: Started session-16.scope - Session 16 of User core. May 8 08:01:07.264000 sshd[4523]: Connection closed by 172.24.4.1 port 54782 May 8 08:01:07.262584 sshd-session[4521]: pam_unix(sshd:session): session closed for user core May 8 08:01:07.272750 systemd[1]: sshd@13-172.24.4.249:22-172.24.4.1:54782.service: Deactivated successfully. May 8 08:01:07.279083 systemd[1]: session-16.scope: Deactivated successfully. May 8 08:01:07.281395 systemd-logind[1509]: Session 16 logged out. Waiting for processes to exit. May 8 08:01:07.284615 systemd-logind[1509]: Removed session 16. May 8 08:01:12.284216 systemd[1]: Started sshd@14-172.24.4.249:22-172.24.4.1:54788.service - OpenSSH per-connection server daemon (172.24.4.1:54788). May 8 08:01:13.477001 sshd[4537]: Accepted publickey for core from 172.24.4.1 port 54788 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:13.479852 sshd-session[4537]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:13.492583 systemd-logind[1509]: New session 17 of user core. May 8 08:01:13.508286 systemd[1]: Started session-17.scope - Session 17 of User core. May 8 08:01:14.257969 sshd[4539]: Connection closed by 172.24.4.1 port 54788 May 8 08:01:14.258334 sshd-session[4537]: pam_unix(sshd:session): session closed for user core May 8 08:01:14.274891 systemd[1]: sshd@14-172.24.4.249:22-172.24.4.1:54788.service: Deactivated successfully. May 8 08:01:14.281622 systemd[1]: session-17.scope: Deactivated successfully. May 8 08:01:14.285186 systemd-logind[1509]: Session 17 logged out. Waiting for processes to exit. May 8 08:01:14.290803 systemd[1]: Started sshd@15-172.24.4.249:22-172.24.4.1:39688.service - OpenSSH per-connection server daemon (172.24.4.1:39688). May 8 08:01:14.294882 systemd-logind[1509]: Removed session 17. May 8 08:01:15.516205 sshd[4551]: Accepted publickey for core from 172.24.4.1 port 39688 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:15.519065 sshd-session[4551]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:15.531253 systemd-logind[1509]: New session 18 of user core. May 8 08:01:15.542303 systemd[1]: Started session-18.scope - Session 18 of User core. May 8 08:01:16.280862 sshd[4554]: Connection closed by 172.24.4.1 port 39688 May 8 08:01:16.283295 sshd-session[4551]: pam_unix(sshd:session): session closed for user core May 8 08:01:16.297851 systemd[1]: Started sshd@16-172.24.4.249:22-172.24.4.1:39702.service - OpenSSH per-connection server daemon (172.24.4.1:39702). May 8 08:01:16.300769 systemd[1]: sshd@15-172.24.4.249:22-172.24.4.1:39688.service: Deactivated successfully. May 8 08:01:16.311778 systemd[1]: session-18.scope: Deactivated successfully. May 8 08:01:16.317056 systemd-logind[1509]: Session 18 logged out. Waiting for processes to exit. May 8 08:01:16.326659 systemd-logind[1509]: Removed session 18. May 8 08:01:17.455974 sshd[4561]: Accepted publickey for core from 172.24.4.1 port 39702 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:17.459411 sshd-session[4561]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:17.474030 systemd-logind[1509]: New session 19 of user core. May 8 08:01:17.482278 systemd[1]: Started session-19.scope - Session 19 of User core. May 8 08:01:20.302527 sshd[4566]: Connection closed by 172.24.4.1 port 39702 May 8 08:01:20.302211 sshd-session[4561]: pam_unix(sshd:session): session closed for user core May 8 08:01:20.319798 systemd[1]: sshd@16-172.24.4.249:22-172.24.4.1:39702.service: Deactivated successfully. May 8 08:01:20.327997 systemd[1]: session-19.scope: Deactivated successfully. May 8 08:01:20.334039 systemd-logind[1509]: Session 19 logged out. Waiting for processes to exit. May 8 08:01:20.337468 systemd[1]: Started sshd@17-172.24.4.249:22-172.24.4.1:39714.service - OpenSSH per-connection server daemon (172.24.4.1:39714). May 8 08:01:20.342478 systemd-logind[1509]: Removed session 19. May 8 08:01:21.575094 sshd[4583]: Accepted publickey for core from 172.24.4.1 port 39714 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:21.576035 sshd-session[4583]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:21.590036 systemd-logind[1509]: New session 20 of user core. May 8 08:01:21.595271 systemd[1]: Started session-20.scope - Session 20 of User core. May 8 08:01:22.293701 sshd[4586]: Connection closed by 172.24.4.1 port 39714 May 8 08:01:22.293215 sshd-session[4583]: pam_unix(sshd:session): session closed for user core May 8 08:01:22.304888 systemd[1]: sshd@17-172.24.4.249:22-172.24.4.1:39714.service: Deactivated successfully. May 8 08:01:22.307272 systemd[1]: session-20.scope: Deactivated successfully. May 8 08:01:22.313976 systemd-logind[1509]: Session 20 logged out. Waiting for processes to exit. May 8 08:01:22.319122 systemd[1]: Started sshd@18-172.24.4.249:22-172.24.4.1:39730.service - OpenSSH per-connection server daemon (172.24.4.1:39730). May 8 08:01:22.321794 systemd-logind[1509]: Removed session 20. May 8 08:01:23.517738 sshd[4594]: Accepted publickey for core from 172.24.4.1 port 39730 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:23.520694 sshd-session[4594]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:23.528229 systemd-logind[1509]: New session 21 of user core. May 8 08:01:23.532082 systemd[1]: Started session-21.scope - Session 21 of User core. May 8 08:01:24.282016 sshd[4597]: Connection closed by 172.24.4.1 port 39730 May 8 08:01:24.282571 sshd-session[4594]: pam_unix(sshd:session): session closed for user core May 8 08:01:24.289351 systemd[1]: sshd@18-172.24.4.249:22-172.24.4.1:39730.service: Deactivated successfully. May 8 08:01:24.293029 systemd[1]: session-21.scope: Deactivated successfully. May 8 08:01:24.296503 systemd-logind[1509]: Session 21 logged out. Waiting for processes to exit. May 8 08:01:24.297964 systemd-logind[1509]: Removed session 21. May 8 08:01:29.307782 systemd[1]: Started sshd@19-172.24.4.249:22-172.24.4.1:59244.service - OpenSSH per-connection server daemon (172.24.4.1:59244). May 8 08:01:30.500549 sshd[4614]: Accepted publickey for core from 172.24.4.1 port 59244 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:30.503675 sshd-session[4614]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:30.517072 systemd-logind[1509]: New session 22 of user core. May 8 08:01:30.522469 systemd[1]: Started session-22.scope - Session 22 of User core. May 8 08:01:31.280343 sshd[4616]: Connection closed by 172.24.4.1 port 59244 May 8 08:01:31.281605 sshd-session[4614]: pam_unix(sshd:session): session closed for user core May 8 08:01:31.288672 systemd[1]: sshd@19-172.24.4.249:22-172.24.4.1:59244.service: Deactivated successfully. May 8 08:01:31.295397 systemd[1]: session-22.scope: Deactivated successfully. May 8 08:01:31.299373 systemd-logind[1509]: Session 22 logged out. Waiting for processes to exit. May 8 08:01:31.302163 systemd-logind[1509]: Removed session 22. May 8 08:01:36.309320 systemd[1]: Started sshd@20-172.24.4.249:22-172.24.4.1:35974.service - OpenSSH per-connection server daemon (172.24.4.1:35974). May 8 08:01:37.770858 sshd[4628]: Accepted publickey for core from 172.24.4.1 port 35974 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:37.774194 sshd-session[4628]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:37.787268 systemd-logind[1509]: New session 23 of user core. May 8 08:01:37.795440 systemd[1]: Started session-23.scope - Session 23 of User core. May 8 08:01:38.602984 sshd[4630]: Connection closed by 172.24.4.1 port 35974 May 8 08:01:38.604162 sshd-session[4628]: pam_unix(sshd:session): session closed for user core May 8 08:01:38.612806 systemd[1]: sshd@20-172.24.4.249:22-172.24.4.1:35974.service: Deactivated successfully. May 8 08:01:38.618087 systemd[1]: session-23.scope: Deactivated successfully. May 8 08:01:38.620589 systemd-logind[1509]: Session 23 logged out. Waiting for processes to exit. May 8 08:01:38.623987 systemd-logind[1509]: Removed session 23. May 8 08:01:43.628605 systemd[1]: Started sshd@21-172.24.4.249:22-172.24.4.1:47110.service - OpenSSH per-connection server daemon (172.24.4.1:47110). May 8 08:01:44.831213 sshd[4642]: Accepted publickey for core from 172.24.4.1 port 47110 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:44.834377 sshd-session[4642]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:44.846756 systemd-logind[1509]: New session 24 of user core. May 8 08:01:44.861372 systemd[1]: Started session-24.scope - Session 24 of User core. May 8 08:01:45.665272 sshd[4644]: Connection closed by 172.24.4.1 port 47110 May 8 08:01:45.663792 sshd-session[4642]: pam_unix(sshd:session): session closed for user core May 8 08:01:45.680241 systemd[1]: sshd@21-172.24.4.249:22-172.24.4.1:47110.service: Deactivated successfully. May 8 08:01:45.689487 systemd[1]: session-24.scope: Deactivated successfully. May 8 08:01:45.698562 systemd-logind[1509]: Session 24 logged out. Waiting for processes to exit. May 8 08:01:45.703525 systemd[1]: Started sshd@22-172.24.4.249:22-172.24.4.1:47126.service - OpenSSH per-connection server daemon (172.24.4.1:47126). May 8 08:01:45.708083 systemd-logind[1509]: Removed session 24. May 8 08:01:46.837141 sshd[4655]: Accepted publickey for core from 172.24.4.1 port 47126 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:46.841958 sshd-session[4655]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:46.854263 systemd-logind[1509]: New session 25 of user core. May 8 08:01:46.864278 systemd[1]: Started session-25.scope - Session 25 of User core. May 8 08:01:48.796180 kubelet[3127]: I0508 08:01:48.796032 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-bqt4w" podStartSLOduration=142.795995744 podStartE2EDuration="2m22.795995744s" podCreationTimestamp="2025-05-08 07:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 07:59:49.870621738 +0000 UTC m=+39.401284434" watchObservedRunningTime="2025-05-08 08:01:48.795995744 +0000 UTC m=+158.326658440" May 8 08:01:48.805002 containerd[1535]: time="2025-05-08T08:01:48.804077940Z" level=info msg="StopContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" with timeout 30 (s)" May 8 08:01:48.805825 containerd[1535]: time="2025-05-08T08:01:48.805136627Z" level=info msg="Stop container \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" with signal terminated" May 8 08:01:48.827153 systemd[1]: cri-containerd-92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f.scope: Deactivated successfully. May 8 08:01:48.830707 containerd[1535]: time="2025-05-08T08:01:48.830672209Z" level=info msg="received exit event container_id:\"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" id:\"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" pid:3676 exited_at:{seconds:1746691308 nanos:829424082}" May 8 08:01:48.831673 containerd[1535]: time="2025-05-08T08:01:48.831535837Z" level=info msg="TaskExit event in podsandbox handler container_id:\"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" id:\"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" pid:3676 exited_at:{seconds:1746691308 nanos:829424082}" May 8 08:01:48.853332 containerd[1535]: time="2025-05-08T08:01:48.853257918Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 8 08:01:48.861635 containerd[1535]: time="2025-05-08T08:01:48.861594315Z" level=info msg="TaskExit event in podsandbox handler container_id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" id:\"707c97a05985d6bfbd12e8f715608778a2c14560200222ebe56debbcc71ce53e\" pid:4688 exited_at:{seconds:1746691308 nanos:861123873}" May 8 08:01:48.862008 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f-rootfs.mount: Deactivated successfully. May 8 08:01:48.870126 containerd[1535]: time="2025-05-08T08:01:48.870063014Z" level=info msg="StopContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" with timeout 2 (s)" May 8 08:01:48.870853 containerd[1535]: time="2025-05-08T08:01:48.870699421Z" level=info msg="Stop container \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" with signal terminated" May 8 08:01:48.889113 systemd-networkd[1419]: lxc_health: Link DOWN May 8 08:01:48.889121 systemd-networkd[1419]: lxc_health: Lost carrier May 8 08:01:48.897210 containerd[1535]: time="2025-05-08T08:01:48.897081959Z" level=info msg="StopContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" returns successfully" May 8 08:01:48.898443 containerd[1535]: time="2025-05-08T08:01:48.898055765Z" level=info msg="StopPodSandbox for \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\"" May 8 08:01:48.898443 containerd[1535]: time="2025-05-08T08:01:48.898123164Z" level=info msg="Container to stop \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.908340 systemd[1]: cri-containerd-81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff.scope: Deactivated successfully. May 8 08:01:48.908627 systemd[1]: cri-containerd-81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff.scope: Consumed 8.135s CPU time, 126M memory peak, 128K read from disk, 13.3M written to disk. May 8 08:01:48.910522 systemd[1]: cri-containerd-7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb.scope: Deactivated successfully. May 8 08:01:48.911534 containerd[1535]: time="2025-05-08T08:01:48.911473479Z" level=info msg="TaskExit event in podsandbox handler container_id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" pid:3746 exited_at:{seconds:1746691308 nanos:910821663}" May 8 08:01:48.913181 containerd[1535]: time="2025-05-08T08:01:48.912635713Z" level=info msg="received exit event container_id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" id:\"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" pid:3746 exited_at:{seconds:1746691308 nanos:910821663}" May 8 08:01:48.913181 containerd[1535]: time="2025-05-08T08:01:48.913128227Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" id:\"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" pid:3386 exit_status:137 exited_at:{seconds:1746691308 nanos:912584706}" May 8 08:01:48.942077 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff-rootfs.mount: Deactivated successfully. May 8 08:01:48.961420 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb-rootfs.mount: Deactivated successfully. May 8 08:01:48.979514 containerd[1535]: time="2025-05-08T08:01:48.979484655Z" level=info msg="shim disconnected" id=7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb namespace=k8s.io May 8 08:01:48.979749 containerd[1535]: time="2025-05-08T08:01:48.979636824Z" level=warning msg="cleaning up after shim disconnected" id=7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb namespace=k8s.io May 8 08:01:48.979749 containerd[1535]: time="2025-05-08T08:01:48.979650831Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 8 08:01:48.991214 containerd[1535]: time="2025-05-08T08:01:48.991104900Z" level=info msg="StopContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" returns successfully" May 8 08:01:48.992034 containerd[1535]: time="2025-05-08T08:01:48.991997182Z" level=info msg="StopPodSandbox for \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\"" May 8 08:01:48.992100 containerd[1535]: time="2025-05-08T08:01:48.992079178Z" level=info msg="Container to stop \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.992137 containerd[1535]: time="2025-05-08T08:01:48.992097422Z" level=info msg="Container to stop \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.992137 containerd[1535]: time="2025-05-08T08:01:48.992114034Z" level=info msg="Container to stop \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.992193 containerd[1535]: time="2025-05-08T08:01:48.992129924Z" level=info msg="Container to stop \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.992193 containerd[1535]: time="2025-05-08T08:01:48.992178135Z" level=info msg="Container to stop \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 08:01:48.999942 containerd[1535]: time="2025-05-08T08:01:48.996021223Z" level=info msg="received exit event sandbox_id:\"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" exit_status:137 exited_at:{seconds:1746691308 nanos:912584706}" May 8 08:01:48.999942 containerd[1535]: time="2025-05-08T08:01:48.997977052Z" level=info msg="TearDown network for sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" successfully" May 8 08:01:48.999942 containerd[1535]: time="2025-05-08T08:01:48.997998382Z" level=info msg="StopPodSandbox for \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" returns successfully" May 8 08:01:48.999907 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb-shm.mount: Deactivated successfully. May 8 08:01:49.001370 systemd[1]: cri-containerd-84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b.scope: Deactivated successfully. May 8 08:01:49.004861 containerd[1535]: time="2025-05-08T08:01:49.004796644Z" level=info msg="TaskExit event in podsandbox handler container_id:\"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" id:\"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" pid:3272 exit_status:137 exited_at:{seconds:1746691309 nanos:4316083}" May 8 08:01:49.035371 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b-rootfs.mount: Deactivated successfully. May 8 08:01:49.061048 containerd[1535]: time="2025-05-08T08:01:49.060608504Z" level=info msg="received exit event sandbox_id:\"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" exit_status:137 exited_at:{seconds:1746691309 nanos:4316083}" May 8 08:01:49.061048 containerd[1535]: time="2025-05-08T08:01:49.060838640Z" level=info msg="TearDown network for sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" successfully" May 8 08:01:49.061048 containerd[1535]: time="2025-05-08T08:01:49.060875060Z" level=info msg="StopPodSandbox for \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" returns successfully" May 8 08:01:49.062612 containerd[1535]: time="2025-05-08T08:01:49.062472148Z" level=info msg="shim disconnected" id=84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b namespace=k8s.io May 8 08:01:49.062612 containerd[1535]: time="2025-05-08T08:01:49.062499009Z" level=warning msg="cleaning up after shim disconnected" id=84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b namespace=k8s.io May 8 08:01:49.062612 containerd[1535]: time="2025-05-08T08:01:49.062531501Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 8 08:01:49.083052 kubelet[3127]: I0508 08:01:49.081990 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtbtb\" (UniqueName: \"kubernetes.io/projected/60355b88-43ea-493c-a59c-b9a39cecea05-kube-api-access-qtbtb\") pod \"60355b88-43ea-493c-a59c-b9a39cecea05\" (UID: \"60355b88-43ea-493c-a59c-b9a39cecea05\") " May 8 08:01:49.083052 kubelet[3127]: I0508 08:01:49.082060 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/60355b88-43ea-493c-a59c-b9a39cecea05-cilium-config-path\") pod \"60355b88-43ea-493c-a59c-b9a39cecea05\" (UID: \"60355b88-43ea-493c-a59c-b9a39cecea05\") " May 8 08:01:49.091316 kubelet[3127]: I0508 08:01:49.091270 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60355b88-43ea-493c-a59c-b9a39cecea05-kube-api-access-qtbtb" (OuterVolumeSpecName: "kube-api-access-qtbtb") pod "60355b88-43ea-493c-a59c-b9a39cecea05" (UID: "60355b88-43ea-493c-a59c-b9a39cecea05"). InnerVolumeSpecName "kube-api-access-qtbtb". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 08:01:49.091704 kubelet[3127]: I0508 08:01:49.091671 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60355b88-43ea-493c-a59c-b9a39cecea05-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "60355b88-43ea-493c-a59c-b9a39cecea05" (UID: "60355b88-43ea-493c-a59c-b9a39cecea05"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183001 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-lib-modules\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183086 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-kernel\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183123 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-cgroup\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183177 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-bpf-maps\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183241 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/01da910d-c3da-436f-ae6b-18a93de1a297-clustermesh-secrets\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184028 kubelet[3127]: I0508 08:01:49.183233 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.184464 kubelet[3127]: I0508 08:01:49.183272 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-run\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184464 kubelet[3127]: I0508 08:01:49.183332 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.184464 kubelet[3127]: I0508 08:01:49.183383 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.184464 kubelet[3127]: I0508 08:01:49.183380 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-etc-cni-netd\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184464 kubelet[3127]: I0508 08:01:49.183418 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183445 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cni-path\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183446 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183491 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-xtables-lock\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183531 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-net\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183581 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-config-path\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.184782 kubelet[3127]: I0508 08:01:49.183648 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-hubble-tls\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183706 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r98vr\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-kube-api-access-r98vr\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183745 3127 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-hostproc\") pod \"01da910d-c3da-436f-ae6b-18a93de1a297\" (UID: \"01da910d-c3da-436f-ae6b-18a93de1a297\") " May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183825 3127 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-lib-modules\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183853 3127 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-kernel\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183878 3127 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-cgroup\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.183976 3127 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-bpf-maps\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.185892 kubelet[3127]: I0508 08:01:49.184014 3127 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-run\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.188432 kubelet[3127]: I0508 08:01:49.184041 3127 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-qtbtb\" (UniqueName: \"kubernetes.io/projected/60355b88-43ea-493c-a59c-b9a39cecea05-kube-api-access-qtbtb\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.188432 kubelet[3127]: I0508 08:01:49.184067 3127 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/60355b88-43ea-493c-a59c-b9a39cecea05-cilium-config-path\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.188432 kubelet[3127]: I0508 08:01:49.184114 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-hostproc" (OuterVolumeSpecName: "hostproc") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.188432 kubelet[3127]: I0508 08:01:49.184159 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.188432 kubelet[3127]: I0508 08:01:49.184195 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cni-path" (OuterVolumeSpecName: "cni-path") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.188710 kubelet[3127]: I0508 08:01:49.184231 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.188710 kubelet[3127]: I0508 08:01:49.184265 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 08:01:49.194521 kubelet[3127]: I0508 08:01:49.194439 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 8 08:01:49.200759 kubelet[3127]: I0508 08:01:49.200687 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01da910d-c3da-436f-ae6b-18a93de1a297-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 8 08:01:49.206024 kubelet[3127]: I0508 08:01:49.205815 3127 scope.go:117] "RemoveContainer" containerID="92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f" May 8 08:01:49.207744 kubelet[3127]: I0508 08:01:49.207652 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 08:01:49.208982 kubelet[3127]: I0508 08:01:49.208726 3127 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-kube-api-access-r98vr" (OuterVolumeSpecName: "kube-api-access-r98vr") pod "01da910d-c3da-436f-ae6b-18a93de1a297" (UID: "01da910d-c3da-436f-ae6b-18a93de1a297"). InnerVolumeSpecName "kube-api-access-r98vr". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 08:01:49.213610 containerd[1535]: time="2025-05-08T08:01:49.213286139Z" level=info msg="RemoveContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\"" May 8 08:01:49.222284 systemd[1]: Removed slice kubepods-besteffort-pod60355b88_43ea_493c_a59c_b9a39cecea05.slice - libcontainer container kubepods-besteffort-pod60355b88_43ea_493c_a59c_b9a39cecea05.slice. May 8 08:01:49.242911 systemd[1]: Removed slice kubepods-burstable-pod01da910d_c3da_436f_ae6b_18a93de1a297.slice - libcontainer container kubepods-burstable-pod01da910d_c3da_436f_ae6b_18a93de1a297.slice. May 8 08:01:49.243298 systemd[1]: kubepods-burstable-pod01da910d_c3da_436f_ae6b_18a93de1a297.slice: Consumed 8.225s CPU time, 126.4M memory peak, 128K read from disk, 13.3M written to disk. May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.284985 3127 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-hubble-tls\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285053 3127 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-hostproc\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285080 3127 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-r98vr\" (UniqueName: \"kubernetes.io/projected/01da910d-c3da-436f-ae6b-18a93de1a297-kube-api-access-r98vr\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285107 3127 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/01da910d-c3da-436f-ae6b-18a93de1a297-clustermesh-secrets\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285278 3127 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-cni-path\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285309 3127 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-etc-cni-netd\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.285508 kubelet[3127]: I0508 08:01:49.285336 3127 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-xtables-lock\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.286514 kubelet[3127]: I0508 08:01:49.285362 3127 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/01da910d-c3da-436f-ae6b-18a93de1a297-host-proc-sys-net\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.286514 kubelet[3127]: I0508 08:01:49.285444 3127 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/01da910d-c3da-436f-ae6b-18a93de1a297-cilium-config-path\") on node \"ci-4327-0-0-w-45c5ab2731.novalocal\" DevicePath \"\"" May 8 08:01:49.352302 containerd[1535]: time="2025-05-08T08:01:49.351817286Z" level=info msg="RemoveContainer for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" returns successfully" May 8 08:01:49.352482 kubelet[3127]: I0508 08:01:49.352390 3127 scope.go:117] "RemoveContainer" containerID="92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f" May 8 08:01:49.353372 containerd[1535]: time="2025-05-08T08:01:49.353286443Z" level=error msg="ContainerStatus for \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\": not found" May 8 08:01:49.353832 kubelet[3127]: E0508 08:01:49.353599 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\": not found" containerID="92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f" May 8 08:01:49.353832 kubelet[3127]: I0508 08:01:49.353680 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f"} err="failed to get container status \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\": rpc error: code = NotFound desc = an error occurred when try to find container \"92e338ad8e28bf3dd762c7967790e53a8bdf744b46c2f21d52e0b97fcfa3313f\": not found" May 8 08:01:49.354144 kubelet[3127]: I0508 08:01:49.353855 3127 scope.go:117] "RemoveContainer" containerID="81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff" May 8 08:01:49.360016 containerd[1535]: time="2025-05-08T08:01:49.359833769Z" level=info msg="RemoveContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\"" May 8 08:01:49.385389 containerd[1535]: time="2025-05-08T08:01:49.385034599Z" level=info msg="RemoveContainer for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" returns successfully" May 8 08:01:49.385952 kubelet[3127]: I0508 08:01:49.385883 3127 scope.go:117] "RemoveContainer" containerID="d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8" May 8 08:01:49.398674 containerd[1535]: time="2025-05-08T08:01:49.394669039Z" level=info msg="RemoveContainer for \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\"" May 8 08:01:49.419166 containerd[1535]: time="2025-05-08T08:01:49.419109017Z" level=info msg="RemoveContainer for \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" returns successfully" May 8 08:01:49.421455 kubelet[3127]: I0508 08:01:49.421418 3127 scope.go:117] "RemoveContainer" containerID="a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3" May 8 08:01:49.426136 containerd[1535]: time="2025-05-08T08:01:49.426077702Z" level=info msg="RemoveContainer for \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\"" May 8 08:01:49.431648 containerd[1535]: time="2025-05-08T08:01:49.431615323Z" level=info msg="RemoveContainer for \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" returns successfully" May 8 08:01:49.431878 kubelet[3127]: I0508 08:01:49.431760 3127 scope.go:117] "RemoveContainer" containerID="03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106" May 8 08:01:49.433776 containerd[1535]: time="2025-05-08T08:01:49.433288977Z" level=info msg="RemoveContainer for \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\"" May 8 08:01:49.439940 containerd[1535]: time="2025-05-08T08:01:49.438963909Z" level=info msg="RemoveContainer for \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" returns successfully" May 8 08:01:49.440893 kubelet[3127]: I0508 08:01:49.440867 3127 scope.go:117] "RemoveContainer" containerID="5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c" May 8 08:01:49.444517 containerd[1535]: time="2025-05-08T08:01:49.444439523Z" level=info msg="RemoveContainer for \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\"" May 8 08:01:49.448643 containerd[1535]: time="2025-05-08T08:01:49.448615553Z" level=info msg="RemoveContainer for \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" returns successfully" May 8 08:01:49.448901 kubelet[3127]: I0508 08:01:49.448858 3127 scope.go:117] "RemoveContainer" containerID="81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff" May 8 08:01:49.449154 containerd[1535]: time="2025-05-08T08:01:49.449105702Z" level=error msg="ContainerStatus for \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\": not found" May 8 08:01:49.449311 kubelet[3127]: E0508 08:01:49.449246 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\": not found" containerID="81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff" May 8 08:01:49.449353 kubelet[3127]: I0508 08:01:49.449304 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff"} err="failed to get container status \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\": rpc error: code = NotFound desc = an error occurred when try to find container \"81b6fb91fcb9c4ada5b71c42b21e866d4cccb5da1e1b5b4b76af5ce31f2b63ff\": not found" May 8 08:01:49.449353 kubelet[3127]: I0508 08:01:49.449326 3127 scope.go:117] "RemoveContainer" containerID="d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8" May 8 08:01:49.449502 containerd[1535]: time="2025-05-08T08:01:49.449472007Z" level=error msg="ContainerStatus for \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\": not found" May 8 08:01:49.449623 kubelet[3127]: E0508 08:01:49.449584 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\": not found" containerID="d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8" May 8 08:01:49.449623 kubelet[3127]: I0508 08:01:49.449609 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8"} err="failed to get container status \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\": rpc error: code = NotFound desc = an error occurred when try to find container \"d930caae8a47a4b607801077cc5e5c10c53b2053137add637a519d5b2a18f4c8\": not found" May 8 08:01:49.449623 kubelet[3127]: I0508 08:01:49.449624 3127 scope.go:117] "RemoveContainer" containerID="a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3" May 8 08:01:49.449791 containerd[1535]: time="2025-05-08T08:01:49.449764231Z" level=error msg="ContainerStatus for \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\": not found" May 8 08:01:49.450024 kubelet[3127]: E0508 08:01:49.449997 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\": not found" containerID="a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3" May 8 08:01:49.450275 kubelet[3127]: I0508 08:01:49.450125 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3"} err="failed to get container status \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\": rpc error: code = NotFound desc = an error occurred when try to find container \"a065458548d19302b285ea2aac58cd303488fb62710518390fc1d04e9ac9a6b3\": not found" May 8 08:01:49.450275 kubelet[3127]: I0508 08:01:49.450157 3127 scope.go:117] "RemoveContainer" containerID="03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106" May 8 08:01:49.450381 containerd[1535]: time="2025-05-08T08:01:49.450339953Z" level=error msg="ContainerStatus for \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\": not found" May 8 08:01:49.450484 kubelet[3127]: E0508 08:01:49.450443 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\": not found" containerID="03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106" May 8 08:01:49.450484 kubelet[3127]: I0508 08:01:49.450462 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106"} err="failed to get container status \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\": rpc error: code = NotFound desc = an error occurred when try to find container \"03a19189b5bb11e2801a9a9fad340ffcea635dbb3a30b3b3e2119db2d9c0a106\": not found" May 8 08:01:49.450484 kubelet[3127]: I0508 08:01:49.450476 3127 scope.go:117] "RemoveContainer" containerID="5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c" May 8 08:01:49.450885 containerd[1535]: time="2025-05-08T08:01:49.450709282Z" level=error msg="ContainerStatus for \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\": not found" May 8 08:01:49.451007 kubelet[3127]: E0508 08:01:49.450831 3127 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\": not found" containerID="5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c" May 8 08:01:49.451007 kubelet[3127]: I0508 08:01:49.450851 3127 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c"} err="failed to get container status \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\": rpc error: code = NotFound desc = an error occurred when try to find container \"5dffd8deb51c1b65addf50fe0f025e92f7e0647f0706fd0bc204e2684ac7852c\": not found" May 8 08:01:49.865054 systemd[1]: var-lib-kubelet-pods-60355b88\x2d43ea\x2d493c\x2da59c\x2db9a39cecea05-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dqtbtb.mount: Deactivated successfully. May 8 08:01:49.865373 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b-shm.mount: Deactivated successfully. May 8 08:01:49.865565 systemd[1]: var-lib-kubelet-pods-01da910d\x2dc3da\x2d436f\x2dae6b\x2d18a93de1a297-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dr98vr.mount: Deactivated successfully. May 8 08:01:49.865758 systemd[1]: var-lib-kubelet-pods-01da910d\x2dc3da\x2d436f\x2dae6b\x2d18a93de1a297-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 8 08:01:49.866690 systemd[1]: var-lib-kubelet-pods-01da910d\x2dc3da\x2d436f\x2dae6b\x2d18a93de1a297-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 8 08:01:50.590134 kubelet[3127]: I0508 08:01:50.589912 3127 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" path="/var/lib/kubelet/pods/01da910d-c3da-436f-ae6b-18a93de1a297/volumes" May 8 08:01:50.591610 kubelet[3127]: I0508 08:01:50.591546 3127 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60355b88-43ea-493c-a59c-b9a39cecea05" path="/var/lib/kubelet/pods/60355b88-43ea-493c-a59c-b9a39cecea05/volumes" May 8 08:01:50.690888 kubelet[3127]: E0508 08:01:50.690785 3127 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 8 08:01:50.835785 sshd[4658]: Connection closed by 172.24.4.1 port 47126 May 8 08:01:50.838070 sshd-session[4655]: pam_unix(sshd:session): session closed for user core May 8 08:01:50.855216 systemd[1]: sshd@22-172.24.4.249:22-172.24.4.1:47126.service: Deactivated successfully. May 8 08:01:50.861468 systemd[1]: session-25.scope: Deactivated successfully. May 8 08:01:50.865061 systemd-logind[1509]: Session 25 logged out. Waiting for processes to exit. May 8 08:01:50.871056 systemd[1]: Started sshd@23-172.24.4.249:22-172.24.4.1:47132.service - OpenSSH per-connection server daemon (172.24.4.1:47132). May 8 08:01:50.875016 systemd-logind[1509]: Removed session 25. May 8 08:01:51.849635 sshd[4813]: Accepted publickey for core from 172.24.4.1 port 47132 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:51.852765 sshd-session[4813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:51.869198 systemd-logind[1509]: New session 26 of user core. May 8 08:01:51.876262 systemd[1]: Started session-26.scope - Session 26 of User core. May 8 08:01:53.351354 kubelet[3127]: I0508 08:01:53.351190 3127 topology_manager.go:215] "Topology Admit Handler" podUID="48fbf70a-4b7f-405b-84ef-e0d51402cc52" podNamespace="kube-system" podName="cilium-fc2qk" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352323 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="mount-cgroup" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352398 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="60355b88-43ea-493c-a59c-b9a39cecea05" containerName="cilium-operator" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352424 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="apply-sysctl-overwrites" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352518 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="mount-bpf-fs" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352533 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="clean-cilium-state" May 8 08:01:53.352899 kubelet[3127]: E0508 08:01:53.352597 3127 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="cilium-agent" May 8 08:01:53.354265 kubelet[3127]: I0508 08:01:53.353826 3127 memory_manager.go:354] "RemoveStaleState removing state" podUID="01da910d-c3da-436f-ae6b-18a93de1a297" containerName="cilium-agent" May 8 08:01:53.354265 kubelet[3127]: I0508 08:01:53.353844 3127 memory_manager.go:354] "RemoveStaleState removing state" podUID="60355b88-43ea-493c-a59c-b9a39cecea05" containerName="cilium-operator" May 8 08:01:53.365586 systemd[1]: Created slice kubepods-burstable-pod48fbf70a_4b7f_405b_84ef_e0d51402cc52.slice - libcontainer container kubepods-burstable-pod48fbf70a_4b7f_405b_84ef_e0d51402cc52.slice. May 8 08:01:53.412740 kubelet[3127]: I0508 08:01:53.412562 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-cni-path\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.412740 kubelet[3127]: I0508 08:01:53.412681 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-bpf-maps\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.412740 kubelet[3127]: I0508 08:01:53.412717 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-host-proc-sys-net\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.412750 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72gzv\" (UniqueName: \"kubernetes.io/projected/48fbf70a-4b7f-405b-84ef-e0d51402cc52-kube-api-access-72gzv\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.412797 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-cilium-run\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.412851 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-etc-cni-netd\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.412886 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/48fbf70a-4b7f-405b-84ef-e0d51402cc52-cilium-ipsec-secrets\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.412990 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-xtables-lock\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413209 kubelet[3127]: I0508 08:01:53.413037 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/48fbf70a-4b7f-405b-84ef-e0d51402cc52-cilium-config-path\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413102 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-host-proc-sys-kernel\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413141 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/48fbf70a-4b7f-405b-84ef-e0d51402cc52-clustermesh-secrets\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413179 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/48fbf70a-4b7f-405b-84ef-e0d51402cc52-hubble-tls\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413235 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-cilium-cgroup\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413291 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-hostproc\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.413399 kubelet[3127]: I0508 08:01:53.413329 3127 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48fbf70a-4b7f-405b-84ef-e0d51402cc52-lib-modules\") pod \"cilium-fc2qk\" (UID: \"48fbf70a-4b7f-405b-84ef-e0d51402cc52\") " pod="kube-system/cilium-fc2qk" May 8 08:01:53.461827 sshd[4817]: Connection closed by 172.24.4.1 port 47132 May 8 08:01:53.463654 sshd-session[4813]: pam_unix(sshd:session): session closed for user core May 8 08:01:53.484391 systemd[1]: sshd@23-172.24.4.249:22-172.24.4.1:47132.service: Deactivated successfully. May 8 08:01:53.488542 systemd[1]: session-26.scope: Deactivated successfully. May 8 08:01:53.491875 systemd-logind[1509]: Session 26 logged out. Waiting for processes to exit. May 8 08:01:53.495497 systemd[1]: Started sshd@24-172.24.4.249:22-172.24.4.1:47138.service - OpenSSH per-connection server daemon (172.24.4.1:47138). May 8 08:01:53.498501 systemd-logind[1509]: Removed session 26. May 8 08:01:53.550444 kubelet[3127]: I0508 08:01:53.550297 3127 setters.go:580] "Node became not ready" node="ci-4327-0-0-w-45c5ab2731.novalocal" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-08T08:01:53Z","lastTransitionTime":"2025-05-08T08:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 8 08:01:53.675065 containerd[1535]: time="2025-05-08T08:01:53.674973890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fc2qk,Uid:48fbf70a-4b7f-405b-84ef-e0d51402cc52,Namespace:kube-system,Attempt:0,}" May 8 08:01:53.702283 containerd[1535]: time="2025-05-08T08:01:53.700059193Z" level=info msg="connecting to shim 7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" namespace=k8s.io protocol=ttrpc version=3 May 8 08:01:53.735071 systemd[1]: Started cri-containerd-7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b.scope - libcontainer container 7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b. May 8 08:01:53.774460 containerd[1535]: time="2025-05-08T08:01:53.773898838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fc2qk,Uid:48fbf70a-4b7f-405b-84ef-e0d51402cc52,Namespace:kube-system,Attempt:0,} returns sandbox id \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\"" May 8 08:01:53.779205 containerd[1535]: time="2025-05-08T08:01:53.779122325Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 8 08:01:53.787265 containerd[1535]: time="2025-05-08T08:01:53.787221960Z" level=info msg="Container b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6: CDI devices from CRI Config.CDIDevices: []" May 8 08:01:53.800053 containerd[1535]: time="2025-05-08T08:01:53.800008633Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\"" May 8 08:01:53.800846 containerd[1535]: time="2025-05-08T08:01:53.800746443Z" level=info msg="StartContainer for \"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\"" May 8 08:01:53.801616 containerd[1535]: time="2025-05-08T08:01:53.801585355Z" level=info msg="connecting to shim b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" protocol=ttrpc version=3 May 8 08:01:53.823052 systemd[1]: Started cri-containerd-b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6.scope - libcontainer container b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6. May 8 08:01:53.861177 containerd[1535]: time="2025-05-08T08:01:53.861043930Z" level=info msg="StartContainer for \"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\" returns successfully" May 8 08:01:53.863287 systemd[1]: cri-containerd-b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6.scope: Deactivated successfully. May 8 08:01:53.865897 containerd[1535]: time="2025-05-08T08:01:53.865843103Z" level=info msg="received exit event container_id:\"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\" id:\"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\" pid:4893 exited_at:{seconds:1746691313 nanos:865300634}" May 8 08:01:53.866533 containerd[1535]: time="2025-05-08T08:01:53.866141529Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\" id:\"b1f49c105202eb3dffddeaebad32ed74804b702e05ffbb6af5e550e679a663e6\" pid:4893 exited_at:{seconds:1746691313 nanos:865300634}" May 8 08:01:54.271417 containerd[1535]: time="2025-05-08T08:01:54.271290974Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 8 08:01:54.289976 containerd[1535]: time="2025-05-08T08:01:54.288524556Z" level=info msg="Container da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0: CDI devices from CRI Config.CDIDevices: []" May 8 08:01:54.308976 containerd[1535]: time="2025-05-08T08:01:54.308863598Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\"" May 8 08:01:54.318022 containerd[1535]: time="2025-05-08T08:01:54.316281931Z" level=info msg="StartContainer for \"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\"" May 8 08:01:54.323488 containerd[1535]: time="2025-05-08T08:01:54.323421585Z" level=info msg="connecting to shim da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" protocol=ttrpc version=3 May 8 08:01:54.365156 systemd[1]: Started cri-containerd-da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0.scope - libcontainer container da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0. May 8 08:01:54.405670 containerd[1535]: time="2025-05-08T08:01:54.405618641Z" level=info msg="StartContainer for \"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\" returns successfully" May 8 08:01:54.408188 systemd[1]: cri-containerd-da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0.scope: Deactivated successfully. May 8 08:01:54.410596 containerd[1535]: time="2025-05-08T08:01:54.410546519Z" level=info msg="received exit event container_id:\"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\" id:\"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\" pid:4937 exited_at:{seconds:1746691314 nanos:409420593}" May 8 08:01:54.411312 containerd[1535]: time="2025-05-08T08:01:54.410796273Z" level=info msg="TaskExit event in podsandbox handler container_id:\"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\" id:\"da7de82c6bf9b2fe748b1e0577fceca3cc1efdad5b0102d9a3dd5ebbb7046ab0\" pid:4937 exited_at:{seconds:1746691314 nanos:409420593}" May 8 08:01:54.726429 sshd[4827]: Accepted publickey for core from 172.24.4.1 port 47138 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:54.729130 sshd-session[4827]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:54.740078 systemd-logind[1509]: New session 27 of user core. May 8 08:01:54.747199 systemd[1]: Started session-27.scope - Session 27 of User core. May 8 08:01:55.284704 containerd[1535]: time="2025-05-08T08:01:55.284491539Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 8 08:01:55.330989 containerd[1535]: time="2025-05-08T08:01:55.327713835Z" level=info msg="Container ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883: CDI devices from CRI Config.CDIDevices: []" May 8 08:01:55.357158 containerd[1535]: time="2025-05-08T08:01:55.356974656Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\"" May 8 08:01:55.358738 containerd[1535]: time="2025-05-08T08:01:55.358673549Z" level=info msg="StartContainer for \"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\"" May 8 08:01:55.360474 containerd[1535]: time="2025-05-08T08:01:55.360423959Z" level=info msg="connecting to shim ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" protocol=ttrpc version=3 May 8 08:01:55.384106 systemd[1]: Started cri-containerd-ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883.scope - libcontainer container ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883. May 8 08:01:55.428779 systemd[1]: cri-containerd-ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883.scope: Deactivated successfully. May 8 08:01:55.431640 containerd[1535]: time="2025-05-08T08:01:55.431609065Z" level=info msg="received exit event container_id:\"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\" id:\"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\" pid:4982 exited_at:{seconds:1746691315 nanos:431216530}" May 8 08:01:55.432071 containerd[1535]: time="2025-05-08T08:01:55.431996159Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\" id:\"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\" pid:4982 exited_at:{seconds:1746691315 nanos:431216530}" May 8 08:01:55.434167 containerd[1535]: time="2025-05-08T08:01:55.433905811Z" level=info msg="StartContainer for \"ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883\" returns successfully" May 8 08:01:55.455192 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ca84f9bc88df63d16b4e72050b40d2a0682b4ead6244c683d9caf963e7524883-rootfs.mount: Deactivated successfully. May 8 08:01:55.474945 sshd[4967]: Connection closed by 172.24.4.1 port 47138 May 8 08:01:55.474885 sshd-session[4827]: pam_unix(sshd:session): session closed for user core May 8 08:01:55.486658 systemd[1]: sshd@24-172.24.4.249:22-172.24.4.1:47138.service: Deactivated successfully. May 8 08:01:55.489199 systemd[1]: session-27.scope: Deactivated successfully. May 8 08:01:55.491441 systemd-logind[1509]: Session 27 logged out. Waiting for processes to exit. May 8 08:01:55.494122 systemd[1]: Started sshd@25-172.24.4.249:22-172.24.4.1:52008.service - OpenSSH per-connection server daemon (172.24.4.1:52008). May 8 08:01:55.495819 systemd-logind[1509]: Removed session 27. May 8 08:01:55.692736 kubelet[3127]: E0508 08:01:55.692589 3127 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 8 08:01:56.289895 containerd[1535]: time="2025-05-08T08:01:56.289328993Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 8 08:01:56.313338 containerd[1535]: time="2025-05-08T08:01:56.313266759Z" level=info msg="Container e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198: CDI devices from CRI Config.CDIDevices: []" May 8 08:01:56.339123 containerd[1535]: time="2025-05-08T08:01:56.338865193Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\"" May 8 08:01:56.343829 containerd[1535]: time="2025-05-08T08:01:56.342235118Z" level=info msg="StartContainer for \"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\"" May 8 08:01:56.343829 containerd[1535]: time="2025-05-08T08:01:56.343701870Z" level=info msg="connecting to shim e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" protocol=ttrpc version=3 May 8 08:01:56.369078 systemd[1]: Started cri-containerd-e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198.scope - libcontainer container e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198. May 8 08:01:56.396158 systemd[1]: cri-containerd-e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198.scope: Deactivated successfully. May 8 08:01:56.397787 containerd[1535]: time="2025-05-08T08:01:56.397741475Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\" id:\"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\" pid:5029 exited_at:{seconds:1746691316 nanos:397298333}" May 8 08:01:56.399511 containerd[1535]: time="2025-05-08T08:01:56.399473690Z" level=info msg="received exit event container_id:\"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\" id:\"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\" pid:5029 exited_at:{seconds:1746691316 nanos:397298333}" May 8 08:01:56.407113 containerd[1535]: time="2025-05-08T08:01:56.407075273Z" level=info msg="StartContainer for \"e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198\" returns successfully" May 8 08:01:56.423784 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e209dc162fdf5da2fa31c365fb2761f53fed2602e1a93728c2095b6e55df0198-rootfs.mount: Deactivated successfully. May 8 08:01:56.759582 sshd[5013]: Accepted publickey for core from 172.24.4.1 port 52008 ssh2: RSA SHA256:wgD4/RwgaU1oYCs7bHtnBTlVci798e2qou6xtbcPlVg May 8 08:01:56.763118 sshd-session[5013]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 8 08:01:56.775613 systemd-logind[1509]: New session 28 of user core. May 8 08:01:56.785352 systemd[1]: Started session-28.scope - Session 28 of User core. May 8 08:01:57.315244 containerd[1535]: time="2025-05-08T08:01:57.315139905Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 8 08:01:57.360248 containerd[1535]: time="2025-05-08T08:01:57.358627534Z" level=info msg="Container 7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458: CDI devices from CRI Config.CDIDevices: []" May 8 08:01:57.365291 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2431839134.mount: Deactivated successfully. May 8 08:01:57.381078 containerd[1535]: time="2025-05-08T08:01:57.380831543Z" level=info msg="CreateContainer within sandbox \"7bfaa31ba60fcee279588650ec0896506cf283f76687422a8177f78cb829b85b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\"" May 8 08:01:57.381574 containerd[1535]: time="2025-05-08T08:01:57.381442161Z" level=info msg="StartContainer for \"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\"" May 8 08:01:57.382855 containerd[1535]: time="2025-05-08T08:01:57.382515958Z" level=info msg="connecting to shim 7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458" address="unix:///run/containerd/s/19287934453aab4753e9ac72cdf5923a6ff00aa1a710db889511f97d2c7e6bc1" protocol=ttrpc version=3 May 8 08:01:57.410270 systemd[1]: Started cri-containerd-7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458.scope - libcontainer container 7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458. May 8 08:01:57.469825 containerd[1535]: time="2025-05-08T08:01:57.469788662Z" level=info msg="StartContainer for \"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" returns successfully" May 8 08:01:57.539298 containerd[1535]: time="2025-05-08T08:01:57.539260623Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"f003e458ca26db6e39aefdf2ca7f4bcf9fc591e4c50a233e46d6dca9b23e1796\" pid:5104 exited_at:{seconds:1746691317 nanos:538959751}" May 8 08:01:57.822024 kernel: cryptd: max_cpu_qlen set to 1000 May 8 08:01:57.873010 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm_base(ctr(aes-generic),ghash-generic)))) May 8 08:01:58.359982 kubelet[3127]: I0508 08:01:58.359817 3127 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-fc2qk" podStartSLOduration=5.359782614 podStartE2EDuration="5.359782614s" podCreationTimestamp="2025-05-08 08:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 08:01:58.357564715 +0000 UTC m=+167.888227461" watchObservedRunningTime="2025-05-08 08:01:58.359782614 +0000 UTC m=+167.890445370" May 8 08:01:59.608425 containerd[1535]: time="2025-05-08T08:01:59.608360747Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"b709b78cae800a10c83afef8af4431083f2c0d98fe0d67ffb88d8ea0176eb047\" pid:5299 exit_status:1 exited_at:{seconds:1746691319 nanos:607755909}" May 8 08:02:00.966420 systemd-networkd[1419]: lxc_health: Link UP May 8 08:02:00.971358 systemd-networkd[1419]: lxc_health: Gained carrier May 8 08:02:01.881829 containerd[1535]: time="2025-05-08T08:02:01.881763541Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"85e26c20c01a5b6198cfcc677e0c77161ed1c2fd8940c210b64deeaca3cfdfd6\" pid:5676 exited_at:{seconds:1746691321 nanos:879477372}" May 8 08:02:01.896610 kubelet[3127]: E0508 08:02:01.895156 3127 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 127.0.0.1:41838->127.0.0.1:38619: read tcp 127.0.0.1:41838->127.0.0.1:38619: read: connection reset by peer May 8 08:02:02.467057 systemd-networkd[1419]: lxc_health: Gained IPv6LL May 8 08:02:04.025137 containerd[1535]: time="2025-05-08T08:02:04.025093381Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"597a78c545944604ba6a83e7187176b875db6267c9ba4f8a510d79e695e0e634\" pid:5709 exited_at:{seconds:1746691324 nanos:23861493}" May 8 08:02:06.212514 containerd[1535]: time="2025-05-08T08:02:06.212379647Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"484e15d7bb770978bead9ba03ae8e42964062d121fa2fa207376f887cdd6586e\" pid:5743 exited_at:{seconds:1746691326 nanos:211403144}" May 8 08:02:06.215755 kubelet[3127]: E0508 08:02:06.215695 3127 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:34584->127.0.0.1:38619: write tcp 127.0.0.1:34584->127.0.0.1:38619: write: broken pipe May 8 08:02:08.457996 containerd[1535]: time="2025-05-08T08:02:08.457950518Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b35240562ccedfdb4dfcada53d15bbdae5b72ec0892294de3354d0c72c29458\" id:\"41c6d603d7cf28685d03498f5f8733279d7b7be2425c4ec16fbb5c4c45185064\" pid:5767 exited_at:{seconds:1746691328 nanos:457529900}" May 8 08:02:08.769418 sshd[5055]: Connection closed by 172.24.4.1 port 52008 May 8 08:02:08.771892 sshd-session[5013]: pam_unix(sshd:session): session closed for user core May 8 08:02:08.787454 systemd[1]: sshd@25-172.24.4.249:22-172.24.4.1:52008.service: Deactivated successfully. May 8 08:02:08.794681 systemd[1]: session-28.scope: Deactivated successfully. May 8 08:02:08.797881 systemd-logind[1509]: Session 28 logged out. Waiting for processes to exit. May 8 08:02:08.803146 systemd-logind[1509]: Removed session 28. May 8 08:02:10.588513 containerd[1535]: time="2025-05-08T08:02:10.587970117Z" level=info msg="StopPodSandbox for \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\"" May 8 08:02:10.594438 containerd[1535]: time="2025-05-08T08:02:10.588879663Z" level=info msg="TearDown network for sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" successfully" May 8 08:02:10.594438 containerd[1535]: time="2025-05-08T08:02:10.589049175Z" level=info msg="StopPodSandbox for \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" returns successfully" May 8 08:02:10.594438 containerd[1535]: time="2025-05-08T08:02:10.590270013Z" level=info msg="RemovePodSandbox for \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\"" May 8 08:02:10.594438 containerd[1535]: time="2025-05-08T08:02:10.590325468Z" level=info msg="Forcibly stopping sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\"" May 8 08:02:10.594438 containerd[1535]: time="2025-05-08T08:02:10.590905840Z" level=info msg="TearDown network for sandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" successfully" May 8 08:02:10.600258 containerd[1535]: time="2025-05-08T08:02:10.600149861Z" level=info msg="Ensure that sandbox 7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb in task-service has been cleanup successfully" May 8 08:02:10.610674 containerd[1535]: time="2025-05-08T08:02:10.610533134Z" level=info msg="RemovePodSandbox \"7f56e85a6f3970ad2214a19418636c017c8e246b6a6b666ad50ffc6dadec51eb\" returns successfully" May 8 08:02:10.611911 containerd[1535]: time="2025-05-08T08:02:10.611677428Z" level=info msg="StopPodSandbox for \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\"" May 8 08:02:10.612661 containerd[1535]: time="2025-05-08T08:02:10.612567176Z" level=info msg="TearDown network for sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" successfully" May 8 08:02:10.612661 containerd[1535]: time="2025-05-08T08:02:10.612637280Z" level=info msg="StopPodSandbox for \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" returns successfully" May 8 08:02:10.613634 containerd[1535]: time="2025-05-08T08:02:10.613564890Z" level=info msg="RemovePodSandbox for \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\"" May 8 08:02:10.613856 containerd[1535]: time="2025-05-08T08:02:10.613647517Z" level=info msg="Forcibly stopping sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\"" May 8 08:02:10.614525 containerd[1535]: time="2025-05-08T08:02:10.614313953Z" level=info msg="TearDown network for sandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" successfully" May 8 08:02:10.618978 containerd[1535]: time="2025-05-08T08:02:10.618862819Z" level=info msg="Ensure that sandbox 84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b in task-service has been cleanup successfully" May 8 08:02:10.628707 containerd[1535]: time="2025-05-08T08:02:10.628406038Z" level=info msg="RemovePodSandbox \"84b224d9e71688a60d9292fb036cca813469669d580304fbdb1ba29540b3876b\" returns successfully"