Dec 12 18:45:03.915241 kernel: Linux version 6.12.61-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Fri Dec 12 15:21:28 -00 2025 Dec 12 18:45:03.915271 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:45:03.915280 kernel: BIOS-provided physical RAM map: Dec 12 18:45:03.915289 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Dec 12 18:45:03.915304 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Dec 12 18:45:03.915311 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Dec 12 18:45:03.915328 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Dec 12 18:45:03.915335 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Dec 12 18:45:03.915341 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Dec 12 18:45:03.915365 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Dec 12 18:45:03.915372 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Dec 12 18:45:03.915378 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Dec 12 18:45:03.915388 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Dec 12 18:45:03.915395 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Dec 12 18:45:03.915403 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Dec 12 18:45:03.915410 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Dec 12 18:45:03.915417 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Dec 12 18:45:03.915426 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Dec 12 18:45:03.915433 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Dec 12 18:45:03.915440 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Dec 12 18:45:03.915447 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Dec 12 18:45:03.915454 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Dec 12 18:45:03.915461 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Dec 12 18:45:03.915468 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 12 18:45:03.915475 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Dec 12 18:45:03.915482 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Dec 12 18:45:03.915489 kernel: NX (Execute Disable) protection: active Dec 12 18:45:03.915505 kernel: APIC: Static calls initialized Dec 12 18:45:03.915514 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Dec 12 18:45:03.915521 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Dec 12 18:45:03.915528 kernel: extended physical RAM map: Dec 12 18:45:03.915536 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Dec 12 18:45:03.915543 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Dec 12 18:45:03.915550 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Dec 12 18:45:03.915557 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Dec 12 18:45:03.915564 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Dec 12 18:45:03.915571 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Dec 12 18:45:03.915578 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Dec 12 18:45:03.915585 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Dec 12 18:45:03.915594 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Dec 12 18:45:03.915605 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Dec 12 18:45:03.915612 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Dec 12 18:45:03.915619 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Dec 12 18:45:03.915626 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Dec 12 18:45:03.915635 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Dec 12 18:45:03.915643 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Dec 12 18:45:03.915650 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Dec 12 18:45:03.915657 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Dec 12 18:45:03.915665 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Dec 12 18:45:03.915672 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Dec 12 18:45:03.915679 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Dec 12 18:45:03.915686 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Dec 12 18:45:03.915694 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Dec 12 18:45:03.915701 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Dec 12 18:45:03.915708 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Dec 12 18:45:03.915718 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 12 18:45:03.915725 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Dec 12 18:45:03.915732 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Dec 12 18:45:03.915739 kernel: efi: EFI v2.7 by EDK II Dec 12 18:45:03.915747 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Dec 12 18:45:03.915754 kernel: random: crng init done Dec 12 18:45:03.915761 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Dec 12 18:45:03.915768 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Dec 12 18:45:03.915776 kernel: secureboot: Secure boot disabled Dec 12 18:45:03.915783 kernel: SMBIOS 2.8 present. Dec 12 18:45:03.915790 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Dec 12 18:45:03.915800 kernel: DMI: Memory slots populated: 1/1 Dec 12 18:45:03.915807 kernel: Hypervisor detected: KVM Dec 12 18:45:03.915814 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Dec 12 18:45:03.915822 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 12 18:45:03.915829 kernel: kvm-clock: using sched offset of 5660456061 cycles Dec 12 18:45:03.915836 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 12 18:45:03.915844 kernel: tsc: Detected 2794.750 MHz processor Dec 12 18:45:03.915852 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 12 18:45:03.915859 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 12 18:45:03.915867 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Dec 12 18:45:03.915874 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Dec 12 18:45:03.915884 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 12 18:45:03.915891 kernel: Using GB pages for direct mapping Dec 12 18:45:03.915899 kernel: ACPI: Early table checksum verification disabled Dec 12 18:45:03.915906 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Dec 12 18:45:03.915914 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Dec 12 18:45:03.915922 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915929 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915936 kernel: ACPI: FACS 0x000000009CBDD000 000040 Dec 12 18:45:03.915944 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915954 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915962 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915969 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 12 18:45:03.915977 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Dec 12 18:45:03.915984 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Dec 12 18:45:03.915992 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Dec 12 18:45:03.915999 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Dec 12 18:45:03.916007 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Dec 12 18:45:03.916014 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Dec 12 18:45:03.916024 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Dec 12 18:45:03.916031 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Dec 12 18:45:03.916038 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Dec 12 18:45:03.916046 kernel: No NUMA configuration found Dec 12 18:45:03.916053 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Dec 12 18:45:03.916061 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Dec 12 18:45:03.916068 kernel: Zone ranges: Dec 12 18:45:03.916075 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 12 18:45:03.916083 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Dec 12 18:45:03.916092 kernel: Normal empty Dec 12 18:45:03.916100 kernel: Device empty Dec 12 18:45:03.916107 kernel: Movable zone start for each node Dec 12 18:45:03.916114 kernel: Early memory node ranges Dec 12 18:45:03.916122 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Dec 12 18:45:03.916129 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Dec 12 18:45:03.916136 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Dec 12 18:45:03.916144 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Dec 12 18:45:03.916151 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Dec 12 18:45:03.916159 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Dec 12 18:45:03.916168 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Dec 12 18:45:03.916176 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Dec 12 18:45:03.916183 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Dec 12 18:45:03.916191 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 12 18:45:03.916205 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Dec 12 18:45:03.916215 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Dec 12 18:45:03.916223 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 12 18:45:03.916231 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Dec 12 18:45:03.916238 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Dec 12 18:45:03.916246 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Dec 12 18:45:03.916254 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Dec 12 18:45:03.916262 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Dec 12 18:45:03.916272 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 12 18:45:03.916284 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 12 18:45:03.916296 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 12 18:45:03.916311 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 12 18:45:03.916333 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 12 18:45:03.916363 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 12 18:45:03.916371 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 12 18:45:03.916379 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 12 18:45:03.916387 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 12 18:45:03.916394 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 12 18:45:03.916402 kernel: TSC deadline timer available Dec 12 18:45:03.916410 kernel: CPU topo: Max. logical packages: 1 Dec 12 18:45:03.916418 kernel: CPU topo: Max. logical dies: 1 Dec 12 18:45:03.916425 kernel: CPU topo: Max. dies per package: 1 Dec 12 18:45:03.916435 kernel: CPU topo: Max. threads per core: 1 Dec 12 18:45:03.916443 kernel: CPU topo: Num. cores per package: 4 Dec 12 18:45:03.916451 kernel: CPU topo: Num. threads per package: 4 Dec 12 18:45:03.916458 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Dec 12 18:45:03.916466 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 12 18:45:03.916474 kernel: kvm-guest: KVM setup pv remote TLB flush Dec 12 18:45:03.916482 kernel: kvm-guest: setup PV sched yield Dec 12 18:45:03.916490 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Dec 12 18:45:03.916504 kernel: Booting paravirtualized kernel on KVM Dec 12 18:45:03.916515 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 12 18:45:03.916523 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Dec 12 18:45:03.916530 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Dec 12 18:45:03.916538 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Dec 12 18:45:03.916546 kernel: pcpu-alloc: [0] 0 1 2 3 Dec 12 18:45:03.916554 kernel: kvm-guest: PV spinlocks enabled Dec 12 18:45:03.916561 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Dec 12 18:45:03.916570 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:45:03.916581 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Dec 12 18:45:03.916589 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 12 18:45:03.916596 kernel: Fallback order for Node 0: 0 Dec 12 18:45:03.916604 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Dec 12 18:45:03.916612 kernel: Policy zone: DMA32 Dec 12 18:45:03.916620 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 12 18:45:03.916627 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Dec 12 18:45:03.916645 kernel: ftrace: allocating 40103 entries in 157 pages Dec 12 18:45:03.916662 kernel: ftrace: allocated 157 pages with 5 groups Dec 12 18:45:03.916672 kernel: Dynamic Preempt: voluntary Dec 12 18:45:03.916680 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 12 18:45:03.916689 kernel: rcu: RCU event tracing is enabled. Dec 12 18:45:03.916697 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Dec 12 18:45:03.916705 kernel: Trampoline variant of Tasks RCU enabled. Dec 12 18:45:03.916713 kernel: Rude variant of Tasks RCU enabled. Dec 12 18:45:03.916721 kernel: Tracing variant of Tasks RCU enabled. Dec 12 18:45:03.916729 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 12 18:45:03.916737 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Dec 12 18:45:03.916744 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 12 18:45:03.916755 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 12 18:45:03.916763 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Dec 12 18:45:03.916771 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Dec 12 18:45:03.916778 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 12 18:45:03.916786 kernel: Console: colour dummy device 80x25 Dec 12 18:45:03.916794 kernel: printk: legacy console [ttyS0] enabled Dec 12 18:45:03.916802 kernel: ACPI: Core revision 20240827 Dec 12 18:45:03.916810 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 12 18:45:03.916820 kernel: APIC: Switch to symmetric I/O mode setup Dec 12 18:45:03.916828 kernel: x2apic enabled Dec 12 18:45:03.916835 kernel: APIC: Switched APIC routing to: physical x2apic Dec 12 18:45:03.916843 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Dec 12 18:45:03.916851 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Dec 12 18:45:03.916859 kernel: kvm-guest: setup PV IPIs Dec 12 18:45:03.916866 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 12 18:45:03.916874 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848e100549, max_idle_ns: 440795215505 ns Dec 12 18:45:03.916882 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Dec 12 18:45:03.916893 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 12 18:45:03.916901 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 12 18:45:03.916908 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 12 18:45:03.916916 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 12 18:45:03.916924 kernel: Spectre V2 : Mitigation: Retpolines Dec 12 18:45:03.916932 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Dec 12 18:45:03.916939 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 12 18:45:03.916947 kernel: active return thunk: retbleed_return_thunk Dec 12 18:45:03.916955 kernel: RETBleed: Mitigation: untrained return thunk Dec 12 18:45:03.916965 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 12 18:45:03.916973 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 12 18:45:03.916981 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Dec 12 18:45:03.916989 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Dec 12 18:45:03.916997 kernel: active return thunk: srso_return_thunk Dec 12 18:45:03.917005 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Dec 12 18:45:03.917013 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 12 18:45:03.917020 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 12 18:45:03.917028 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 12 18:45:03.917038 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 12 18:45:03.917046 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Dec 12 18:45:03.917054 kernel: Freeing SMP alternatives memory: 32K Dec 12 18:45:03.917062 kernel: pid_max: default: 32768 minimum: 301 Dec 12 18:45:03.917069 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Dec 12 18:45:03.917077 kernel: landlock: Up and running. Dec 12 18:45:03.917085 kernel: SELinux: Initializing. Dec 12 18:45:03.917093 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 12 18:45:03.917100 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 12 18:45:03.917110 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 12 18:45:03.917118 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 12 18:45:03.917126 kernel: ... version: 0 Dec 12 18:45:03.917133 kernel: ... bit width: 48 Dec 12 18:45:03.917141 kernel: ... generic registers: 6 Dec 12 18:45:03.917149 kernel: ... value mask: 0000ffffffffffff Dec 12 18:45:03.917157 kernel: ... max period: 00007fffffffffff Dec 12 18:45:03.917164 kernel: ... fixed-purpose events: 0 Dec 12 18:45:03.917172 kernel: ... event mask: 000000000000003f Dec 12 18:45:03.917181 kernel: signal: max sigframe size: 1776 Dec 12 18:45:03.917189 kernel: rcu: Hierarchical SRCU implementation. Dec 12 18:45:03.917197 kernel: rcu: Max phase no-delay instances is 400. Dec 12 18:45:03.917205 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Dec 12 18:45:03.917213 kernel: smp: Bringing up secondary CPUs ... Dec 12 18:45:03.917220 kernel: smpboot: x86: Booting SMP configuration: Dec 12 18:45:03.917228 kernel: .... node #0, CPUs: #1 #2 #3 Dec 12 18:45:03.917236 kernel: smp: Brought up 1 node, 4 CPUs Dec 12 18:45:03.917244 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Dec 12 18:45:03.917254 kernel: Memory: 2414476K/2565800K available (14336K kernel code, 2444K rwdata, 26064K rodata, 46188K init, 2572K bss, 145388K reserved, 0K cma-reserved) Dec 12 18:45:03.917262 kernel: devtmpfs: initialized Dec 12 18:45:03.917269 kernel: x86/mm: Memory block size: 128MB Dec 12 18:45:03.917277 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Dec 12 18:45:03.917285 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Dec 12 18:45:03.917293 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Dec 12 18:45:03.917301 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Dec 12 18:45:03.917309 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Dec 12 18:45:03.917317 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Dec 12 18:45:03.917327 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 12 18:45:03.917335 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Dec 12 18:45:03.917342 kernel: pinctrl core: initialized pinctrl subsystem Dec 12 18:45:03.917378 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 12 18:45:03.917386 kernel: audit: initializing netlink subsys (disabled) Dec 12 18:45:03.917394 kernel: audit: type=2000 audit(1765565099.321:1): state=initialized audit_enabled=0 res=1 Dec 12 18:45:03.917402 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 12 18:45:03.917410 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 12 18:45:03.917420 kernel: cpuidle: using governor menu Dec 12 18:45:03.917428 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 12 18:45:03.917436 kernel: dca service started, version 1.12.1 Dec 12 18:45:03.917444 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Dec 12 18:45:03.917451 kernel: PCI: Using configuration type 1 for base access Dec 12 18:45:03.917459 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 12 18:45:03.917467 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 12 18:45:03.917475 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Dec 12 18:45:03.917483 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 12 18:45:03.917493 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 12 18:45:03.917508 kernel: ACPI: Added _OSI(Module Device) Dec 12 18:45:03.917515 kernel: ACPI: Added _OSI(Processor Device) Dec 12 18:45:03.917523 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 12 18:45:03.917532 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 12 18:45:03.917539 kernel: ACPI: Interpreter enabled Dec 12 18:45:03.917547 kernel: ACPI: PM: (supports S0 S3 S5) Dec 12 18:45:03.917555 kernel: ACPI: Using IOAPIC for interrupt routing Dec 12 18:45:03.917563 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 12 18:45:03.917573 kernel: PCI: Using E820 reservations for host bridge windows Dec 12 18:45:03.917581 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 12 18:45:03.917588 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 12 18:45:03.917828 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 12 18:45:03.918060 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 12 18:45:03.918189 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 12 18:45:03.918201 kernel: PCI host bridge to bus 0000:00 Dec 12 18:45:03.918334 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 12 18:45:03.918470 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 12 18:45:03.918589 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 12 18:45:03.918705 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Dec 12 18:45:03.918811 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Dec 12 18:45:03.918929 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Dec 12 18:45:03.919062 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 12 18:45:03.919322 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Dec 12 18:45:03.919504 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Dec 12 18:45:03.919627 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Dec 12 18:45:03.919743 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Dec 12 18:45:03.919861 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Dec 12 18:45:03.919976 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 12 18:45:03.920114 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Dec 12 18:45:03.920237 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Dec 12 18:45:03.920372 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Dec 12 18:45:03.920493 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Dec 12 18:45:03.920647 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Dec 12 18:45:03.920774 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Dec 12 18:45:03.920892 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Dec 12 18:45:03.921016 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Dec 12 18:45:03.921197 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Dec 12 18:45:03.921323 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Dec 12 18:45:03.921463 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Dec 12 18:45:03.921592 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Dec 12 18:45:03.921709 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Dec 12 18:45:03.921840 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Dec 12 18:45:03.921963 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 12 18:45:03.922093 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Dec 12 18:45:03.922218 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Dec 12 18:45:03.922397 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Dec 12 18:45:03.922585 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Dec 12 18:45:03.922705 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Dec 12 18:45:03.922716 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 12 18:45:03.922728 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 12 18:45:03.922737 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 12 18:45:03.922745 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 12 18:45:03.922753 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 12 18:45:03.922761 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 12 18:45:03.922769 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 12 18:45:03.922778 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 12 18:45:03.922786 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 12 18:45:03.922794 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 12 18:45:03.922804 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 12 18:45:03.922812 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 12 18:45:03.922820 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 12 18:45:03.922828 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 12 18:45:03.922836 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 12 18:45:03.922844 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 12 18:45:03.922852 kernel: iommu: Default domain type: Translated Dec 12 18:45:03.922860 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 12 18:45:03.922868 kernel: efivars: Registered efivars operations Dec 12 18:45:03.922879 kernel: PCI: Using ACPI for IRQ routing Dec 12 18:45:03.922890 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 12 18:45:03.922900 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Dec 12 18:45:03.922910 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Dec 12 18:45:03.922920 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Dec 12 18:45:03.922931 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Dec 12 18:45:03.922941 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Dec 12 18:45:03.922952 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Dec 12 18:45:03.922963 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Dec 12 18:45:03.922976 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Dec 12 18:45:03.923100 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 12 18:45:03.923216 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 12 18:45:03.923331 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 12 18:45:03.923341 kernel: vgaarb: loaded Dec 12 18:45:03.923383 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 12 18:45:03.923391 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 12 18:45:03.923399 kernel: clocksource: Switched to clocksource kvm-clock Dec 12 18:45:03.923411 kernel: VFS: Disk quotas dquot_6.6.0 Dec 12 18:45:03.923419 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 12 18:45:03.923428 kernel: pnp: PnP ACPI init Dec 12 18:45:03.923625 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Dec 12 18:45:03.923645 kernel: pnp: PnP ACPI: found 6 devices Dec 12 18:45:03.923653 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 12 18:45:03.923662 kernel: NET: Registered PF_INET protocol family Dec 12 18:45:03.923671 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Dec 12 18:45:03.923682 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Dec 12 18:45:03.923690 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 12 18:45:03.923699 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 12 18:45:03.923707 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Dec 12 18:45:03.923716 kernel: TCP: Hash tables configured (established 32768 bind 32768) Dec 12 18:45:03.923724 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 12 18:45:03.923733 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 12 18:45:03.923741 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 12 18:45:03.923750 kernel: NET: Registered PF_XDP protocol family Dec 12 18:45:03.923872 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Dec 12 18:45:03.923989 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Dec 12 18:45:03.924099 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 12 18:45:03.924206 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 12 18:45:03.924312 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 12 18:45:03.924441 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Dec 12 18:45:03.924562 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Dec 12 18:45:03.924674 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Dec 12 18:45:03.924685 kernel: PCI: CLS 0 bytes, default 64 Dec 12 18:45:03.924694 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848e100549, max_idle_ns: 440795215505 ns Dec 12 18:45:03.924705 kernel: Initialise system trusted keyrings Dec 12 18:45:03.924714 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Dec 12 18:45:03.924722 kernel: Key type asymmetric registered Dec 12 18:45:03.924732 kernel: Asymmetric key parser 'x509' registered Dec 12 18:45:03.924741 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Dec 12 18:45:03.924749 kernel: io scheduler mq-deadline registered Dec 12 18:45:03.924757 kernel: io scheduler kyber registered Dec 12 18:45:03.924766 kernel: io scheduler bfq registered Dec 12 18:45:03.924774 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 12 18:45:03.924783 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 12 18:45:03.924791 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 12 18:45:03.924799 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Dec 12 18:45:03.924810 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 12 18:45:03.924818 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 12 18:45:03.924827 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 12 18:45:03.924837 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 12 18:45:03.924845 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 12 18:45:03.924980 kernel: rtc_cmos 00:04: RTC can wake from S4 Dec 12 18:45:03.924993 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 12 18:45:03.925102 kernel: rtc_cmos 00:04: registered as rtc0 Dec 12 18:45:03.925216 kernel: rtc_cmos 00:04: setting system clock to 2025-12-12T18:45:03 UTC (1765565103) Dec 12 18:45:03.925325 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Dec 12 18:45:03.925335 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 12 18:45:03.925358 kernel: efifb: probing for efifb Dec 12 18:45:03.925367 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Dec 12 18:45:03.925376 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Dec 12 18:45:03.925384 kernel: efifb: scrolling: redraw Dec 12 18:45:03.925392 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Dec 12 18:45:03.925400 kernel: Console: switching to colour frame buffer device 160x50 Dec 12 18:45:03.925412 kernel: fb0: EFI VGA frame buffer device Dec 12 18:45:03.925420 kernel: pstore: Using crash dump compression: deflate Dec 12 18:45:03.925428 kernel: pstore: Registered efi_pstore as persistent store backend Dec 12 18:45:03.925437 kernel: NET: Registered PF_INET6 protocol family Dec 12 18:45:03.925445 kernel: Segment Routing with IPv6 Dec 12 18:45:03.925453 kernel: In-situ OAM (IOAM) with IPv6 Dec 12 18:45:03.925461 kernel: NET: Registered PF_PACKET protocol family Dec 12 18:45:03.925470 kernel: Key type dns_resolver registered Dec 12 18:45:03.925478 kernel: IPI shorthand broadcast: enabled Dec 12 18:45:03.925488 kernel: sched_clock: Marking stable (4354002985, 303641097)->(4726957003, -69312921) Dec 12 18:45:03.925505 kernel: registered taskstats version 1 Dec 12 18:45:03.925513 kernel: Loading compiled-in X.509 certificates Dec 12 18:45:03.925522 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.61-flatcar: 0d0c78e6590cb40d27f1cef749ef9f2f3425f38d' Dec 12 18:45:03.925530 kernel: Demotion targets for Node 0: null Dec 12 18:45:03.925538 kernel: Key type .fscrypt registered Dec 12 18:45:03.925547 kernel: Key type fscrypt-provisioning registered Dec 12 18:45:03.925555 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 12 18:45:03.925563 kernel: ima: Allocated hash algorithm: sha1 Dec 12 18:45:03.925574 kernel: ima: No architecture policies found Dec 12 18:45:03.925583 kernel: clk: Disabling unused clocks Dec 12 18:45:03.925593 kernel: Warning: unable to open an initial console. Dec 12 18:45:03.925602 kernel: Freeing unused kernel image (initmem) memory: 46188K Dec 12 18:45:03.925611 kernel: Write protecting the kernel read-only data: 40960k Dec 12 18:45:03.925621 kernel: Freeing unused kernel image (rodata/data gap) memory: 560K Dec 12 18:45:03.925630 kernel: Run /init as init process Dec 12 18:45:03.925640 kernel: with arguments: Dec 12 18:45:03.925649 kernel: /init Dec 12 18:45:03.925659 kernel: with environment: Dec 12 18:45:03.925667 kernel: HOME=/ Dec 12 18:45:03.925675 kernel: TERM=linux Dec 12 18:45:03.925685 systemd[1]: Successfully made /usr/ read-only. Dec 12 18:45:03.925697 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 12 18:45:03.925707 systemd[1]: Detected virtualization kvm. Dec 12 18:45:03.925716 systemd[1]: Detected architecture x86-64. Dec 12 18:45:03.925725 systemd[1]: Running in initrd. Dec 12 18:45:03.925735 systemd[1]: No hostname configured, using default hostname. Dec 12 18:45:03.925745 systemd[1]: Hostname set to . Dec 12 18:45:03.925753 systemd[1]: Initializing machine ID from VM UUID. Dec 12 18:45:03.925762 systemd[1]: Queued start job for default target initrd.target. Dec 12 18:45:03.925771 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:45:03.925780 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:45:03.925790 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 12 18:45:03.925799 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 12 18:45:03.925810 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 12 18:45:03.925820 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 12 18:45:03.925830 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 12 18:45:03.925839 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 12 18:45:03.925848 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:45:03.925857 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:45:03.925867 systemd[1]: Reached target paths.target - Path Units. Dec 12 18:45:03.925876 systemd[1]: Reached target slices.target - Slice Units. Dec 12 18:45:03.925885 systemd[1]: Reached target swap.target - Swaps. Dec 12 18:45:03.925894 systemd[1]: Reached target timers.target - Timer Units. Dec 12 18:45:03.925905 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 12 18:45:03.925914 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 12 18:45:03.925922 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 12 18:45:03.925931 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Dec 12 18:45:03.925940 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:45:03.925951 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 12 18:45:03.925960 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:45:03.925969 systemd[1]: Reached target sockets.target - Socket Units. Dec 12 18:45:03.925977 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 12 18:45:03.925986 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 12 18:45:03.925995 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 12 18:45:03.926004 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Dec 12 18:45:03.926013 systemd[1]: Starting systemd-fsck-usr.service... Dec 12 18:45:03.926024 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 12 18:45:03.926033 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 12 18:45:03.926042 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:03.926050 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 12 18:45:03.926060 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:45:03.926071 systemd[1]: Finished systemd-fsck-usr.service. Dec 12 18:45:03.926080 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 12 18:45:03.926118 systemd-journald[200]: Collecting audit messages is disabled. Dec 12 18:45:03.926139 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:03.926151 systemd-journald[200]: Journal started Dec 12 18:45:03.926170 systemd-journald[200]: Runtime Journal (/run/log/journal/6bf5c48166de46c5bf99d47cad4bb3db) is 6M, max 48.1M, 42.1M free. Dec 12 18:45:03.916684 systemd-modules-load[202]: Inserted module 'overlay' Dec 12 18:45:03.932586 systemd[1]: Started systemd-journald.service - Journal Service. Dec 12 18:45:03.934010 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 12 18:45:03.937861 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 12 18:45:03.943448 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 12 18:45:03.952368 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 12 18:45:03.955172 systemd-modules-load[202]: Inserted module 'br_netfilter' Dec 12 18:45:03.956638 kernel: Bridge firewalling registered Dec 12 18:45:03.962198 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 12 18:45:03.963111 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 12 18:45:03.965262 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:45:03.980099 systemd-tmpfiles[223]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Dec 12 18:45:03.981885 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:45:03.986461 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:45:03.989023 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 12 18:45:03.990673 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 12 18:45:03.994319 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:45:03.999259 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 12 18:45:04.020039 dracut-cmdline[240]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a214a2d85e162c493e8b13db2df50a43e1005a0e4854a1ae089a14f442a30022 Dec 12 18:45:04.041957 systemd-resolved[243]: Positive Trust Anchors: Dec 12 18:45:04.041974 systemd-resolved[243]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 12 18:45:04.042004 systemd-resolved[243]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 12 18:45:04.044982 systemd-resolved[243]: Defaulting to hostname 'linux'. Dec 12 18:45:04.046230 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 12 18:45:04.047138 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:45:04.160957 kernel: SCSI subsystem initialized Dec 12 18:45:04.172383 kernel: Loading iSCSI transport class v2.0-870. Dec 12 18:45:04.185394 kernel: iscsi: registered transport (tcp) Dec 12 18:45:04.212391 kernel: iscsi: registered transport (qla4xxx) Dec 12 18:45:04.212479 kernel: QLogic iSCSI HBA Driver Dec 12 18:45:04.235522 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 12 18:45:04.263220 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:45:04.267005 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 12 18:45:04.328849 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 12 18:45:04.332564 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 12 18:45:04.392385 kernel: raid6: avx2x4 gen() 28517 MB/s Dec 12 18:45:04.409381 kernel: raid6: avx2x2 gen() 28756 MB/s Dec 12 18:45:04.427188 kernel: raid6: avx2x1 gen() 22823 MB/s Dec 12 18:45:04.427224 kernel: raid6: using algorithm avx2x2 gen() 28756 MB/s Dec 12 18:45:04.445222 kernel: raid6: .... xor() 19219 MB/s, rmw enabled Dec 12 18:45:04.445252 kernel: raid6: using avx2x2 recovery algorithm Dec 12 18:45:04.467376 kernel: xor: automatically using best checksumming function avx Dec 12 18:45:04.642388 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 12 18:45:04.651127 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 12 18:45:04.654564 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:45:04.685098 systemd-udevd[453]: Using default interface naming scheme 'v255'. Dec 12 18:45:04.691724 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:45:04.695169 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 12 18:45:04.729731 dracut-pre-trigger[460]: rd.md=0: removing MD RAID activation Dec 12 18:45:04.759209 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 12 18:45:04.764163 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 12 18:45:04.841580 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:45:04.847768 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 12 18:45:04.888406 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Dec 12 18:45:04.894378 kernel: cryptd: max_cpu_qlen set to 1000 Dec 12 18:45:04.898634 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Dec 12 18:45:04.915479 kernel: AES CTR mode by8 optimization enabled Dec 12 18:45:04.915527 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Dec 12 18:45:04.923961 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 12 18:45:04.923994 kernel: GPT:9289727 != 19775487 Dec 12 18:45:04.924009 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 12 18:45:04.924022 kernel: GPT:9289727 != 19775487 Dec 12 18:45:04.924035 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 12 18:45:04.924047 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 12 18:45:04.948377 kernel: libata version 3.00 loaded. Dec 12 18:45:04.958452 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:45:04.958749 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:04.963838 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:04.975515 kernel: ahci 0000:00:1f.2: version 3.0 Dec 12 18:45:04.975923 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 12 18:45:04.976001 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:04.981999 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:45:04.989864 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Dec 12 18:45:04.990095 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Dec 12 18:45:04.990238 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 12 18:45:04.993385 kernel: scsi host0: ahci Dec 12 18:45:04.995377 kernel: scsi host1: ahci Dec 12 18:45:04.998372 kernel: scsi host2: ahci Dec 12 18:45:05.001394 kernel: scsi host3: ahci Dec 12 18:45:05.005492 kernel: scsi host4: ahci Dec 12 18:45:05.005675 kernel: scsi host5: ahci Dec 12 18:45:05.008028 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 1 Dec 12 18:45:05.008051 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 1 Dec 12 18:45:05.010390 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 1 Dec 12 18:45:05.012924 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 1 Dec 12 18:45:05.012958 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 1 Dec 12 18:45:05.012969 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 1 Dec 12 18:45:05.013131 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Dec 12 18:45:05.033937 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Dec 12 18:45:05.041849 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Dec 12 18:45:05.042895 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Dec 12 18:45:05.057285 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 12 18:45:05.058866 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 12 18:45:05.062226 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:45:05.062283 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:05.069186 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:05.090028 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:05.095021 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:45:05.100574 disk-uuid[616]: Primary Header is updated. Dec 12 18:45:05.100574 disk-uuid[616]: Secondary Entries is updated. Dec 12 18:45:05.100574 disk-uuid[616]: Secondary Header is updated. Dec 12 18:45:05.106169 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 12 18:45:05.108380 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 12 18:45:05.119138 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:05.325383 kernel: ata1: SATA link down (SStatus 0 SControl 300) Dec 12 18:45:05.327370 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 12 18:45:05.327399 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 12 18:45:05.328368 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 12 18:45:05.330391 kernel: ata3.00: LPM support broken, forcing max_power Dec 12 18:45:05.331844 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 12 18:45:05.331863 kernel: ata3.00: applying bridge limits Dec 12 18:45:05.333380 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 12 18:45:05.334378 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 12 18:45:05.335382 kernel: ata3.00: LPM support broken, forcing max_power Dec 12 18:45:05.337397 kernel: ata3.00: configured for UDMA/100 Dec 12 18:45:05.339407 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 12 18:45:05.395893 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 12 18:45:05.396144 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 12 18:45:05.409381 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Dec 12 18:45:05.771896 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 12 18:45:05.776530 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 12 18:45:05.780904 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:45:05.785298 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 12 18:45:05.790146 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 12 18:45:05.819267 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 12 18:45:06.113372 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 12 18:45:06.114411 disk-uuid[618]: The operation has completed successfully. Dec 12 18:45:06.142401 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 12 18:45:06.142559 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 12 18:45:06.187806 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 12 18:45:06.204538 sh[651]: Success Dec 12 18:45:06.225391 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 12 18:45:06.225429 kernel: device-mapper: uevent: version 1.0.3 Dec 12 18:45:06.227092 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Dec 12 18:45:06.237379 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Dec 12 18:45:06.267287 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 12 18:45:06.270074 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 12 18:45:06.284675 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 12 18:45:06.293745 kernel: BTRFS: device fsid a6ae7f96-a076-4d3c-81ed-46dd341492f8 devid 1 transid 37 /dev/mapper/usr (253:0) scanned by mount (663) Dec 12 18:45:06.293809 kernel: BTRFS info (device dm-0): first mount of filesystem a6ae7f96-a076-4d3c-81ed-46dd341492f8 Dec 12 18:45:06.293821 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:45:06.302720 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 12 18:45:06.302777 kernel: BTRFS info (device dm-0): enabling free space tree Dec 12 18:45:06.304212 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 12 18:45:06.305572 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Dec 12 18:45:06.307881 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 12 18:45:06.308969 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 12 18:45:06.313019 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 12 18:45:06.344091 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (695) Dec 12 18:45:06.344149 kernel: BTRFS info (device vda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:45:06.344167 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:45:06.354961 kernel: BTRFS info (device vda6): turning on async discard Dec 12 18:45:06.354996 kernel: BTRFS info (device vda6): enabling free space tree Dec 12 18:45:06.361369 kernel: BTRFS info (device vda6): last unmount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:45:06.362398 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 12 18:45:06.367914 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 12 18:45:06.465206 ignition[753]: Ignition 2.22.0 Dec 12 18:45:06.465219 ignition[753]: Stage: fetch-offline Dec 12 18:45:06.465246 ignition[753]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:06.465255 ignition[753]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:06.469894 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 12 18:45:06.465327 ignition[753]: parsed url from cmdline: "" Dec 12 18:45:06.475388 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 12 18:45:06.465330 ignition[753]: no config URL provided Dec 12 18:45:06.465335 ignition[753]: reading system config file "/usr/lib/ignition/user.ign" Dec 12 18:45:06.465343 ignition[753]: no config at "/usr/lib/ignition/user.ign" Dec 12 18:45:06.465380 ignition[753]: op(1): [started] loading QEMU firmware config module Dec 12 18:45:06.465385 ignition[753]: op(1): executing: "modprobe" "qemu_fw_cfg" Dec 12 18:45:06.479414 ignition[753]: op(1): [finished] loading QEMU firmware config module Dec 12 18:45:06.542816 systemd-networkd[840]: lo: Link UP Dec 12 18:45:06.542827 systemd-networkd[840]: lo: Gained carrier Dec 12 18:45:06.545086 systemd-networkd[840]: Enumeration completed Dec 12 18:45:06.545463 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 12 18:45:06.547122 systemd[1]: Reached target network.target - Network. Dec 12 18:45:06.547317 systemd-networkd[840]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:45:06.547322 systemd-networkd[840]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:45:06.548690 systemd-networkd[840]: eth0: Link UP Dec 12 18:45:06.549501 systemd-networkd[840]: eth0: Gained carrier Dec 12 18:45:06.549519 systemd-networkd[840]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:45:06.581412 systemd-networkd[840]: eth0: DHCPv4 address 10.0.0.119/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 12 18:45:06.606046 ignition[753]: parsing config with SHA512: 01b728c5e5da4185a89eff5add996234556b93f4687807b54b1ecfce3ac9ed29d4c47395e45f3b64f585aabe33d34572d0ba8638bd17f7c3d9d32f01336c99fc Dec 12 18:45:06.612705 unknown[753]: fetched base config from "system" Dec 12 18:45:06.612722 unknown[753]: fetched user config from "qemu" Dec 12 18:45:06.613230 ignition[753]: fetch-offline: fetch-offline passed Dec 12 18:45:06.613314 ignition[753]: Ignition finished successfully Dec 12 18:45:06.619847 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 12 18:45:06.621069 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Dec 12 18:45:06.623887 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 12 18:45:06.656777 ignition[845]: Ignition 2.22.0 Dec 12 18:45:06.656794 ignition[845]: Stage: kargs Dec 12 18:45:06.656970 ignition[845]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:06.656984 ignition[845]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:06.658109 ignition[845]: kargs: kargs passed Dec 12 18:45:06.658153 ignition[845]: Ignition finished successfully Dec 12 18:45:06.665727 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 12 18:45:06.670558 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 12 18:45:06.705904 ignition[853]: Ignition 2.22.0 Dec 12 18:45:06.705922 ignition[853]: Stage: disks Dec 12 18:45:06.706106 ignition[853]: no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:06.706118 ignition[853]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:06.707133 ignition[853]: disks: disks passed Dec 12 18:45:06.707189 ignition[853]: Ignition finished successfully Dec 12 18:45:06.713916 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 12 18:45:06.715619 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 12 18:45:06.718995 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 12 18:45:06.722391 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 12 18:45:06.726270 systemd[1]: Reached target sysinit.target - System Initialization. Dec 12 18:45:06.729905 systemd[1]: Reached target basic.target - Basic System. Dec 12 18:45:06.735466 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 12 18:45:06.762995 systemd-fsck[864]: ROOT: clean, 15/553520 files, 52789/553472 blocks Dec 12 18:45:06.773684 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 12 18:45:06.778018 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 12 18:45:06.903391 kernel: EXT4-fs (vda9): mounted filesystem e48ca59c-1206-4abd-b121-5e9b35e49852 r/w with ordered data mode. Quota mode: none. Dec 12 18:45:06.903922 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 12 18:45:06.905302 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 12 18:45:06.910411 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 12 18:45:06.911908 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 12 18:45:06.917052 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Dec 12 18:45:06.917110 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 12 18:45:06.917139 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 12 18:45:06.941700 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 12 18:45:06.950108 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (872) Dec 12 18:45:06.950147 kernel: BTRFS info (device vda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:45:06.950159 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:45:06.944732 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 12 18:45:06.956385 kernel: BTRFS info (device vda6): turning on async discard Dec 12 18:45:06.956410 kernel: BTRFS info (device vda6): enabling free space tree Dec 12 18:45:06.958376 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 12 18:45:06.990165 initrd-setup-root[896]: cut: /sysroot/etc/passwd: No such file or directory Dec 12 18:45:06.995212 initrd-setup-root[903]: cut: /sysroot/etc/group: No such file or directory Dec 12 18:45:06.999570 initrd-setup-root[910]: cut: /sysroot/etc/shadow: No such file or directory Dec 12 18:45:07.005113 initrd-setup-root[917]: cut: /sysroot/etc/gshadow: No such file or directory Dec 12 18:45:07.113190 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 12 18:45:07.116303 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 12 18:45:07.119601 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 12 18:45:07.137394 kernel: BTRFS info (device vda6): last unmount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:45:07.151258 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 12 18:45:07.169788 ignition[986]: INFO : Ignition 2.22.0 Dec 12 18:45:07.169788 ignition[986]: INFO : Stage: mount Dec 12 18:45:07.172329 ignition[986]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:07.172329 ignition[986]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:07.172329 ignition[986]: INFO : mount: mount passed Dec 12 18:45:07.172329 ignition[986]: INFO : Ignition finished successfully Dec 12 18:45:07.173598 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 12 18:45:07.176871 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 12 18:45:07.297730 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 12 18:45:07.299316 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 12 18:45:07.332852 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (998) Dec 12 18:45:07.332896 kernel: BTRFS info (device vda6): first mount of filesystem 7e9ead35-f0ec-40e8-bc31-5061934f865a Dec 12 18:45:07.332908 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 12 18:45:07.339980 kernel: BTRFS info (device vda6): turning on async discard Dec 12 18:45:07.340063 kernel: BTRFS info (device vda6): enabling free space tree Dec 12 18:45:07.342953 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 12 18:45:07.379575 ignition[1015]: INFO : Ignition 2.22.0 Dec 12 18:45:07.379575 ignition[1015]: INFO : Stage: files Dec 12 18:45:07.382473 ignition[1015]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:07.382473 ignition[1015]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:07.382473 ignition[1015]: DEBUG : files: compiled without relabeling support, skipping Dec 12 18:45:07.382473 ignition[1015]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 12 18:45:07.382473 ignition[1015]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 12 18:45:07.393076 ignition[1015]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 12 18:45:07.393076 ignition[1015]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 12 18:45:07.393076 ignition[1015]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 12 18:45:07.393076 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Dec 12 18:45:07.393076 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Dec 12 18:45:07.384607 unknown[1015]: wrote ssh authorized keys file for user: core Dec 12 18:45:07.445653 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 12 18:45:07.501243 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Dec 12 18:45:07.501243 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 12 18:45:07.507739 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Dec 12 18:45:07.742147 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Dec 12 18:45:07.830598 systemd-networkd[840]: eth0: Gained IPv6LL Dec 12 18:45:07.939936 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Dec 12 18:45:07.939936 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Dec 12 18:45:07.945806 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Dec 12 18:45:07.945806 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 12 18:45:07.951297 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 12 18:45:07.954081 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 12 18:45:07.957039 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 12 18:45:07.959981 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 12 18:45:07.962959 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 12 18:45:07.969570 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 12 18:45:07.972511 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 12 18:45:07.972511 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Dec 12 18:45:07.980100 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Dec 12 18:45:07.980100 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Dec 12 18:45:07.987770 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.1-x86-64.raw: attempt #1 Dec 12 18:45:08.200431 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Dec 12 18:45:08.568987 ignition[1015]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Dec 12 18:45:08.568987 ignition[1015]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Dec 12 18:45:08.575411 ignition[1015]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 12 18:45:08.582259 ignition[1015]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 12 18:45:08.582259 ignition[1015]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Dec 12 18:45:08.582259 ignition[1015]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Dec 12 18:45:08.592626 ignition[1015]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 12 18:45:08.592626 ignition[1015]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 12 18:45:08.592626 ignition[1015]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Dec 12 18:45:08.592626 ignition[1015]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Dec 12 18:45:08.615652 ignition[1015]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Dec 12 18:45:08.621820 ignition[1015]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 12 18:45:08.625392 ignition[1015]: INFO : files: files passed Dec 12 18:45:08.625392 ignition[1015]: INFO : Ignition finished successfully Dec 12 18:45:08.636331 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 12 18:45:08.645279 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 12 18:45:08.653761 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 12 18:45:08.675835 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 12 18:45:08.676005 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 12 18:45:08.677452 initrd-setup-root-after-ignition[1044]: grep: /sysroot/oem/oem-release: No such file or directory Dec 12 18:45:08.687315 initrd-setup-root-after-ignition[1046]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:45:08.687315 initrd-setup-root-after-ignition[1046]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:45:08.693085 initrd-setup-root-after-ignition[1050]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 12 18:45:08.695292 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 12 18:45:08.698743 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 12 18:45:08.700301 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 12 18:45:08.747886 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 12 18:45:08.749480 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 12 18:45:08.753634 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 12 18:45:08.756784 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 12 18:45:08.757820 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 12 18:45:08.760204 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 12 18:45:08.791336 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 12 18:45:08.793665 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 12 18:45:08.823520 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:45:08.824275 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:45:08.828956 systemd[1]: Stopped target timers.target - Timer Units. Dec 12 18:45:08.832114 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 12 18:45:08.832240 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 12 18:45:08.836771 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 12 18:45:08.837995 systemd[1]: Stopped target basic.target - Basic System. Dec 12 18:45:08.842368 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 12 18:45:08.845297 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 12 18:45:08.848889 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 12 18:45:08.849775 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Dec 12 18:45:08.856181 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 12 18:45:08.859915 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 12 18:45:08.863219 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 12 18:45:08.867211 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 12 18:45:08.870800 systemd[1]: Stopped target swap.target - Swaps. Dec 12 18:45:08.873998 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 12 18:45:08.874165 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 12 18:45:08.878928 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:45:08.882192 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:45:08.883069 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 12 18:45:08.883223 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:45:08.883957 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 12 18:45:08.884101 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 12 18:45:08.894439 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 12 18:45:08.894556 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 12 18:45:08.897780 systemd[1]: Stopped target paths.target - Path Units. Dec 12 18:45:08.898834 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 12 18:45:08.902440 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:45:08.905754 systemd[1]: Stopped target slices.target - Slice Units. Dec 12 18:45:08.906884 systemd[1]: Stopped target sockets.target - Socket Units. Dec 12 18:45:08.910851 systemd[1]: iscsid.socket: Deactivated successfully. Dec 12 18:45:08.910950 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 12 18:45:08.914333 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 12 18:45:08.914455 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 12 18:45:08.917314 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 12 18:45:08.917537 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 12 18:45:08.920306 systemd[1]: ignition-files.service: Deactivated successfully. Dec 12 18:45:08.920494 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 12 18:45:08.924612 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 12 18:45:08.926336 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 12 18:45:08.926511 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:45:08.927669 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 12 18:45:08.934999 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 12 18:45:08.935282 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:45:08.938868 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 12 18:45:08.939054 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 12 18:45:08.949960 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 12 18:45:08.953527 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 12 18:45:08.971738 ignition[1070]: INFO : Ignition 2.22.0 Dec 12 18:45:08.971738 ignition[1070]: INFO : Stage: umount Dec 12 18:45:08.974563 ignition[1070]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 12 18:45:08.974563 ignition[1070]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 12 18:45:08.974563 ignition[1070]: INFO : umount: umount passed Dec 12 18:45:08.974563 ignition[1070]: INFO : Ignition finished successfully Dec 12 18:45:08.981007 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 12 18:45:08.981184 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 12 18:45:08.983108 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 12 18:45:08.985504 systemd[1]: Stopped target network.target - Network. Dec 12 18:45:08.986435 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 12 18:45:08.986515 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 12 18:45:08.986982 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 12 18:45:08.987029 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 12 18:45:08.996563 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 12 18:45:08.999932 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 12 18:45:09.004870 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 12 18:45:09.004933 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 12 18:45:09.010302 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 12 18:45:09.012308 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 12 18:45:09.025846 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 12 18:45:09.026036 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 12 18:45:09.032326 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Dec 12 18:45:09.032690 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 12 18:45:09.032854 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 12 18:45:09.038185 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Dec 12 18:45:09.038846 systemd[1]: Stopped target network-pre.target - Preparation for Network. Dec 12 18:45:09.040902 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 12 18:45:09.040944 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:45:09.044687 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 12 18:45:09.048421 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 12 18:45:09.048480 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 12 18:45:09.050549 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 12 18:45:09.052058 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:45:09.055767 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 12 18:45:09.055817 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 12 18:45:09.056939 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 12 18:45:09.056983 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:45:09.064521 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:45:09.066983 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 12 18:45:09.067064 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:45:09.089837 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 12 18:45:09.089986 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 12 18:45:09.105035 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 12 18:45:09.105220 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:45:09.106258 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 12 18:45:09.106303 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 12 18:45:09.111103 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 12 18:45:09.111138 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:45:09.114497 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 12 18:45:09.114548 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 12 18:45:09.120166 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 12 18:45:09.120220 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 12 18:45:09.121888 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 12 18:45:09.121944 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 12 18:45:09.131214 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 12 18:45:09.132039 systemd[1]: systemd-network-generator.service: Deactivated successfully. Dec 12 18:45:09.132094 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:45:09.139823 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 12 18:45:09.139881 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:45:09.145523 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 12 18:45:09.145597 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:09.152298 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Dec 12 18:45:09.152387 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Dec 12 18:45:09.152447 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Dec 12 18:45:09.163342 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 12 18:45:09.163517 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 12 18:45:09.373856 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 12 18:45:09.374028 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 12 18:45:09.375948 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 12 18:45:09.378885 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 12 18:45:09.378975 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 12 18:45:09.384826 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 12 18:45:09.410126 systemd[1]: Switching root. Dec 12 18:45:09.451988 systemd-journald[200]: Journal stopped Dec 12 18:45:10.939640 systemd-journald[200]: Received SIGTERM from PID 1 (systemd). Dec 12 18:45:10.939711 kernel: SELinux: policy capability network_peer_controls=1 Dec 12 18:45:10.939724 kernel: SELinux: policy capability open_perms=1 Dec 12 18:45:10.939736 kernel: SELinux: policy capability extended_socket_class=1 Dec 12 18:45:10.939751 kernel: SELinux: policy capability always_check_network=0 Dec 12 18:45:10.939767 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 12 18:45:10.939782 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 12 18:45:10.939795 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 12 18:45:10.939809 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 12 18:45:10.939828 kernel: SELinux: policy capability userspace_initial_context=0 Dec 12 18:45:10.939842 kernel: audit: type=1403 audit(1765565110.068:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 12 18:45:10.939854 systemd[1]: Successfully loaded SELinux policy in 70.256ms. Dec 12 18:45:10.939876 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 8.488ms. Dec 12 18:45:10.939889 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 12 18:45:10.939911 systemd[1]: Detected virtualization kvm. Dec 12 18:45:10.939927 systemd[1]: Detected architecture x86-64. Dec 12 18:45:10.939939 systemd[1]: Detected first boot. Dec 12 18:45:10.939951 systemd[1]: Initializing machine ID from VM UUID. Dec 12 18:45:10.939963 zram_generator::config[1116]: No configuration found. Dec 12 18:45:10.939976 kernel: Guest personality initialized and is inactive Dec 12 18:45:10.939992 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Dec 12 18:45:10.940003 kernel: Initialized host personality Dec 12 18:45:10.940014 kernel: NET: Registered PF_VSOCK protocol family Dec 12 18:45:10.940028 systemd[1]: Populated /etc with preset unit settings. Dec 12 18:45:10.940046 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Dec 12 18:45:10.940058 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 12 18:45:10.940070 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 12 18:45:10.940082 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 12 18:45:10.940103 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 12 18:45:10.940124 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 12 18:45:10.940147 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 12 18:45:10.940170 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 12 18:45:10.940191 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 12 18:45:10.940211 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 12 18:45:10.940238 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 12 18:45:10.940260 systemd[1]: Created slice user.slice - User and Session Slice. Dec 12 18:45:10.940281 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 12 18:45:10.940301 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 12 18:45:10.940319 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 12 18:45:10.940366 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 12 18:45:10.940383 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 12 18:45:10.940395 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 12 18:45:10.940408 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 12 18:45:10.940419 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 12 18:45:10.940431 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 12 18:45:10.940455 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 12 18:45:10.940478 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 12 18:45:10.940505 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 12 18:45:10.940525 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 12 18:45:10.940546 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 12 18:45:10.940566 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 12 18:45:10.940586 systemd[1]: Reached target slices.target - Slice Units. Dec 12 18:45:10.940608 systemd[1]: Reached target swap.target - Swaps. Dec 12 18:45:10.940627 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 12 18:45:10.940639 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 12 18:45:10.940651 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Dec 12 18:45:10.940663 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 12 18:45:10.940678 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 12 18:45:10.940690 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 12 18:45:10.940702 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 12 18:45:10.940714 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 12 18:45:10.940726 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 12 18:45:10.940738 systemd[1]: Mounting media.mount - External Media Directory... Dec 12 18:45:10.940750 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:10.940762 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 12 18:45:10.940774 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 12 18:45:10.940788 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 12 18:45:10.940801 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 12 18:45:10.940813 systemd[1]: Reached target machines.target - Containers. Dec 12 18:45:10.940825 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 12 18:45:10.940837 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:45:10.940849 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 12 18:45:10.940860 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 12 18:45:10.940873 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 12 18:45:10.940887 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 12 18:45:10.940901 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 12 18:45:10.940915 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 12 18:45:10.940927 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 12 18:45:10.940940 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 12 18:45:10.940953 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 12 18:45:10.940965 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 12 18:45:10.940976 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 12 18:45:10.940994 systemd[1]: Stopped systemd-fsck-usr.service. Dec 12 18:45:10.941010 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:45:10.941024 kernel: fuse: init (API version 7.41) Dec 12 18:45:10.941035 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 12 18:45:10.941047 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 12 18:45:10.941058 kernel: loop: module loaded Dec 12 18:45:10.941070 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 12 18:45:10.941082 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 12 18:45:10.941094 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Dec 12 18:45:10.941109 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 12 18:45:10.941124 systemd[1]: verity-setup.service: Deactivated successfully. Dec 12 18:45:10.941138 systemd[1]: Stopped verity-setup.service. Dec 12 18:45:10.941176 systemd-journald[1195]: Collecting audit messages is disabled. Dec 12 18:45:10.941197 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:10.941210 systemd-journald[1195]: Journal started Dec 12 18:45:10.941279 systemd-journald[1195]: Runtime Journal (/run/log/journal/6bf5c48166de46c5bf99d47cad4bb3db) is 6M, max 48.1M, 42.1M free. Dec 12 18:45:10.620486 systemd[1]: Queued start job for default target multi-user.target. Dec 12 18:45:10.635638 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Dec 12 18:45:10.636184 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 12 18:45:10.947376 systemd[1]: Started systemd-journald.service - Journal Service. Dec 12 18:45:10.947413 kernel: ACPI: bus type drm_connector registered Dec 12 18:45:10.952093 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 12 18:45:10.954255 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 12 18:45:10.956336 systemd[1]: Mounted media.mount - External Media Directory. Dec 12 18:45:10.958195 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 12 18:45:10.960196 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 12 18:45:10.962214 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 12 18:45:10.964295 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 12 18:45:10.966761 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 12 18:45:10.969241 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 12 18:45:10.969643 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 12 18:45:10.972138 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 12 18:45:10.972613 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 12 18:45:10.974780 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 12 18:45:10.975004 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 12 18:45:10.977016 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 12 18:45:10.977554 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 12 18:45:10.979767 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 12 18:45:10.980009 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 12 18:45:10.982039 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 12 18:45:10.982250 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 12 18:45:10.984302 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 12 18:45:10.986639 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 12 18:45:10.988984 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 12 18:45:10.991380 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Dec 12 18:45:11.011771 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 12 18:45:11.016153 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 12 18:45:11.019846 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 12 18:45:11.021962 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 12 18:45:11.022096 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 12 18:45:11.024362 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Dec 12 18:45:11.032503 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 12 18:45:11.034770 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:45:11.036541 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 12 18:45:11.040418 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 12 18:45:11.042817 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 12 18:45:11.043915 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 12 18:45:11.046053 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 12 18:45:11.049022 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:45:11.052479 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 12 18:45:11.059272 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 12 18:45:11.067543 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 12 18:45:11.070427 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 12 18:45:11.074251 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 12 18:45:11.075209 systemd-journald[1195]: Time spent on flushing to /var/log/journal/6bf5c48166de46c5bf99d47cad4bb3db is 42.188ms for 1079 entries. Dec 12 18:45:11.075209 systemd-journald[1195]: System Journal (/var/log/journal/6bf5c48166de46c5bf99d47cad4bb3db) is 8M, max 195.6M, 187.6M free. Dec 12 18:45:11.138697 systemd-journald[1195]: Received client request to flush runtime journal. Dec 12 18:45:11.138784 kernel: loop0: detected capacity change from 0 to 219144 Dec 12 18:45:11.138998 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 12 18:45:11.087543 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 12 18:45:11.092314 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 12 18:45:11.099040 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Dec 12 18:45:11.112645 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:45:11.135303 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 12 18:45:11.141524 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 12 18:45:11.144749 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 12 18:45:11.149386 kernel: loop1: detected capacity change from 0 to 128560 Dec 12 18:45:11.154731 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Dec 12 18:45:11.177376 kernel: loop2: detected capacity change from 0 to 110984 Dec 12 18:45:11.180706 systemd-tmpfiles[1250]: ACLs are not supported, ignoring. Dec 12 18:45:11.180724 systemd-tmpfiles[1250]: ACLs are not supported, ignoring. Dec 12 18:45:11.188381 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 12 18:45:11.206376 kernel: loop3: detected capacity change from 0 to 219144 Dec 12 18:45:11.218396 kernel: loop4: detected capacity change from 0 to 128560 Dec 12 18:45:11.235382 kernel: loop5: detected capacity change from 0 to 110984 Dec 12 18:45:11.242868 (sd-merge)[1257]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Dec 12 18:45:11.243607 (sd-merge)[1257]: Merged extensions into '/usr'. Dec 12 18:45:11.249432 systemd[1]: Reload requested from client PID 1235 ('systemd-sysext') (unit systemd-sysext.service)... Dec 12 18:45:11.249455 systemd[1]: Reloading... Dec 12 18:45:11.308416 zram_generator::config[1280]: No configuration found. Dec 12 18:45:11.458156 ldconfig[1230]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 12 18:45:11.519133 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 12 18:45:11.519214 systemd[1]: Reloading finished in 269 ms. Dec 12 18:45:11.550076 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 12 18:45:11.552614 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 12 18:45:11.574654 systemd[1]: Starting ensure-sysext.service... Dec 12 18:45:11.577020 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 12 18:45:11.595978 systemd[1]: Reload requested from client PID 1320 ('systemctl') (unit ensure-sysext.service)... Dec 12 18:45:11.595996 systemd[1]: Reloading... Dec 12 18:45:11.597230 systemd-tmpfiles[1321]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Dec 12 18:45:11.597435 systemd-tmpfiles[1321]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Dec 12 18:45:11.597823 systemd-tmpfiles[1321]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 12 18:45:11.598118 systemd-tmpfiles[1321]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 12 18:45:11.599080 systemd-tmpfiles[1321]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 12 18:45:11.599375 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. Dec 12 18:45:11.599453 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. Dec 12 18:45:11.603997 systemd-tmpfiles[1321]: Detected autofs mount point /boot during canonicalization of boot. Dec 12 18:45:11.604101 systemd-tmpfiles[1321]: Skipping /boot Dec 12 18:45:11.614540 systemd-tmpfiles[1321]: Detected autofs mount point /boot during canonicalization of boot. Dec 12 18:45:11.614552 systemd-tmpfiles[1321]: Skipping /boot Dec 12 18:45:11.648390 zram_generator::config[1346]: No configuration found. Dec 12 18:45:11.836943 systemd[1]: Reloading finished in 240 ms. Dec 12 18:45:11.866065 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 12 18:45:11.896249 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 12 18:45:11.906414 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 12 18:45:11.909542 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 12 18:45:11.928934 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 12 18:45:11.934561 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 12 18:45:11.938310 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 12 18:45:11.943722 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 12 18:45:11.951279 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:11.951633 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:45:11.952982 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 12 18:45:11.959419 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 12 18:45:11.963192 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 12 18:45:11.965114 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:45:11.965311 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:45:11.967933 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 12 18:45:11.969676 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:11.976044 systemd-udevd[1392]: Using default interface naming scheme 'v255'. Dec 12 18:45:11.978178 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 12 18:45:11.978780 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 12 18:45:11.981417 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 12 18:45:11.981783 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 12 18:45:11.984564 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 12 18:45:11.988702 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 12 18:45:11.991602 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 12 18:45:11.996049 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 12 18:45:12.003292 systemd[1]: Finished ensure-sysext.service. Dec 12 18:45:12.007018 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:12.007233 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 12 18:45:12.008542 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 12 18:45:12.010283 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 12 18:45:12.010327 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 12 18:45:12.010396 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 12 18:45:12.010456 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 12 18:45:12.012217 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 12 18:45:12.014553 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 12 18:45:12.061553 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 12 18:45:12.070126 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 12 18:45:12.070382 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 12 18:45:12.093065 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 12 18:45:12.113040 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 12 18:45:12.117626 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 12 18:45:12.127938 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 12 18:45:12.144542 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 12 18:45:12.157008 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 12 18:45:12.170826 augenrules[1461]: No rules Dec 12 18:45:12.173701 systemd[1]: audit-rules.service: Deactivated successfully. Dec 12 18:45:12.174478 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 12 18:45:12.193022 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 12 18:45:12.283673 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 12 18:45:12.287271 systemd-networkd[1432]: lo: Link UP Dec 12 18:45:12.287282 systemd-networkd[1432]: lo: Gained carrier Dec 12 18:45:12.288437 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 12 18:45:12.290374 kernel: mousedev: PS/2 mouse device common for all mice Dec 12 18:45:12.290516 systemd-networkd[1432]: Enumeration completed Dec 12 18:45:12.290900 systemd-networkd[1432]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:45:12.290910 systemd-networkd[1432]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 12 18:45:12.291594 systemd-networkd[1432]: eth0: Link UP Dec 12 18:45:12.291739 systemd-networkd[1432]: eth0: Gained carrier Dec 12 18:45:12.291747 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 12 18:45:12.291758 systemd-networkd[1432]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 12 18:45:12.295375 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Dec 12 18:45:12.298830 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Dec 12 18:45:12.302317 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 12 18:45:12.305415 systemd-networkd[1432]: eth0: DHCPv4 address 10.0.0.119/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 12 18:45:12.311387 kernel: ACPI: button: Power Button [PWRF] Dec 12 18:45:12.314639 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 12 18:45:12.326772 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Dec 12 18:45:12.327030 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 12 18:45:12.328905 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 12 18:45:12.329989 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Dec 12 18:45:12.380113 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 12 18:45:12.382321 systemd[1]: Reached target time-set.target - System Time Set. Dec 12 18:45:12.383909 systemd-timesyncd[1420]: Contacted time server 10.0.0.1:123 (10.0.0.1). Dec 12 18:45:12.383971 systemd-timesyncd[1420]: Initial clock synchronization to Fri 2025-12-12 18:45:12.676761 UTC. Dec 12 18:45:12.412779 systemd-resolved[1390]: Positive Trust Anchors: Dec 12 18:45:12.412796 systemd-resolved[1390]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 12 18:45:12.412827 systemd-resolved[1390]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 12 18:45:12.417755 systemd-resolved[1390]: Defaulting to hostname 'linux'. Dec 12 18:45:12.423468 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 12 18:45:12.426203 systemd[1]: Reached target network.target - Network. Dec 12 18:45:12.427918 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 12 18:45:12.430952 systemd[1]: Reached target sysinit.target - System Initialization. Dec 12 18:45:12.432959 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 12 18:45:12.434975 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 12 18:45:12.438451 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Dec 12 18:45:12.441673 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 12 18:45:12.443548 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 12 18:45:12.445542 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 12 18:45:12.448422 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 12 18:45:12.448463 systemd[1]: Reached target paths.target - Path Units. Dec 12 18:45:12.450034 systemd[1]: Reached target timers.target - Timer Units. Dec 12 18:45:12.453668 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 12 18:45:12.457574 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 12 18:45:12.461978 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Dec 12 18:45:12.464253 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Dec 12 18:45:12.466448 systemd[1]: Reached target ssh-access.target - SSH Access Available. Dec 12 18:45:12.513129 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 12 18:45:12.516073 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Dec 12 18:45:12.519094 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 12 18:45:12.525041 kernel: kvm_amd: TSC scaling supported Dec 12 18:45:12.525101 kernel: kvm_amd: Nested Virtualization enabled Dec 12 18:45:12.525119 kernel: kvm_amd: Nested Paging enabled Dec 12 18:45:12.525797 kernel: kvm_amd: LBR virtualization supported Dec 12 18:45:12.526668 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Dec 12 18:45:12.527608 kernel: kvm_amd: Virtual GIF supported Dec 12 18:45:12.533071 systemd[1]: Reached target sockets.target - Socket Units. Dec 12 18:45:12.535043 systemd[1]: Reached target basic.target - Basic System. Dec 12 18:45:12.536973 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 12 18:45:12.537381 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 12 18:45:12.540970 systemd[1]: Starting containerd.service - containerd container runtime... Dec 12 18:45:12.544517 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 12 18:45:12.555451 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 12 18:45:12.560567 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 12 18:45:12.562558 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 12 18:45:12.564247 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 12 18:45:12.566378 kernel: EDAC MC: Ver: 3.0.0 Dec 12 18:45:12.568488 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Dec 12 18:45:12.573712 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 12 18:45:12.574710 jq[1507]: false Dec 12 18:45:12.576960 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 12 18:45:12.580064 extend-filesystems[1508]: Found /dev/vda6 Dec 12 18:45:12.582830 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 12 18:45:12.583269 extend-filesystems[1508]: Found /dev/vda9 Dec 12 18:45:12.586376 extend-filesystems[1508]: Checking size of /dev/vda9 Dec 12 18:45:12.587168 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 12 18:45:12.593796 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 12 18:45:12.596745 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Refreshing passwd entry cache Dec 12 18:45:12.597033 oslogin_cache_refresh[1510]: Refreshing passwd entry cache Dec 12 18:45:12.598617 extend-filesystems[1508]: Resized partition /dev/vda9 Dec 12 18:45:12.603422 extend-filesystems[1527]: resize2fs 1.47.3 (8-Jul-2025) Dec 12 18:45:12.603693 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 12 18:45:12.610795 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Dec 12 18:45:12.611516 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 12 18:45:12.612194 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 12 18:45:12.613827 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Failure getting users, quitting Dec 12 18:45:12.613820 oslogin_cache_refresh[1510]: Failure getting users, quitting Dec 12 18:45:12.613902 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 12 18:45:12.613902 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Refreshing group entry cache Dec 12 18:45:12.613842 oslogin_cache_refresh[1510]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 12 18:45:12.613891 oslogin_cache_refresh[1510]: Refreshing group entry cache Dec 12 18:45:12.614643 systemd[1]: Starting update-engine.service - Update Engine... Dec 12 18:45:12.619532 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 12 18:45:12.621782 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Failure getting groups, quitting Dec 12 18:45:12.621782 google_oslogin_nss_cache[1510]: oslogin_cache_refresh[1510]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 12 18:45:12.620449 oslogin_cache_refresh[1510]: Failure getting groups, quitting Dec 12 18:45:12.620462 oslogin_cache_refresh[1510]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 12 18:45:12.629395 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 12 18:45:12.632641 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 12 18:45:12.633020 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 12 18:45:12.633735 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Dec 12 18:45:12.634206 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Dec 12 18:45:12.637166 systemd[1]: motdgen.service: Deactivated successfully. Dec 12 18:45:12.640623 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 12 18:45:12.644469 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 12 18:45:12.645210 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 12 18:45:12.662158 (ntainerd)[1543]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 12 18:45:12.675162 jq[1535]: true Dec 12 18:45:12.675521 update_engine[1533]: I20251212 18:45:12.675335 1533 main.cc:92] Flatcar Update Engine starting Dec 12 18:45:12.681294 systemd-logind[1524]: Watching system buttons on /dev/input/event2 (Power Button) Dec 12 18:45:12.681320 systemd-logind[1524]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 12 18:45:12.682555 systemd-logind[1524]: New seat seat0. Dec 12 18:45:12.685387 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Dec 12 18:45:12.700756 systemd[1]: Started systemd-logind.service - User Login Management. Dec 12 18:45:12.704268 tar[1541]: linux-amd64/LICENSE Dec 12 18:45:12.707236 tar[1541]: linux-amd64/helm Dec 12 18:45:12.709250 extend-filesystems[1527]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 12 18:45:12.709250 extend-filesystems[1527]: old_desc_blocks = 1, new_desc_blocks = 1 Dec 12 18:45:12.709250 extend-filesystems[1527]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Dec 12 18:45:12.717212 extend-filesystems[1508]: Resized filesystem in /dev/vda9 Dec 12 18:45:12.710984 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 12 18:45:12.719556 jq[1552]: true Dec 12 18:45:12.711257 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 12 18:45:12.729180 dbus-daemon[1505]: [system] SELinux support is enabled Dec 12 18:45:12.735422 update_engine[1533]: I20251212 18:45:12.735375 1533 update_check_scheduler.cc:74] Next update check in 8m56s Dec 12 18:45:12.762789 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 12 18:45:12.767112 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 12 18:45:12.775549 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 12 18:45:12.775581 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 12 18:45:12.777655 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 12 18:45:12.777674 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 12 18:45:12.778664 dbus-daemon[1505]: [system] Successfully activated service 'org.freedesktop.systemd1' Dec 12 18:45:12.780463 systemd[1]: Started update-engine.service - Update Engine. Dec 12 18:45:12.785105 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 12 18:45:12.816190 sshd_keygen[1536]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 12 18:45:12.834372 bash[1583]: Updated "/home/core/.ssh/authorized_keys" Dec 12 18:45:12.835244 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 12 18:45:12.839275 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Dec 12 18:45:12.843816 locksmithd[1568]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 12 18:45:12.847484 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 12 18:45:12.852048 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 12 18:45:12.859900 containerd[1543]: time="2025-12-12T18:45:12Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Dec 12 18:45:12.860529 containerd[1543]: time="2025-12-12T18:45:12.860493683Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Dec 12 18:45:12.870183 containerd[1543]: time="2025-12-12T18:45:12.870116225Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.152µs" Dec 12 18:45:12.870183 containerd[1543]: time="2025-12-12T18:45:12.870174494Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Dec 12 18:45:12.870260 containerd[1543]: time="2025-12-12T18:45:12.870200082Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Dec 12 18:45:12.870524 containerd[1543]: time="2025-12-12T18:45:12.870489725Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Dec 12 18:45:12.870524 containerd[1543]: time="2025-12-12T18:45:12.870523208Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Dec 12 18:45:12.870615 containerd[1543]: time="2025-12-12T18:45:12.870556180Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 12 18:45:12.870669 containerd[1543]: time="2025-12-12T18:45:12.870639556Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 12 18:45:12.870669 containerd[1543]: time="2025-12-12T18:45:12.870661447Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871005 containerd[1543]: time="2025-12-12T18:45:12.870971028Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871005 containerd[1543]: time="2025-12-12T18:45:12.870995123Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871066 containerd[1543]: time="2025-12-12T18:45:12.871010592Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871066 containerd[1543]: time="2025-12-12T18:45:12.871020320Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871119 containerd[1543]: time="2025-12-12T18:45:12.871110409Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871405 containerd[1543]: time="2025-12-12T18:45:12.871368924Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871453 containerd[1543]: time="2025-12-12T18:45:12.871420761Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 12 18:45:12.871453 containerd[1543]: time="2025-12-12T18:45:12.871438464Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Dec 12 18:45:12.871507 containerd[1543]: time="2025-12-12T18:45:12.871481976Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Dec 12 18:45:12.872546 containerd[1543]: time="2025-12-12T18:45:12.872515734Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Dec 12 18:45:12.872621 containerd[1543]: time="2025-12-12T18:45:12.872595293Z" level=info msg="metadata content store policy set" policy=shared Dec 12 18:45:12.875043 systemd[1]: issuegen.service: Deactivated successfully. Dec 12 18:45:12.875370 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 12 18:45:12.878238 containerd[1543]: time="2025-12-12T18:45:12.878205092Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Dec 12 18:45:12.878290 containerd[1543]: time="2025-12-12T18:45:12.878248142Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Dec 12 18:45:12.878290 containerd[1543]: time="2025-12-12T18:45:12.878262149Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Dec 12 18:45:12.878290 containerd[1543]: time="2025-12-12T18:45:12.878273470Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Dec 12 18:45:12.878405 containerd[1543]: time="2025-12-12T18:45:12.878302043Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Dec 12 18:45:12.878405 containerd[1543]: time="2025-12-12T18:45:12.878315198Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Dec 12 18:45:12.878405 containerd[1543]: time="2025-12-12T18:45:12.878328784Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Dec 12 18:45:12.878480 containerd[1543]: time="2025-12-12T18:45:12.878339995Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Dec 12 18:45:12.878480 containerd[1543]: time="2025-12-12T18:45:12.878441485Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Dec 12 18:45:12.878531 containerd[1543]: time="2025-12-12T18:45:12.878502820Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Dec 12 18:45:12.878531 containerd[1543]: time="2025-12-12T18:45:12.878517327Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Dec 12 18:45:12.878577 containerd[1543]: time="2025-12-12T18:45:12.878533748Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Dec 12 18:45:12.878702 containerd[1543]: time="2025-12-12T18:45:12.878671196Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Dec 12 18:45:12.878702 containerd[1543]: time="2025-12-12T18:45:12.878695541Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Dec 12 18:45:12.878743 containerd[1543]: time="2025-12-12T18:45:12.878710519Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Dec 12 18:45:12.878743 containerd[1543]: time="2025-12-12T18:45:12.878736057Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Dec 12 18:45:12.878790 containerd[1543]: time="2025-12-12T18:45:12.878746948Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Dec 12 18:45:12.878790 containerd[1543]: time="2025-12-12T18:45:12.878757497Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Dec 12 18:45:12.878790 containerd[1543]: time="2025-12-12T18:45:12.878768308Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Dec 12 18:45:12.878790 containerd[1543]: time="2025-12-12T18:45:12.878778787Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Dec 12 18:45:12.878873 containerd[1543]: time="2025-12-12T18:45:12.878793265Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Dec 12 18:45:12.878873 containerd[1543]: time="2025-12-12T18:45:12.878805517Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Dec 12 18:45:12.878873 containerd[1543]: time="2025-12-12T18:45:12.878820265Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Dec 12 18:45:12.878873 containerd[1543]: time="2025-12-12T18:45:12.878864117Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Dec 12 18:45:12.878873 containerd[1543]: time="2025-12-12T18:45:12.878876190Z" level=info msg="Start snapshots syncer" Dec 12 18:45:12.878967 containerd[1543]: time="2025-12-12T18:45:12.878922537Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Dec 12 18:45:12.879270 containerd[1543]: time="2025-12-12T18:45:12.879229182Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Dec 12 18:45:12.879260 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 12 18:45:12.879447 containerd[1543]: time="2025-12-12T18:45:12.879290106Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Dec 12 18:45:12.879447 containerd[1543]: time="2025-12-12T18:45:12.879339859Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879493548Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879515038Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879525207Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879534595Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879557457Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879568067Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879577635Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879596250Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879606249Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879616157Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879658677Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879675188Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 12 18:45:12.881402 containerd[1543]: time="2025-12-12T18:45:12.879683363Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879691759Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879754517Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879763774Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879791286Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879813768Z" level=info msg="runtime interface created" Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879826342Z" level=info msg="created NRI interface" Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879837492Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879847852Z" level=info msg="Connect containerd service" Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.879865004Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 12 18:45:12.881743 containerd[1543]: time="2025-12-12T18:45:12.880763549Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 12 18:45:12.899323 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 12 18:45:12.904422 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 12 18:45:12.908501 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 12 18:45:12.911191 systemd[1]: Reached target getty.target - Login Prompts. Dec 12 18:45:12.967750 containerd[1543]: time="2025-12-12T18:45:12.967690687Z" level=info msg="Start subscribing containerd event" Dec 12 18:45:12.967750 containerd[1543]: time="2025-12-12T18:45:12.967752523Z" level=info msg="Start recovering state" Dec 12 18:45:12.967878 containerd[1543]: time="2025-12-12T18:45:12.967850787Z" level=info msg="Start event monitor" Dec 12 18:45:12.967878 containerd[1543]: time="2025-12-12T18:45:12.967869232Z" level=info msg="Start cni network conf syncer for default" Dec 12 18:45:12.967878 containerd[1543]: time="2025-12-12T18:45:12.967876626Z" level=info msg="Start streaming server" Dec 12 18:45:12.967951 containerd[1543]: time="2025-12-12T18:45:12.967887105Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Dec 12 18:45:12.967951 containerd[1543]: time="2025-12-12T18:45:12.967894379Z" level=info msg="runtime interface starting up..." Dec 12 18:45:12.967951 containerd[1543]: time="2025-12-12T18:45:12.967900320Z" level=info msg="starting plugins..." Dec 12 18:45:12.967951 containerd[1543]: time="2025-12-12T18:45:12.967919556Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Dec 12 18:45:12.967951 containerd[1543]: time="2025-12-12T18:45:12.967941898Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 12 18:45:12.968045 containerd[1543]: time="2025-12-12T18:45:12.967993445Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 12 18:45:12.968082 containerd[1543]: time="2025-12-12T18:45:12.968060971Z" level=info msg="containerd successfully booted in 0.108683s" Dec 12 18:45:12.968203 systemd[1]: Started containerd.service - containerd container runtime. Dec 12 18:45:13.024271 tar[1541]: linux-amd64/README.md Dec 12 18:45:13.052112 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 12 18:45:13.464274 systemd-networkd[1432]: eth0: Gained IPv6LL Dec 12 18:45:13.467911 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 12 18:45:13.470950 systemd[1]: Reached target network-online.target - Network is Online. Dec 12 18:45:13.474759 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Dec 12 18:45:13.478718 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:13.482110 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 12 18:45:13.518707 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 12 18:45:13.521024 systemd[1]: coreos-metadata.service: Deactivated successfully. Dec 12 18:45:13.521309 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Dec 12 18:45:13.524257 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 12 18:45:14.248669 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:14.251245 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 12 18:45:14.254105 systemd[1]: Startup finished in 4.426s (kernel) + 6.394s (initrd) + 4.255s (userspace) = 15.076s. Dec 12 18:45:14.254134 (kubelet)[1650]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:45:14.666569 kubelet[1650]: E1212 18:45:14.666497 1650 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:45:14.670502 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:45:14.670731 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:45:14.671227 systemd[1]: kubelet.service: Consumed 961ms CPU time, 256.4M memory peak. Dec 12 18:45:17.592336 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 12 18:45:17.593675 systemd[1]: Started sshd@0-10.0.0.119:22-10.0.0.1:42794.service - OpenSSH per-connection server daemon (10.0.0.1:42794). Dec 12 18:45:17.674521 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 42794 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:17.676643 sshd-session[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:17.684707 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 12 18:45:17.686092 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 12 18:45:17.693889 systemd-logind[1524]: New session 1 of user core. Dec 12 18:45:17.745320 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 12 18:45:17.748832 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 12 18:45:17.771227 (systemd)[1669]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 12 18:45:17.774099 systemd-logind[1524]: New session c1 of user core. Dec 12 18:45:17.930522 systemd[1669]: Queued start job for default target default.target. Dec 12 18:45:17.945852 systemd[1669]: Created slice app.slice - User Application Slice. Dec 12 18:45:17.945880 systemd[1669]: Reached target paths.target - Paths. Dec 12 18:45:17.945926 systemd[1669]: Reached target timers.target - Timers. Dec 12 18:45:17.947602 systemd[1669]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 12 18:45:17.961342 systemd[1669]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 12 18:45:17.961533 systemd[1669]: Reached target sockets.target - Sockets. Dec 12 18:45:17.961592 systemd[1669]: Reached target basic.target - Basic System. Dec 12 18:45:17.961643 systemd[1669]: Reached target default.target - Main User Target. Dec 12 18:45:17.961692 systemd[1669]: Startup finished in 180ms. Dec 12 18:45:17.961869 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 12 18:45:17.963445 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 12 18:45:18.031503 systemd[1]: Started sshd@1-10.0.0.119:22-10.0.0.1:42802.service - OpenSSH per-connection server daemon (10.0.0.1:42802). Dec 12 18:45:18.105289 sshd[1680]: Accepted publickey for core from 10.0.0.1 port 42802 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:18.107128 sshd-session[1680]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:18.111800 systemd-logind[1524]: New session 2 of user core. Dec 12 18:45:18.125701 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 12 18:45:18.182194 sshd[1683]: Connection closed by 10.0.0.1 port 42802 Dec 12 18:45:18.182548 sshd-session[1680]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:18.192765 systemd[1]: sshd@1-10.0.0.119:22-10.0.0.1:42802.service: Deactivated successfully. Dec 12 18:45:18.194953 systemd[1]: session-2.scope: Deactivated successfully. Dec 12 18:45:18.195848 systemd-logind[1524]: Session 2 logged out. Waiting for processes to exit. Dec 12 18:45:18.198821 systemd[1]: Started sshd@2-10.0.0.119:22-10.0.0.1:42812.service - OpenSSH per-connection server daemon (10.0.0.1:42812). Dec 12 18:45:18.199759 systemd-logind[1524]: Removed session 2. Dec 12 18:45:18.265091 sshd[1689]: Accepted publickey for core from 10.0.0.1 port 42812 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:18.267111 sshd-session[1689]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:18.272491 systemd-logind[1524]: New session 3 of user core. Dec 12 18:45:18.282555 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 12 18:45:18.334034 sshd[1692]: Connection closed by 10.0.0.1 port 42812 Dec 12 18:45:18.334494 sshd-session[1689]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:18.348310 systemd[1]: sshd@2-10.0.0.119:22-10.0.0.1:42812.service: Deactivated successfully. Dec 12 18:45:18.350140 systemd[1]: session-3.scope: Deactivated successfully. Dec 12 18:45:18.351003 systemd-logind[1524]: Session 3 logged out. Waiting for processes to exit. Dec 12 18:45:18.353808 systemd[1]: Started sshd@3-10.0.0.119:22-10.0.0.1:42826.service - OpenSSH per-connection server daemon (10.0.0.1:42826). Dec 12 18:45:18.354641 systemd-logind[1524]: Removed session 3. Dec 12 18:45:18.435093 sshd[1698]: Accepted publickey for core from 10.0.0.1 port 42826 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:18.436612 sshd-session[1698]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:18.441887 systemd-logind[1524]: New session 4 of user core. Dec 12 18:45:18.451657 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 12 18:45:18.509231 sshd[1701]: Connection closed by 10.0.0.1 port 42826 Dec 12 18:45:18.509651 sshd-session[1698]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:18.522201 systemd[1]: sshd@3-10.0.0.119:22-10.0.0.1:42826.service: Deactivated successfully. Dec 12 18:45:18.524106 systemd[1]: session-4.scope: Deactivated successfully. Dec 12 18:45:18.524927 systemd-logind[1524]: Session 4 logged out. Waiting for processes to exit. Dec 12 18:45:18.527818 systemd[1]: Started sshd@4-10.0.0.119:22-10.0.0.1:42836.service - OpenSSH per-connection server daemon (10.0.0.1:42836). Dec 12 18:45:18.528448 systemd-logind[1524]: Removed session 4. Dec 12 18:45:18.589299 sshd[1707]: Accepted publickey for core from 10.0.0.1 port 42836 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:18.591136 sshd-session[1707]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:18.596719 systemd-logind[1524]: New session 5 of user core. Dec 12 18:45:18.610563 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 12 18:45:18.670302 sudo[1711]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 12 18:45:18.670656 sudo[1711]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:45:18.684897 sudo[1711]: pam_unix(sudo:session): session closed for user root Dec 12 18:45:18.686925 sshd[1710]: Connection closed by 10.0.0.1 port 42836 Dec 12 18:45:18.686948 sshd-session[1707]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:18.701124 systemd[1]: sshd@4-10.0.0.119:22-10.0.0.1:42836.service: Deactivated successfully. Dec 12 18:45:18.702862 systemd[1]: session-5.scope: Deactivated successfully. Dec 12 18:45:18.703683 systemd-logind[1524]: Session 5 logged out. Waiting for processes to exit. Dec 12 18:45:18.706559 systemd[1]: Started sshd@5-10.0.0.119:22-10.0.0.1:42848.service - OpenSSH per-connection server daemon (10.0.0.1:42848). Dec 12 18:45:18.707240 systemd-logind[1524]: Removed session 5. Dec 12 18:45:18.770644 sshd[1717]: Accepted publickey for core from 10.0.0.1 port 42848 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:18.772601 sshd-session[1717]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:18.777432 systemd-logind[1524]: New session 6 of user core. Dec 12 18:45:18.798782 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 12 18:45:18.855547 sudo[1722]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 12 18:45:18.855920 sudo[1722]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:45:19.291046 sudo[1722]: pam_unix(sudo:session): session closed for user root Dec 12 18:45:19.298946 sudo[1721]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Dec 12 18:45:19.299395 sudo[1721]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:45:19.310660 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 12 18:45:19.352827 augenrules[1744]: No rules Dec 12 18:45:19.353853 systemd[1]: audit-rules.service: Deactivated successfully. Dec 12 18:45:19.354228 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 12 18:45:19.355645 sudo[1721]: pam_unix(sudo:session): session closed for user root Dec 12 18:45:19.357705 sshd[1720]: Connection closed by 10.0.0.1 port 42848 Dec 12 18:45:19.358024 sshd-session[1717]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:19.372432 systemd[1]: sshd@5-10.0.0.119:22-10.0.0.1:42848.service: Deactivated successfully. Dec 12 18:45:19.374429 systemd[1]: session-6.scope: Deactivated successfully. Dec 12 18:45:19.375233 systemd-logind[1524]: Session 6 logged out. Waiting for processes to exit. Dec 12 18:45:19.378233 systemd[1]: Started sshd@6-10.0.0.119:22-10.0.0.1:42864.service - OpenSSH per-connection server daemon (10.0.0.1:42864). Dec 12 18:45:19.378871 systemd-logind[1524]: Removed session 6. Dec 12 18:45:19.448645 sshd[1753]: Accepted publickey for core from 10.0.0.1 port 42864 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:45:19.451654 sshd-session[1753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:45:19.457876 systemd-logind[1524]: New session 7 of user core. Dec 12 18:45:19.467622 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 12 18:45:19.526486 sudo[1757]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 12 18:45:19.526984 sudo[1757]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 12 18:45:20.410268 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 12 18:45:20.427768 (dockerd)[1778]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 12 18:45:20.961501 dockerd[1778]: time="2025-12-12T18:45:20.961414728Z" level=info msg="Starting up" Dec 12 18:45:20.962565 dockerd[1778]: time="2025-12-12T18:45:20.962500650Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Dec 12 18:45:21.044297 dockerd[1778]: time="2025-12-12T18:45:21.044198888Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Dec 12 18:45:21.111311 dockerd[1778]: time="2025-12-12T18:45:21.111224661Z" level=info msg="Loading containers: start." Dec 12 18:45:21.125573 kernel: Initializing XFRM netlink socket Dec 12 18:45:21.523241 systemd-networkd[1432]: docker0: Link UP Dec 12 18:45:21.528015 dockerd[1778]: time="2025-12-12T18:45:21.527959401Z" level=info msg="Loading containers: done." Dec 12 18:45:21.548915 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2856616415-merged.mount: Deactivated successfully. Dec 12 18:45:21.550866 dockerd[1778]: time="2025-12-12T18:45:21.550784876Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 12 18:45:21.550941 dockerd[1778]: time="2025-12-12T18:45:21.550910986Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Dec 12 18:45:21.551020 dockerd[1778]: time="2025-12-12T18:45:21.551001742Z" level=info msg="Initializing buildkit" Dec 12 18:45:21.583472 dockerd[1778]: time="2025-12-12T18:45:21.583416636Z" level=info msg="Completed buildkit initialization" Dec 12 18:45:21.587523 dockerd[1778]: time="2025-12-12T18:45:21.587483391Z" level=info msg="Daemon has completed initialization" Dec 12 18:45:21.587644 dockerd[1778]: time="2025-12-12T18:45:21.587548785Z" level=info msg="API listen on /run/docker.sock" Dec 12 18:45:21.587725 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 12 18:45:22.365397 containerd[1543]: time="2025-12-12T18:45:22.365315304Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\"" Dec 12 18:45:23.052804 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4033148926.mount: Deactivated successfully. Dec 12 18:45:24.246471 containerd[1543]: time="2025-12-12T18:45:24.246407698Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:24.247219 containerd[1543]: time="2025-12-12T18:45:24.247187425Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.3: active requests=0, bytes read=27068073" Dec 12 18:45:24.248688 containerd[1543]: time="2025-12-12T18:45:24.248647889Z" level=info msg="ImageCreate event name:\"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:24.251774 containerd[1543]: time="2025-12-12T18:45:24.251698966Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:24.252856 containerd[1543]: time="2025-12-12T18:45:24.252805867Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.3\" with image id \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\", size \"27064672\" in 1.88741508s" Dec 12 18:45:24.252856 containerd[1543]: time="2025-12-12T18:45:24.252850715Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\" returns image reference \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\"" Dec 12 18:45:24.253674 containerd[1543]: time="2025-12-12T18:45:24.253645768Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\"" Dec 12 18:45:24.755237 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 12 18:45:24.757303 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:25.149636 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:25.155815 (kubelet)[2059]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:45:25.375157 kubelet[2059]: E1212 18:45:25.375067 2059 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:45:25.385640 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:45:25.385848 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:45:25.386332 systemd[1]: kubelet.service: Consumed 506ms CPU time, 109.4M memory peak. Dec 12 18:45:25.781833 containerd[1543]: time="2025-12-12T18:45:25.781742178Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:25.782681 containerd[1543]: time="2025-12-12T18:45:25.782605335Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.3: active requests=0, bytes read=21162440" Dec 12 18:45:25.787190 containerd[1543]: time="2025-12-12T18:45:25.787111415Z" level=info msg="ImageCreate event name:\"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:25.790362 containerd[1543]: time="2025-12-12T18:45:25.790295868Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:25.791616 containerd[1543]: time="2025-12-12T18:45:25.791566132Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.3\" with image id \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\", size \"22819474\" in 1.537854117s" Dec 12 18:45:25.791616 containerd[1543]: time="2025-12-12T18:45:25.791607311Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\" returns image reference \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\"" Dec 12 18:45:25.792521 containerd[1543]: time="2025-12-12T18:45:25.792458507Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\"" Dec 12 18:45:28.060156 containerd[1543]: time="2025-12-12T18:45:28.060075444Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:28.060969 containerd[1543]: time="2025-12-12T18:45:28.060936151Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.3: active requests=0, bytes read=15725927" Dec 12 18:45:28.062279 containerd[1543]: time="2025-12-12T18:45:28.062239703Z" level=info msg="ImageCreate event name:\"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:28.065515 containerd[1543]: time="2025-12-12T18:45:28.065453088Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:28.066786 containerd[1543]: time="2025-12-12T18:45:28.066738959Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.3\" with image id \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\", size \"17382979\" in 2.274231466s" Dec 12 18:45:28.066786 containerd[1543]: time="2025-12-12T18:45:28.066779806Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\" returns image reference \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\"" Dec 12 18:45:28.069438 containerd[1543]: time="2025-12-12T18:45:28.069374827Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\"" Dec 12 18:45:29.242972 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2874527865.mount: Deactivated successfully. Dec 12 18:45:29.474533 containerd[1543]: time="2025-12-12T18:45:29.474449158Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:29.475668 containerd[1543]: time="2025-12-12T18:45:29.475630972Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.3: active requests=0, bytes read=25965293" Dec 12 18:45:29.477085 containerd[1543]: time="2025-12-12T18:45:29.477043681Z" level=info msg="ImageCreate event name:\"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:29.479720 containerd[1543]: time="2025-12-12T18:45:29.479682316Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:29.480413 containerd[1543]: time="2025-12-12T18:45:29.480344745Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.3\" with image id \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\", repo tag \"registry.k8s.io/kube-proxy:v1.34.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\", size \"25964312\" in 1.410903139s" Dec 12 18:45:29.480413 containerd[1543]: time="2025-12-12T18:45:29.480407824Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\" returns image reference \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\"" Dec 12 18:45:29.481296 containerd[1543]: time="2025-12-12T18:45:29.481012304Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Dec 12 18:45:30.066732 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3095507168.mount: Deactivated successfully. Dec 12 18:45:30.823872 containerd[1543]: time="2025-12-12T18:45:30.823795668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:30.824613 containerd[1543]: time="2025-12-12T18:45:30.824594079Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22388007" Dec 12 18:45:30.826171 containerd[1543]: time="2025-12-12T18:45:30.826114298Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:30.829343 containerd[1543]: time="2025-12-12T18:45:30.829300056Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:30.830272 containerd[1543]: time="2025-12-12T18:45:30.830247144Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 1.349189494s" Dec 12 18:45:30.830322 containerd[1543]: time="2025-12-12T18:45:30.830275353Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Dec 12 18:45:30.830883 containerd[1543]: time="2025-12-12T18:45:30.830848021Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Dec 12 18:45:31.799510 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3641076285.mount: Deactivated successfully. Dec 12 18:45:31.832907 containerd[1543]: time="2025-12-12T18:45:31.832795240Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:31.833703 containerd[1543]: time="2025-12-12T18:45:31.833656049Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=321218" Dec 12 18:45:31.834903 containerd[1543]: time="2025-12-12T18:45:31.834853478Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:31.836746 containerd[1543]: time="2025-12-12T18:45:31.836710354Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:31.837293 containerd[1543]: time="2025-12-12T18:45:31.837259924Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 1.006388574s" Dec 12 18:45:31.837293 containerd[1543]: time="2025-12-12T18:45:31.837286602Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Dec 12 18:45:31.837979 containerd[1543]: time="2025-12-12T18:45:31.837801905Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\"" Dec 12 18:45:32.403478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount461134679.mount: Deactivated successfully. Dec 12 18:45:35.172642 containerd[1543]: time="2025-12-12T18:45:35.172577029Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.4-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:35.173619 containerd[1543]: time="2025-12-12T18:45:35.173569910Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.4-0: active requests=0, bytes read=74166814" Dec 12 18:45:35.174971 containerd[1543]: time="2025-12-12T18:45:35.174938109Z" level=info msg="ImageCreate event name:\"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:35.177793 containerd[1543]: time="2025-12-12T18:45:35.177758017Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:45:35.179005 containerd[1543]: time="2025-12-12T18:45:35.178972061Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.4-0\" with image id \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\", repo tag \"registry.k8s.io/etcd:3.6.4-0\", repo digest \"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\", size \"74311308\" in 3.341142069s" Dec 12 18:45:35.179005 containerd[1543]: time="2025-12-12T18:45:35.179002858Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\" returns image reference \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\"" Dec 12 18:45:35.398921 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 12 18:45:35.400618 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:35.607124 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:35.628708 (kubelet)[2226]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 12 18:45:35.668467 kubelet[2226]: E1212 18:45:35.668388 2226 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 12 18:45:35.673109 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 12 18:45:35.673330 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 12 18:45:35.673751 systemd[1]: kubelet.service: Consumed 222ms CPU time, 109.8M memory peak. Dec 12 18:45:38.774292 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:38.774502 systemd[1]: kubelet.service: Consumed 222ms CPU time, 109.8M memory peak. Dec 12 18:45:38.776720 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:38.806112 systemd[1]: Reload requested from client PID 2243 ('systemctl') (unit session-7.scope)... Dec 12 18:45:38.806127 systemd[1]: Reloading... Dec 12 18:45:38.919435 zram_generator::config[2291]: No configuration found. Dec 12 18:45:39.262078 systemd[1]: Reloading finished in 455 ms. Dec 12 18:45:39.350322 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 12 18:45:39.350482 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 12 18:45:39.350923 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:39.350984 systemd[1]: kubelet.service: Consumed 168ms CPU time, 98.2M memory peak. Dec 12 18:45:39.353231 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:39.583407 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:39.596792 (kubelet)[2333]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 12 18:45:39.630309 kubelet[2333]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 12 18:45:39.630309 kubelet[2333]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:45:39.630725 kubelet[2333]: I1212 18:45:39.630615 2333 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 12 18:45:39.859610 kubelet[2333]: I1212 18:45:39.859502 2333 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Dec 12 18:45:39.859610 kubelet[2333]: I1212 18:45:39.859528 2333 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 12 18:45:39.860105 kubelet[2333]: I1212 18:45:39.860086 2333 watchdog_linux.go:95] "Systemd watchdog is not enabled" Dec 12 18:45:39.860135 kubelet[2333]: I1212 18:45:39.860110 2333 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 12 18:45:39.860383 kubelet[2333]: I1212 18:45:39.860368 2333 server.go:956] "Client rotation is on, will bootstrap in background" Dec 12 18:45:40.972423 kubelet[2333]: E1212 18:45:40.972367 2333 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.119:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Dec 12 18:45:40.977473 kubelet[2333]: I1212 18:45:40.977421 2333 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 12 18:45:40.984819 kubelet[2333]: I1212 18:45:40.984786 2333 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 12 18:45:40.989694 kubelet[2333]: I1212 18:45:40.989660 2333 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Dec 12 18:45:40.990459 kubelet[2333]: I1212 18:45:40.990415 2333 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 12 18:45:40.990599 kubelet[2333]: I1212 18:45:40.990442 2333 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 12 18:45:40.990599 kubelet[2333]: I1212 18:45:40.990598 2333 topology_manager.go:138] "Creating topology manager with none policy" Dec 12 18:45:40.990802 kubelet[2333]: I1212 18:45:40.990608 2333 container_manager_linux.go:306] "Creating device plugin manager" Dec 12 18:45:40.990802 kubelet[2333]: I1212 18:45:40.990721 2333 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Dec 12 18:45:41.229835 kubelet[2333]: I1212 18:45:41.229695 2333 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:45:41.230009 kubelet[2333]: I1212 18:45:41.229919 2333 kubelet.go:475] "Attempting to sync node with API server" Dec 12 18:45:41.230009 kubelet[2333]: I1212 18:45:41.229936 2333 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 12 18:45:41.230009 kubelet[2333]: I1212 18:45:41.229957 2333 kubelet.go:387] "Adding apiserver pod source" Dec 12 18:45:41.230009 kubelet[2333]: I1212 18:45:41.229974 2333 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 12 18:45:41.230931 kubelet[2333]: E1212 18:45:41.230663 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.119:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 12 18:45:41.231086 kubelet[2333]: E1212 18:45:41.231052 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.119:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 12 18:45:41.233103 kubelet[2333]: I1212 18:45:41.233080 2333 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Dec 12 18:45:41.233657 kubelet[2333]: I1212 18:45:41.233627 2333 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Dec 12 18:45:41.233657 kubelet[2333]: I1212 18:45:41.233655 2333 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Dec 12 18:45:41.233716 kubelet[2333]: W1212 18:45:41.233710 2333 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 12 18:45:41.237639 kubelet[2333]: I1212 18:45:41.237608 2333 server.go:1262] "Started kubelet" Dec 12 18:45:41.239001 kubelet[2333]: I1212 18:45:41.238068 2333 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Dec 12 18:45:41.239001 kubelet[2333]: I1212 18:45:41.238808 2333 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 12 18:45:41.239166 kubelet[2333]: I1212 18:45:41.239104 2333 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 12 18:45:41.239248 kubelet[2333]: I1212 18:45:41.239233 2333 server_v1.go:49] "podresources" method="list" useActivePods=true Dec 12 18:45:41.239729 kubelet[2333]: I1212 18:45:41.239705 2333 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 12 18:45:41.241044 kubelet[2333]: I1212 18:45:41.241020 2333 server.go:310] "Adding debug handlers to kubelet server" Dec 12 18:45:41.242009 kubelet[2333]: E1212 18:45:41.241988 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:41.242123 kubelet[2333]: I1212 18:45:41.242025 2333 volume_manager.go:313] "Starting Kubelet Volume Manager" Dec 12 18:45:41.242192 kubelet[2333]: I1212 18:45:41.242174 2333 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 12 18:45:41.242223 kubelet[2333]: I1212 18:45:41.242219 2333 reconciler.go:29] "Reconciler: start to sync state" Dec 12 18:45:41.242661 kubelet[2333]: E1212 18:45:41.242630 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.119:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 12 18:45:41.242897 kubelet[2333]: E1212 18:45:41.242713 2333 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 12 18:45:41.242997 kubelet[2333]: E1212 18:45:41.242932 2333 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.119:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.119:6443: connect: connection refused" interval="200ms" Dec 12 18:45:41.242997 kubelet[2333]: I1212 18:45:41.242990 2333 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 12 18:45:41.243477 kubelet[2333]: E1212 18:45:41.240785 2333 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.119:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.119:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.18808c243f47300d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-12-12 18:45:41.237575693 +0000 UTC m=+1.637128540,LastTimestamp:2025-12-12 18:45:41.237575693 +0000 UTC m=+1.637128540,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Dec 12 18:45:41.247622 kubelet[2333]: I1212 18:45:41.247497 2333 factory.go:223] Registration of the containerd container factory successfully Dec 12 18:45:41.247622 kubelet[2333]: I1212 18:45:41.247532 2333 factory.go:223] Registration of the systemd container factory successfully Dec 12 18:45:41.247717 kubelet[2333]: I1212 18:45:41.247636 2333 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 12 18:45:41.261498 kubelet[2333]: I1212 18:45:41.261466 2333 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 12 18:45:41.261498 kubelet[2333]: I1212 18:45:41.261489 2333 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 12 18:45:41.261682 kubelet[2333]: I1212 18:45:41.261510 2333 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:45:41.262729 kubelet[2333]: I1212 18:45:41.262687 2333 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Dec 12 18:45:41.264407 kubelet[2333]: I1212 18:45:41.264375 2333 policy_none.go:49] "None policy: Start" Dec 12 18:45:41.264407 kubelet[2333]: I1212 18:45:41.264403 2333 memory_manager.go:187] "Starting memorymanager" policy="None" Dec 12 18:45:41.264519 kubelet[2333]: I1212 18:45:41.264419 2333 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Dec 12 18:45:41.264992 kubelet[2333]: I1212 18:45:41.264970 2333 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Dec 12 18:45:41.265024 kubelet[2333]: I1212 18:45:41.264994 2333 status_manager.go:244] "Starting to sync pod status with apiserver" Dec 12 18:45:41.265050 kubelet[2333]: I1212 18:45:41.265025 2333 kubelet.go:2427] "Starting kubelet main sync loop" Dec 12 18:45:41.265089 kubelet[2333]: E1212 18:45:41.265066 2333 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 12 18:45:41.265772 kubelet[2333]: I1212 18:45:41.265742 2333 policy_none.go:47] "Start" Dec 12 18:45:41.265811 kubelet[2333]: E1212 18:45:41.265782 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.119:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 12 18:45:41.271618 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 12 18:45:41.293840 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 12 18:45:41.310672 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 12 18:45:41.312102 kubelet[2333]: E1212 18:45:41.312074 2333 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Dec 12 18:45:41.312337 kubelet[2333]: I1212 18:45:41.312317 2333 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 12 18:45:41.312456 kubelet[2333]: I1212 18:45:41.312332 2333 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 12 18:45:41.312631 kubelet[2333]: I1212 18:45:41.312607 2333 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 12 18:45:41.313748 kubelet[2333]: E1212 18:45:41.313721 2333 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 12 18:45:41.313794 kubelet[2333]: E1212 18:45:41.313765 2333 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Dec 12 18:45:41.390750 systemd[1]: Created slice kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice - libcontainer container kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice. Dec 12 18:45:41.398230 kubelet[2333]: E1212 18:45:41.398192 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:41.401935 systemd[1]: Created slice kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice - libcontainer container kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice. Dec 12 18:45:41.403812 kubelet[2333]: E1212 18:45:41.403777 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:41.405525 systemd[1]: Created slice kubepods-burstable-pod2f29938ea9ffdae1ba75459fa38a7152.slice - libcontainer container kubepods-burstable-pod2f29938ea9ffdae1ba75459fa38a7152.slice. Dec 12 18:45:41.407079 kubelet[2333]: E1212 18:45:41.407062 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:41.414039 kubelet[2333]: I1212 18:45:41.414004 2333 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Dec 12 18:45:41.414416 kubelet[2333]: E1212 18:45:41.414382 2333 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.119:6443/api/v1/nodes\": dial tcp 10.0.0.119:6443: connect: connection refused" node="localhost" Dec 12 18:45:41.444091 kubelet[2333]: E1212 18:45:41.444031 2333 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.119:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.119:6443: connect: connection refused" interval="400ms" Dec 12 18:45:41.543590 kubelet[2333]: I1212 18:45:41.543525 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:41.543590 kubelet[2333]: I1212 18:45:41.543572 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:41.543590 kubelet[2333]: I1212 18:45:41.543596 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:41.543590 kubelet[2333]: I1212 18:45:41.543613 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:41.543863 kubelet[2333]: I1212 18:45:41.543634 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:41.543863 kubelet[2333]: I1212 18:45:41.543649 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:41.543863 kubelet[2333]: I1212 18:45:41.543664 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:41.543863 kubelet[2333]: I1212 18:45:41.543703 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:41.543863 kubelet[2333]: I1212 18:45:41.543741 2333 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:41.615839 kubelet[2333]: I1212 18:45:41.615790 2333 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Dec 12 18:45:41.616195 kubelet[2333]: E1212 18:45:41.616157 2333 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.119:6443/api/v1/nodes\": dial tcp 10.0.0.119:6443: connect: connection refused" node="localhost" Dec 12 18:45:41.702757 containerd[1543]: time="2025-12-12T18:45:41.702702342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:41.707784 containerd[1543]: time="2025-12-12T18:45:41.707712628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:41.710362 containerd[1543]: time="2025-12-12T18:45:41.710330649Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2f29938ea9ffdae1ba75459fa38a7152,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:41.845458 kubelet[2333]: E1212 18:45:41.845312 2333 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.119:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.119:6443: connect: connection refused" interval="800ms" Dec 12 18:45:42.017465 kubelet[2333]: I1212 18:45:42.017416 2333 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Dec 12 18:45:42.017878 kubelet[2333]: E1212 18:45:42.017736 2333 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.119:6443/api/v1/nodes\": dial tcp 10.0.0.119:6443: connect: connection refused" node="localhost" Dec 12 18:45:42.207943 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount595973242.mount: Deactivated successfully. Dec 12 18:45:42.214882 containerd[1543]: time="2025-12-12T18:45:42.214825199Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:45:42.215100 kubelet[2333]: E1212 18:45:42.215063 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.119:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 12 18:45:42.216662 containerd[1543]: time="2025-12-12T18:45:42.216628731Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Dec 12 18:45:42.219426 containerd[1543]: time="2025-12-12T18:45:42.219396875Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:45:42.220444 containerd[1543]: time="2025-12-12T18:45:42.220420122Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:45:42.222390 containerd[1543]: time="2025-12-12T18:45:42.222325803Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Dec 12 18:45:42.223400 containerd[1543]: time="2025-12-12T18:45:42.223376692Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:45:42.224228 containerd[1543]: time="2025-12-12T18:45:42.224197825Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Dec 12 18:45:42.225205 containerd[1543]: time="2025-12-12T18:45:42.225166979Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 12 18:45:42.225811 containerd[1543]: time="2025-12-12T18:45:42.225760433Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 514.223906ms" Dec 12 18:45:42.229325 containerd[1543]: time="2025-12-12T18:45:42.229275675Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 524.838425ms" Dec 12 18:45:42.231298 containerd[1543]: time="2025-12-12T18:45:42.231264767Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 520.756008ms" Dec 12 18:45:42.254831 containerd[1543]: time="2025-12-12T18:45:42.254481582Z" level=info msg="connecting to shim 5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6" address="unix:///run/containerd/s/9b229002a9298de529a80089e8305a184116bf9078232b4c3d39733b0836f812" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:42.258726 containerd[1543]: time="2025-12-12T18:45:42.258672605Z" level=info msg="connecting to shim 7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60" address="unix:///run/containerd/s/a258b8af54a8da5a6855467942c313802e831cdb6a0f7d891031cfa14702cf60" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:42.268473 containerd[1543]: time="2025-12-12T18:45:42.268422707Z" level=info msg="connecting to shim 865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363" address="unix:///run/containerd/s/0c274cb32bf202a1ce57bb8eb8750b42ec13f66562157906c8145207193b4cab" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:42.296547 systemd[1]: Started cri-containerd-5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6.scope - libcontainer container 5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6. Dec 12 18:45:42.302122 systemd[1]: Started cri-containerd-7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60.scope - libcontainer container 7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60. Dec 12 18:45:42.304459 systemd[1]: Started cri-containerd-865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363.scope - libcontainer container 865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363. Dec 12 18:45:42.357271 containerd[1543]: time="2025-12-12T18:45:42.357123672Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2f29938ea9ffdae1ba75459fa38a7152,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6\"" Dec 12 18:45:42.360208 containerd[1543]: time="2025-12-12T18:45:42.360154769Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,} returns sandbox id \"7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60\"" Dec 12 18:45:42.365055 containerd[1543]: time="2025-12-12T18:45:42.365013241Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,} returns sandbox id \"865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363\"" Dec 12 18:45:42.365248 containerd[1543]: time="2025-12-12T18:45:42.365213839Z" level=info msg="CreateContainer within sandbox \"5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 12 18:45:42.367445 containerd[1543]: time="2025-12-12T18:45:42.367406667Z" level=info msg="CreateContainer within sandbox \"7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 12 18:45:42.370897 containerd[1543]: time="2025-12-12T18:45:42.370848948Z" level=info msg="CreateContainer within sandbox \"865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 12 18:45:42.376923 containerd[1543]: time="2025-12-12T18:45:42.376881553Z" level=info msg="Container 4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:45:42.381978 containerd[1543]: time="2025-12-12T18:45:42.381953358Z" level=info msg="Container e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:45:42.385026 containerd[1543]: time="2025-12-12T18:45:42.384976915Z" level=info msg="Container e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:45:42.388511 containerd[1543]: time="2025-12-12T18:45:42.388456844Z" level=info msg="CreateContainer within sandbox \"5e0027a6ae4f3b530092e77dc485e6517306a5a526834b94225aa26a671d32f6\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc\"" Dec 12 18:45:42.389070 containerd[1543]: time="2025-12-12T18:45:42.389034827Z" level=info msg="StartContainer for \"4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc\"" Dec 12 18:45:42.390201 containerd[1543]: time="2025-12-12T18:45:42.390167011Z" level=info msg="connecting to shim 4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc" address="unix:///run/containerd/s/9b229002a9298de529a80089e8305a184116bf9078232b4c3d39733b0836f812" protocol=ttrpc version=3 Dec 12 18:45:42.391833 containerd[1543]: time="2025-12-12T18:45:42.391796544Z" level=info msg="CreateContainer within sandbox \"7e668eb20c2c344b2588c83eeee89beda1781b7df386226289f7a9c8e1097c60\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2\"" Dec 12 18:45:42.392671 containerd[1543]: time="2025-12-12T18:45:42.392126062Z" level=info msg="StartContainer for \"e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2\"" Dec 12 18:45:42.393368 containerd[1543]: time="2025-12-12T18:45:42.393308969Z" level=info msg="connecting to shim e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2" address="unix:///run/containerd/s/a258b8af54a8da5a6855467942c313802e831cdb6a0f7d891031cfa14702cf60" protocol=ttrpc version=3 Dec 12 18:45:42.396396 containerd[1543]: time="2025-12-12T18:45:42.396026730Z" level=info msg="CreateContainer within sandbox \"865846fad184b97b26e414cbb1cad4125324c4f9712aaadd204fbbbe627d8363\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767\"" Dec 12 18:45:42.397091 containerd[1543]: time="2025-12-12T18:45:42.397057647Z" level=info msg="StartContainer for \"e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767\"" Dec 12 18:45:42.398462 containerd[1543]: time="2025-12-12T18:45:42.398429462Z" level=info msg="connecting to shim e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767" address="unix:///run/containerd/s/0c274cb32bf202a1ce57bb8eb8750b42ec13f66562157906c8145207193b4cab" protocol=ttrpc version=3 Dec 12 18:45:42.412521 systemd[1]: Started cri-containerd-4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc.scope - libcontainer container 4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc. Dec 12 18:45:42.419449 systemd[1]: Started cri-containerd-e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767.scope - libcontainer container e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767. Dec 12 18:45:42.421233 systemd[1]: Started cri-containerd-e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2.scope - libcontainer container e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2. Dec 12 18:45:42.431247 kubelet[2333]: E1212 18:45:42.431196 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.119:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 12 18:45:42.481827 containerd[1543]: time="2025-12-12T18:45:42.480895398Z" level=info msg="StartContainer for \"e5f24dc6e8e812f246432e869ada340c385a4eb4569135ff8439d82620e5f767\" returns successfully" Dec 12 18:45:42.486571 containerd[1543]: time="2025-12-12T18:45:42.486505600Z" level=info msg="StartContainer for \"4f4f22575e075ddeb0b3ea0bd134cc2dc92c6ce6d88bffa0aa7f9c364e8a6bdc\" returns successfully" Dec 12 18:45:42.491630 containerd[1543]: time="2025-12-12T18:45:42.491588032Z" level=info msg="StartContainer for \"e86429fd73ec7f49e9c81922f220a5b2298450fdb267a3c0e4ad71434d369dc2\" returns successfully" Dec 12 18:45:42.508682 kubelet[2333]: E1212 18:45:42.508628 2333 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.119:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.119:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 12 18:45:42.819246 kubelet[2333]: I1212 18:45:42.819211 2333 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Dec 12 18:45:43.273592 kubelet[2333]: E1212 18:45:43.273549 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:43.276529 kubelet[2333]: E1212 18:45:43.276424 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:43.279003 kubelet[2333]: E1212 18:45:43.278974 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:43.818459 kubelet[2333]: E1212 18:45:43.818403 2333 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Dec 12 18:45:43.911299 kubelet[2333]: I1212 18:45:43.911133 2333 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Dec 12 18:45:43.911299 kubelet[2333]: E1212 18:45:43.911200 2333 kubelet_node_status.go:486] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Dec 12 18:45:43.923549 kubelet[2333]: E1212 18:45:43.923492 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.024428 kubelet[2333]: E1212 18:45:44.024329 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.125338 kubelet[2333]: E1212 18:45:44.125186 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.225943 kubelet[2333]: E1212 18:45:44.225888 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.280729 kubelet[2333]: E1212 18:45:44.280664 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:44.281181 kubelet[2333]: E1212 18:45:44.281030 2333 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Dec 12 18:45:44.326663 kubelet[2333]: E1212 18:45:44.326591 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.427465 kubelet[2333]: E1212 18:45:44.427307 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.528468 kubelet[2333]: E1212 18:45:44.528398 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.629309 kubelet[2333]: E1212 18:45:44.629251 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.729564 kubelet[2333]: E1212 18:45:44.729398 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.830425 kubelet[2333]: E1212 18:45:44.830372 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:44.931325 kubelet[2333]: E1212 18:45:44.931277 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:45.032237 kubelet[2333]: E1212 18:45:45.032173 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:45.133145 kubelet[2333]: E1212 18:45:45.133084 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:45.234157 kubelet[2333]: E1212 18:45:45.234111 2333 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:45.343538 kubelet[2333]: I1212 18:45:45.343379 2333 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:45.354308 kubelet[2333]: I1212 18:45:45.354275 2333 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:45.358286 kubelet[2333]: I1212 18:45:45.358241 2333 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:46.034020 systemd[1]: Reload requested from client PID 2623 ('systemctl') (unit session-7.scope)... Dec 12 18:45:46.034038 systemd[1]: Reloading... Dec 12 18:45:46.112009 zram_generator::config[2669]: No configuration found. Dec 12 18:45:46.234146 kubelet[2333]: I1212 18:45:46.234061 2333 apiserver.go:52] "Watching apiserver" Dec 12 18:45:46.242763 kubelet[2333]: I1212 18:45:46.242719 2333 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 12 18:45:46.347168 systemd[1]: Reloading finished in 312 ms. Dec 12 18:45:46.375429 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:46.393101 systemd[1]: kubelet.service: Deactivated successfully. Dec 12 18:45:46.393580 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:46.393667 systemd[1]: kubelet.service: Consumed 753ms CPU time, 127.4M memory peak. Dec 12 18:45:46.396307 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 12 18:45:46.621752 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 12 18:45:46.638665 (kubelet)[2712]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 12 18:45:46.682041 kubelet[2712]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 12 18:45:46.682041 kubelet[2712]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 18:45:46.682489 kubelet[2712]: I1212 18:45:46.682068 2712 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 12 18:45:46.689105 kubelet[2712]: I1212 18:45:46.689055 2712 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Dec 12 18:45:46.689105 kubelet[2712]: I1212 18:45:46.689084 2712 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 12 18:45:46.689105 kubelet[2712]: I1212 18:45:46.689111 2712 watchdog_linux.go:95] "Systemd watchdog is not enabled" Dec 12 18:45:46.689105 kubelet[2712]: I1212 18:45:46.689118 2712 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 12 18:45:46.689406 kubelet[2712]: I1212 18:45:46.689391 2712 server.go:956] "Client rotation is on, will bootstrap in background" Dec 12 18:45:46.690572 kubelet[2712]: I1212 18:45:46.690548 2712 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Dec 12 18:45:46.692978 kubelet[2712]: I1212 18:45:46.692930 2712 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 12 18:45:46.697440 kubelet[2712]: I1212 18:45:46.697400 2712 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 12 18:45:46.702520 kubelet[2712]: I1212 18:45:46.702490 2712 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Dec 12 18:45:46.702765 kubelet[2712]: I1212 18:45:46.702713 2712 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 12 18:45:46.702914 kubelet[2712]: I1212 18:45:46.702745 2712 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 12 18:45:46.702914 kubelet[2712]: I1212 18:45:46.702905 2712 topology_manager.go:138] "Creating topology manager with none policy" Dec 12 18:45:46.702914 kubelet[2712]: I1212 18:45:46.702913 2712 container_manager_linux.go:306] "Creating device plugin manager" Dec 12 18:45:46.703085 kubelet[2712]: I1212 18:45:46.702935 2712 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Dec 12 18:45:46.703844 kubelet[2712]: I1212 18:45:46.703810 2712 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:45:46.704007 kubelet[2712]: I1212 18:45:46.703981 2712 kubelet.go:475] "Attempting to sync node with API server" Dec 12 18:45:46.704007 kubelet[2712]: I1212 18:45:46.704002 2712 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 12 18:45:46.704056 kubelet[2712]: I1212 18:45:46.704022 2712 kubelet.go:387] "Adding apiserver pod source" Dec 12 18:45:46.704056 kubelet[2712]: I1212 18:45:46.704038 2712 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 12 18:45:46.707108 kubelet[2712]: I1212 18:45:46.705457 2712 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Dec 12 18:45:46.707108 kubelet[2712]: I1212 18:45:46.705983 2712 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Dec 12 18:45:46.707108 kubelet[2712]: I1212 18:45:46.706016 2712 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Dec 12 18:45:46.709747 kubelet[2712]: I1212 18:45:46.709714 2712 server.go:1262] "Started kubelet" Dec 12 18:45:46.710142 kubelet[2712]: I1212 18:45:46.710094 2712 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 12 18:45:46.710424 kubelet[2712]: I1212 18:45:46.710213 2712 server_v1.go:49] "podresources" method="list" useActivePods=true Dec 12 18:45:46.711273 kubelet[2712]: I1212 18:45:46.710538 2712 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 12 18:45:46.711273 kubelet[2712]: I1212 18:45:46.710611 2712 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Dec 12 18:45:46.711952 kubelet[2712]: I1212 18:45:46.711918 2712 server.go:310] "Adding debug handlers to kubelet server" Dec 12 18:45:46.716887 kubelet[2712]: I1212 18:45:46.716677 2712 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 12 18:45:46.717850 kubelet[2712]: I1212 18:45:46.717815 2712 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 12 18:45:46.718545 kubelet[2712]: E1212 18:45:46.718514 2712 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Dec 12 18:45:46.718592 kubelet[2712]: I1212 18:45:46.718567 2712 volume_manager.go:313] "Starting Kubelet Volume Manager" Dec 12 18:45:46.718830 kubelet[2712]: I1212 18:45:46.718803 2712 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 12 18:45:46.719305 kubelet[2712]: I1212 18:45:46.719274 2712 reconciler.go:29] "Reconciler: start to sync state" Dec 12 18:45:46.723569 kubelet[2712]: I1212 18:45:46.723425 2712 factory.go:223] Registration of the systemd container factory successfully Dec 12 18:45:46.723569 kubelet[2712]: E1212 18:45:46.723498 2712 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 12 18:45:46.723569 kubelet[2712]: I1212 18:45:46.723531 2712 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 12 18:45:46.725366 kubelet[2712]: I1212 18:45:46.725319 2712 factory.go:223] Registration of the containerd container factory successfully Dec 12 18:45:46.733127 kubelet[2712]: I1212 18:45:46.733090 2712 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Dec 12 18:45:46.734752 kubelet[2712]: I1212 18:45:46.734434 2712 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Dec 12 18:45:46.734752 kubelet[2712]: I1212 18:45:46.734457 2712 status_manager.go:244] "Starting to sync pod status with apiserver" Dec 12 18:45:46.734752 kubelet[2712]: I1212 18:45:46.734482 2712 kubelet.go:2427] "Starting kubelet main sync loop" Dec 12 18:45:46.734752 kubelet[2712]: E1212 18:45:46.734521 2712 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 12 18:45:46.765906 kubelet[2712]: I1212 18:45:46.765874 2712 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 12 18:45:46.765906 kubelet[2712]: I1212 18:45:46.765894 2712 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 12 18:45:46.765906 kubelet[2712]: I1212 18:45:46.765916 2712 state_mem.go:36] "Initialized new in-memory state store" Dec 12 18:45:46.766077 kubelet[2712]: I1212 18:45:46.766041 2712 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 12 18:45:46.766077 kubelet[2712]: I1212 18:45:46.766051 2712 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 12 18:45:46.766077 kubelet[2712]: I1212 18:45:46.766069 2712 policy_none.go:49] "None policy: Start" Dec 12 18:45:46.766077 kubelet[2712]: I1212 18:45:46.766078 2712 memory_manager.go:187] "Starting memorymanager" policy="None" Dec 12 18:45:46.766170 kubelet[2712]: I1212 18:45:46.766088 2712 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Dec 12 18:45:46.766192 kubelet[2712]: I1212 18:45:46.766172 2712 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Dec 12 18:45:46.766192 kubelet[2712]: I1212 18:45:46.766179 2712 policy_none.go:47] "Start" Dec 12 18:45:46.770980 kubelet[2712]: E1212 18:45:46.770939 2712 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Dec 12 18:45:46.771131 kubelet[2712]: I1212 18:45:46.771105 2712 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 12 18:45:46.771158 kubelet[2712]: I1212 18:45:46.771119 2712 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 12 18:45:46.771359 kubelet[2712]: I1212 18:45:46.771317 2712 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 12 18:45:46.772217 kubelet[2712]: E1212 18:45:46.772192 2712 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 12 18:45:46.835704 kubelet[2712]: I1212 18:45:46.835654 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:46.835704 kubelet[2712]: I1212 18:45:46.835684 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:46.835870 kubelet[2712]: I1212 18:45:46.835654 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:46.873237 kubelet[2712]: I1212 18:45:46.873114 2712 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Dec 12 18:45:47.020205 kubelet[2712]: I1212 18:45:47.020154 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.020205 kubelet[2712]: I1212 18:45:47.020195 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:47.020205 kubelet[2712]: I1212 18:45:47.020214 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:47.020451 kubelet[2712]: I1212 18:45:47.020228 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.020451 kubelet[2712]: I1212 18:45:47.020243 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.020451 kubelet[2712]: I1212 18:45:47.020280 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:47.020451 kubelet[2712]: I1212 18:45:47.020380 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2f29938ea9ffdae1ba75459fa38a7152-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2f29938ea9ffdae1ba75459fa38a7152\") " pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:47.020451 kubelet[2712]: I1212 18:45:47.020424 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.020582 kubelet[2712]: I1212 18:45:47.020450 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.288252 kubelet[2712]: E1212 18:45:47.288174 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.289487 kubelet[2712]: E1212 18:45:47.289269 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:47.289487 kubelet[2712]: E1212 18:45:47.289419 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:47.289876 kubelet[2712]: I1212 18:45:47.289842 2712 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Dec 12 18:45:47.290012 kubelet[2712]: I1212 18:45:47.289983 2712 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Dec 12 18:45:47.705295 kubelet[2712]: I1212 18:45:47.705123 2712 apiserver.go:52] "Watching apiserver" Dec 12 18:45:47.719825 kubelet[2712]: I1212 18:45:47.719796 2712 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 12 18:45:47.751616 kubelet[2712]: I1212 18:45:47.751557 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:47.752049 kubelet[2712]: I1212 18:45:47.751653 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:47.752049 kubelet[2712]: I1212 18:45:47.751792 2712 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:48.060384 kubelet[2712]: E1212 18:45:48.059737 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Dec 12 18:45:48.060957 kubelet[2712]: E1212 18:45:48.060130 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Dec 12 18:45:48.061307 kubelet[2712]: E1212 18:45:48.060221 2712 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Dec 12 18:45:48.068943 kubelet[2712]: I1212 18:45:48.068846 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.068827203 podStartE2EDuration="3.068827203s" podCreationTimestamp="2025-12-12 18:45:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:45:48.059724591 +0000 UTC m=+1.417337334" watchObservedRunningTime="2025-12-12 18:45:48.068827203 +0000 UTC m=+1.426439946" Dec 12 18:45:48.080734 kubelet[2712]: I1212 18:45:48.080628 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=3.08060921 podStartE2EDuration="3.08060921s" podCreationTimestamp="2025-12-12 18:45:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:45:48.080329174 +0000 UTC m=+1.437941917" watchObservedRunningTime="2025-12-12 18:45:48.08060921 +0000 UTC m=+1.438221953" Dec 12 18:45:48.080902 kubelet[2712]: I1212 18:45:48.080752 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.080747012 podStartE2EDuration="3.080747012s" podCreationTimestamp="2025-12-12 18:45:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:45:48.0696555 +0000 UTC m=+1.427268263" watchObservedRunningTime="2025-12-12 18:45:48.080747012 +0000 UTC m=+1.438359765" Dec 12 18:45:48.103862 sudo[2749]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Dec 12 18:45:48.104266 sudo[2749]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Dec 12 18:45:48.424956 sudo[2749]: pam_unix(sudo:session): session closed for user root Dec 12 18:45:49.707939 sudo[1757]: pam_unix(sudo:session): session closed for user root Dec 12 18:45:49.709457 sshd[1756]: Connection closed by 10.0.0.1 port 42864 Dec 12 18:45:49.709894 sshd-session[1753]: pam_unix(sshd:session): session closed for user core Dec 12 18:45:49.714258 systemd[1]: sshd@6-10.0.0.119:22-10.0.0.1:42864.service: Deactivated successfully. Dec 12 18:45:49.716544 systemd[1]: session-7.scope: Deactivated successfully. Dec 12 18:45:49.716753 systemd[1]: session-7.scope: Consumed 6.391s CPU time, 265.4M memory peak. Dec 12 18:45:49.718335 systemd-logind[1524]: Session 7 logged out. Waiting for processes to exit. Dec 12 18:45:49.719607 systemd-logind[1524]: Removed session 7. Dec 12 18:45:53.159135 kubelet[2712]: I1212 18:45:53.159086 2712 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 12 18:45:53.159777 containerd[1543]: time="2025-12-12T18:45:53.159631235Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 12 18:45:53.160115 kubelet[2712]: I1212 18:45:53.159946 2712 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 12 18:45:53.968175 systemd[1]: Created slice kubepods-besteffort-podb8fb40b5_c7df_4e24_b305_1753cac8b281.slice - libcontainer container kubepods-besteffort-podb8fb40b5_c7df_4e24_b305_1753cac8b281.slice. Dec 12 18:45:53.970995 systemd[1]: Created slice kubepods-burstable-pod4548a03e_d58b_4ff6_a670_4064c33ce862.slice - libcontainer container kubepods-burstable-pod4548a03e_d58b_4ff6_a670_4064c33ce862.slice. Dec 12 18:45:54.063656 kubelet[2712]: I1212 18:45:54.063555 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-lib-modules\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063690 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/b8fb40b5-c7df-4e24-b305-1753cac8b281-kube-proxy\") pod \"kube-proxy-8f2c7\" (UID: \"b8fb40b5-c7df-4e24-b305-1753cac8b281\") " pod="kube-system/kube-proxy-8f2c7" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063712 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-etc-cni-netd\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063727 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-xtables-lock\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063756 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4548a03e-d58b-4ff6-a670-4064c33ce862-clustermesh-secrets\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063773 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-config-path\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.063812 kubelet[2712]: I1212 18:45:54.063792 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-bpf-maps\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064062 kubelet[2712]: I1212 18:45:54.063810 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-net\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064062 kubelet[2712]: I1212 18:45:54.063827 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-kernel\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064062 kubelet[2712]: I1212 18:45:54.063845 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-hubble-tls\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064062 kubelet[2712]: I1212 18:45:54.063863 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b8fb40b5-c7df-4e24-b305-1753cac8b281-lib-modules\") pod \"kube-proxy-8f2c7\" (UID: \"b8fb40b5-c7df-4e24-b305-1753cac8b281\") " pod="kube-system/kube-proxy-8f2c7" Dec 12 18:45:54.064062 kubelet[2712]: I1212 18:45:54.063878 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-run\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063895 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvvq5\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-kube-api-access-dvvq5\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063912 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b8fb40b5-c7df-4e24-b305-1753cac8b281-xtables-lock\") pod \"kube-proxy-8f2c7\" (UID: \"b8fb40b5-c7df-4e24-b305-1753cac8b281\") " pod="kube-system/kube-proxy-8f2c7" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063932 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmnh\" (UniqueName: \"kubernetes.io/projected/b8fb40b5-c7df-4e24-b305-1753cac8b281-kube-api-access-flmnh\") pod \"kube-proxy-8f2c7\" (UID: \"b8fb40b5-c7df-4e24-b305-1753cac8b281\") " pod="kube-system/kube-proxy-8f2c7" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063948 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-hostproc\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063963 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-cgroup\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.064273 kubelet[2712]: I1212 18:45:54.063979 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cni-path\") pod \"cilium-r7fhx\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " pod="kube-system/cilium-r7fhx" Dec 12 18:45:54.289992 containerd[1543]: time="2025-12-12T18:45:54.289925743Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-r7fhx,Uid:4548a03e-d58b-4ff6-a670-4064c33ce862,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:54.294622 containerd[1543]: time="2025-12-12T18:45:54.294551078Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8f2c7,Uid:b8fb40b5-c7df-4e24-b305-1753cac8b281,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:54.346768 containerd[1543]: time="2025-12-12T18:45:54.346628663Z" level=info msg="connecting to shim a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:54.359737 containerd[1543]: time="2025-12-12T18:45:54.359677191Z" level=info msg="connecting to shim 5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c" address="unix:///run/containerd/s/dda0a0b0162b1f11beb2987682693252730f4184f3f3c423b34155475a1eeac7" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:54.378194 systemd[1]: Created slice kubepods-besteffort-pod3dbdbe3f_84a9_4932_9148_cd1f33377b91.slice - libcontainer container kubepods-besteffort-pod3dbdbe3f_84a9_4932_9148_cd1f33377b91.slice. Dec 12 18:45:54.405625 systemd[1]: Started cri-containerd-5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c.scope - libcontainer container 5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c. Dec 12 18:45:54.409896 systemd[1]: Started cri-containerd-a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea.scope - libcontainer container a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea. Dec 12 18:45:54.442363 containerd[1543]: time="2025-12-12T18:45:54.442297954Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8f2c7,Uid:b8fb40b5-c7df-4e24-b305-1753cac8b281,Namespace:kube-system,Attempt:0,} returns sandbox id \"5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c\"" Dec 12 18:45:54.451434 containerd[1543]: time="2025-12-12T18:45:54.451367524Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-r7fhx,Uid:4548a03e-d58b-4ff6-a670-4064c33ce862,Namespace:kube-system,Attempt:0,} returns sandbox id \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\"" Dec 12 18:45:54.452051 containerd[1543]: time="2025-12-12T18:45:54.451878016Z" level=info msg="CreateContainer within sandbox \"5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 12 18:45:54.454018 containerd[1543]: time="2025-12-12T18:45:54.453994484Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Dec 12 18:45:54.468724 kubelet[2712]: I1212 18:45:54.468545 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m44cv\" (UniqueName: \"kubernetes.io/projected/3dbdbe3f-84a9-4932-9148-cd1f33377b91-kube-api-access-m44cv\") pod \"cilium-operator-6f9c7c5859-fvrvw\" (UID: \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\") " pod="kube-system/cilium-operator-6f9c7c5859-fvrvw" Dec 12 18:45:54.468724 kubelet[2712]: I1212 18:45:54.468590 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3dbdbe3f-84a9-4932-9148-cd1f33377b91-cilium-config-path\") pod \"cilium-operator-6f9c7c5859-fvrvw\" (UID: \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\") " pod="kube-system/cilium-operator-6f9c7c5859-fvrvw" Dec 12 18:45:54.470956 containerd[1543]: time="2025-12-12T18:45:54.470906635Z" level=info msg="Container a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:45:54.482568 containerd[1543]: time="2025-12-12T18:45:54.482513958Z" level=info msg="CreateContainer within sandbox \"5654065cdd36ab5192878b2481d841d570b4931b09693027564bf8293695d96c\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8\"" Dec 12 18:45:54.483135 containerd[1543]: time="2025-12-12T18:45:54.483105998Z" level=info msg="StartContainer for \"a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8\"" Dec 12 18:45:54.484592 containerd[1543]: time="2025-12-12T18:45:54.484558188Z" level=info msg="connecting to shim a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8" address="unix:///run/containerd/s/dda0a0b0162b1f11beb2987682693252730f4184f3f3c423b34155475a1eeac7" protocol=ttrpc version=3 Dec 12 18:45:54.509520 systemd[1]: Started cri-containerd-a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8.scope - libcontainer container a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8. Dec 12 18:45:54.610477 containerd[1543]: time="2025-12-12T18:45:54.610325507Z" level=info msg="StartContainer for \"a4682ea6d7da17f302a8c8d1ed9138069fe1e4db5b52ea8efacd095d2e8e95f8\" returns successfully" Dec 12 18:45:54.688290 containerd[1543]: time="2025-12-12T18:45:54.688216804Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6f9c7c5859-fvrvw,Uid:3dbdbe3f-84a9-4932-9148-cd1f33377b91,Namespace:kube-system,Attempt:0,}" Dec 12 18:45:54.730194 containerd[1543]: time="2025-12-12T18:45:54.730123731Z" level=info msg="connecting to shim 7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c" address="unix:///run/containerd/s/45d315c0a23dc3d2817d730d2fa27c40eaf2113da8b5322bb6ccd6417e2e3c64" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:45:54.778658 systemd[1]: Started cri-containerd-7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c.scope - libcontainer container 7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c. Dec 12 18:45:54.793951 kubelet[2712]: I1212 18:45:54.793866 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8f2c7" podStartSLOduration=1.793845886 podStartE2EDuration="1.793845886s" podCreationTimestamp="2025-12-12 18:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:45:54.793124927 +0000 UTC m=+8.150737670" watchObservedRunningTime="2025-12-12 18:45:54.793845886 +0000 UTC m=+8.151458629" Dec 12 18:45:54.843381 containerd[1543]: time="2025-12-12T18:45:54.838491271Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6f9c7c5859-fvrvw,Uid:3dbdbe3f-84a9-4932-9148-cd1f33377b91,Namespace:kube-system,Attempt:0,} returns sandbox id \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\"" Dec 12 18:45:58.180777 update_engine[1533]: I20251212 18:45:58.180663 1533 update_attempter.cc:509] Updating boot flags... Dec 12 18:46:01.706281 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4215756510.mount: Deactivated successfully. Dec 12 18:46:05.197591 containerd[1543]: time="2025-12-12T18:46:05.197492474Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:46:05.198630 containerd[1543]: time="2025-12-12T18:46:05.198596017Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Dec 12 18:46:05.200072 containerd[1543]: time="2025-12-12T18:46:05.199954613Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:46:05.201781 containerd[1543]: time="2025-12-12T18:46:05.201703893Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 10.747486694s" Dec 12 18:46:05.201781 containerd[1543]: time="2025-12-12T18:46:05.201741091Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Dec 12 18:46:05.207883 containerd[1543]: time="2025-12-12T18:46:05.207848452Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Dec 12 18:46:05.219574 containerd[1543]: time="2025-12-12T18:46:05.219503248Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 12 18:46:05.233451 containerd[1543]: time="2025-12-12T18:46:05.233368046Z" level=info msg="Container 0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:05.245607 containerd[1543]: time="2025-12-12T18:46:05.245533550Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\"" Dec 12 18:46:05.246252 containerd[1543]: time="2025-12-12T18:46:05.246189920Z" level=info msg="StartContainer for \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\"" Dec 12 18:46:05.247543 containerd[1543]: time="2025-12-12T18:46:05.247496264Z" level=info msg="connecting to shim 0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" protocol=ttrpc version=3 Dec 12 18:46:05.275896 systemd[1]: Started cri-containerd-0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4.scope - libcontainer container 0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4. Dec 12 18:46:05.328472 containerd[1543]: time="2025-12-12T18:46:05.328410952Z" level=info msg="StartContainer for \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" returns successfully" Dec 12 18:46:05.340440 systemd[1]: cri-containerd-0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4.scope: Deactivated successfully. Dec 12 18:46:05.343710 containerd[1543]: time="2025-12-12T18:46:05.343663018Z" level=info msg="received container exit event container_id:\"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" id:\"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" pid:3157 exited_at:{seconds:1765565165 nanos:342992228}" Dec 12 18:46:05.376286 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4-rootfs.mount: Deactivated successfully. Dec 12 18:46:06.831203 containerd[1543]: time="2025-12-12T18:46:06.831071484Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 12 18:46:06.851886 containerd[1543]: time="2025-12-12T18:46:06.851832133Z" level=info msg="Container ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:06.855713 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3313670337.mount: Deactivated successfully. Dec 12 18:46:06.861980 containerd[1543]: time="2025-12-12T18:46:06.861910758Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\"" Dec 12 18:46:06.862704 containerd[1543]: time="2025-12-12T18:46:06.862657261Z" level=info msg="StartContainer for \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\"" Dec 12 18:46:06.864100 containerd[1543]: time="2025-12-12T18:46:06.863975958Z" level=info msg="connecting to shim ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" protocol=ttrpc version=3 Dec 12 18:46:06.893705 systemd[1]: Started cri-containerd-ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6.scope - libcontainer container ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6. Dec 12 18:46:06.939985 containerd[1543]: time="2025-12-12T18:46:06.939934042Z" level=info msg="StartContainer for \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" returns successfully" Dec 12 18:46:06.957629 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 12 18:46:06.958294 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:46:06.958591 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:46:06.961275 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 12 18:46:06.963760 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 12 18:46:06.966385 systemd[1]: cri-containerd-ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6.scope: Deactivated successfully. Dec 12 18:46:06.968159 containerd[1543]: time="2025-12-12T18:46:06.968097912Z" level=info msg="received container exit event container_id:\"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" id:\"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" pid:3201 exited_at:{seconds:1765565166 nanos:966690326}" Dec 12 18:46:06.993812 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 12 18:46:07.511232 containerd[1543]: time="2025-12-12T18:46:07.510812758Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:46:07.511878 containerd[1543]: time="2025-12-12T18:46:07.511826408Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Dec 12 18:46:07.513450 containerd[1543]: time="2025-12-12T18:46:07.513415733Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 12 18:46:07.514806 containerd[1543]: time="2025-12-12T18:46:07.514764519Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.306690777s" Dec 12 18:46:07.514806 containerd[1543]: time="2025-12-12T18:46:07.514793961Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Dec 12 18:46:07.520426 containerd[1543]: time="2025-12-12T18:46:07.520373847Z" level=info msg="CreateContainer within sandbox \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Dec 12 18:46:07.528500 containerd[1543]: time="2025-12-12T18:46:07.528432164Z" level=info msg="Container 91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:07.541164 containerd[1543]: time="2025-12-12T18:46:07.541101943Z" level=info msg="CreateContainer within sandbox \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\"" Dec 12 18:46:07.542539 containerd[1543]: time="2025-12-12T18:46:07.542115753Z" level=info msg="StartContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\"" Dec 12 18:46:07.544441 containerd[1543]: time="2025-12-12T18:46:07.544410937Z" level=info msg="connecting to shim 91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655" address="unix:///run/containerd/s/45d315c0a23dc3d2817d730d2fa27c40eaf2113da8b5322bb6ccd6417e2e3c64" protocol=ttrpc version=3 Dec 12 18:46:07.570808 systemd[1]: Started cri-containerd-91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655.scope - libcontainer container 91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655. Dec 12 18:46:07.618065 containerd[1543]: time="2025-12-12T18:46:07.617999743Z" level=info msg="StartContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" returns successfully" Dec 12 18:46:07.833924 containerd[1543]: time="2025-12-12T18:46:07.833776124Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 12 18:46:07.853672 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6-rootfs.mount: Deactivated successfully. Dec 12 18:46:07.854450 containerd[1543]: time="2025-12-12T18:46:07.854004054Z" level=info msg="Container 4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:07.869242 containerd[1543]: time="2025-12-12T18:46:07.869174491Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\"" Dec 12 18:46:07.870390 containerd[1543]: time="2025-12-12T18:46:07.869778916Z" level=info msg="StartContainer for \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\"" Dec 12 18:46:07.870453 kubelet[2712]: I1212 18:46:07.869907 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6f9c7c5859-fvrvw" podStartSLOduration=1.199949738 podStartE2EDuration="13.869888477s" podCreationTimestamp="2025-12-12 18:45:54 +0000 UTC" firstStartedPulling="2025-12-12 18:45:54.845558847 +0000 UTC m=+8.203171590" lastFinishedPulling="2025-12-12 18:46:07.515497586 +0000 UTC m=+20.873110329" observedRunningTime="2025-12-12 18:46:07.86866262 +0000 UTC m=+21.226275373" watchObservedRunningTime="2025-12-12 18:46:07.869888477 +0000 UTC m=+21.227501240" Dec 12 18:46:07.871990 containerd[1543]: time="2025-12-12T18:46:07.871947172Z" level=info msg="connecting to shim 4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" protocol=ttrpc version=3 Dec 12 18:46:07.899922 systemd[1]: Started cri-containerd-4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0.scope - libcontainer container 4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0. Dec 12 18:46:07.995633 containerd[1543]: time="2025-12-12T18:46:07.995562809Z" level=info msg="StartContainer for \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" returns successfully" Dec 12 18:46:08.000993 systemd[1]: cri-containerd-4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0.scope: Deactivated successfully. Dec 12 18:46:08.004335 containerd[1543]: time="2025-12-12T18:46:08.004277804Z" level=info msg="received container exit event container_id:\"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" id:\"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" pid:3302 exited_at:{seconds:1765565168 nanos:3827498}" Dec 12 18:46:08.068249 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0-rootfs.mount: Deactivated successfully. Dec 12 18:46:08.851198 containerd[1543]: time="2025-12-12T18:46:08.851143976Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 12 18:46:08.864479 containerd[1543]: time="2025-12-12T18:46:08.864442232Z" level=info msg="Container fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:08.869910 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4053299581.mount: Deactivated successfully. Dec 12 18:46:08.876438 containerd[1543]: time="2025-12-12T18:46:08.876387677Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\"" Dec 12 18:46:08.876930 containerd[1543]: time="2025-12-12T18:46:08.876894291Z" level=info msg="StartContainer for \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\"" Dec 12 18:46:08.878010 containerd[1543]: time="2025-12-12T18:46:08.877850831Z" level=info msg="connecting to shim fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" protocol=ttrpc version=3 Dec 12 18:46:08.907664 systemd[1]: Started cri-containerd-fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58.scope - libcontainer container fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58. Dec 12 18:46:08.939164 systemd[1]: cri-containerd-fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58.scope: Deactivated successfully. Dec 12 18:46:08.942254 containerd[1543]: time="2025-12-12T18:46:08.942218362Z" level=info msg="received container exit event container_id:\"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" id:\"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" pid:3340 exited_at:{seconds:1765565168 nanos:939481483}" Dec 12 18:46:08.950169 containerd[1543]: time="2025-12-12T18:46:08.950120432Z" level=info msg="StartContainer for \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" returns successfully" Dec 12 18:46:08.963860 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58-rootfs.mount: Deactivated successfully. Dec 12 18:46:09.855484 containerd[1543]: time="2025-12-12T18:46:09.855431218Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 12 18:46:09.878379 containerd[1543]: time="2025-12-12T18:46:09.876626312Z" level=info msg="Container 19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:09.897598 containerd[1543]: time="2025-12-12T18:46:09.897490593Z" level=info msg="CreateContainer within sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\"" Dec 12 18:46:09.898488 containerd[1543]: time="2025-12-12T18:46:09.898410888Z" level=info msg="StartContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\"" Dec 12 18:46:09.899496 containerd[1543]: time="2025-12-12T18:46:09.899462357Z" level=info msg="connecting to shim 19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71" address="unix:///run/containerd/s/75228e65dfc349a57161ee5b779a7f5c1022342b08eebd8aba9f0346f747a4b4" protocol=ttrpc version=3 Dec 12 18:46:09.928703 systemd[1]: Started cri-containerd-19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71.scope - libcontainer container 19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71. Dec 12 18:46:09.984615 containerd[1543]: time="2025-12-12T18:46:09.984571618Z" level=info msg="StartContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" returns successfully" Dec 12 18:46:10.121680 kubelet[2712]: I1212 18:46:10.121556 2712 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Dec 12 18:46:10.189701 systemd[1]: Created slice kubepods-burstable-pode7b01855_8751_4669_ab1a_27bd7e844c87.slice - libcontainer container kubepods-burstable-pode7b01855_8751_4669_ab1a_27bd7e844c87.slice. Dec 12 18:46:10.199766 systemd[1]: Created slice kubepods-burstable-podd5ee204c_76a5_420b_9a30_e2349e43a98b.slice - libcontainer container kubepods-burstable-podd5ee204c_76a5_420b_9a30_e2349e43a98b.slice. Dec 12 18:46:10.282968 kubelet[2712]: I1212 18:46:10.282907 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7b01855-8751-4669-ab1a-27bd7e844c87-config-volume\") pod \"coredns-66bc5c9577-nmp52\" (UID: \"e7b01855-8751-4669-ab1a-27bd7e844c87\") " pod="kube-system/coredns-66bc5c9577-nmp52" Dec 12 18:46:10.282968 kubelet[2712]: I1212 18:46:10.282956 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2jbp\" (UniqueName: \"kubernetes.io/projected/d5ee204c-76a5-420b-9a30-e2349e43a98b-kube-api-access-w2jbp\") pod \"coredns-66bc5c9577-wfx25\" (UID: \"d5ee204c-76a5-420b-9a30-e2349e43a98b\") " pod="kube-system/coredns-66bc5c9577-wfx25" Dec 12 18:46:10.282968 kubelet[2712]: I1212 18:46:10.282972 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5ee204c-76a5-420b-9a30-e2349e43a98b-config-volume\") pod \"coredns-66bc5c9577-wfx25\" (UID: \"d5ee204c-76a5-420b-9a30-e2349e43a98b\") " pod="kube-system/coredns-66bc5c9577-wfx25" Dec 12 18:46:10.283187 kubelet[2712]: I1212 18:46:10.283021 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fp6s\" (UniqueName: \"kubernetes.io/projected/e7b01855-8751-4669-ab1a-27bd7e844c87-kube-api-access-2fp6s\") pod \"coredns-66bc5c9577-nmp52\" (UID: \"e7b01855-8751-4669-ab1a-27bd7e844c87\") " pod="kube-system/coredns-66bc5c9577-nmp52" Dec 12 18:46:10.498992 containerd[1543]: time="2025-12-12T18:46:10.498743861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-nmp52,Uid:e7b01855-8751-4669-ab1a-27bd7e844c87,Namespace:kube-system,Attempt:0,}" Dec 12 18:46:10.511251 containerd[1543]: time="2025-12-12T18:46:10.510917190Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-wfx25,Uid:d5ee204c-76a5-420b-9a30-e2349e43a98b,Namespace:kube-system,Attempt:0,}" Dec 12 18:46:10.890521 kubelet[2712]: I1212 18:46:10.890458 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-r7fhx" podStartSLOduration=7.135761736 podStartE2EDuration="17.890443571s" podCreationTimestamp="2025-12-12 18:45:53 +0000 UTC" firstStartedPulling="2025-12-12 18:45:54.452935914 +0000 UTC m=+7.810548647" lastFinishedPulling="2025-12-12 18:46:05.207617729 +0000 UTC m=+18.565230482" observedRunningTime="2025-12-12 18:46:10.88984774 +0000 UTC m=+24.247460483" watchObservedRunningTime="2025-12-12 18:46:10.890443571 +0000 UTC m=+24.248056314" Dec 12 18:46:12.296047 systemd-networkd[1432]: cilium_host: Link UP Dec 12 18:46:12.296454 systemd-networkd[1432]: cilium_net: Link UP Dec 12 18:46:12.296688 systemd-networkd[1432]: cilium_net: Gained carrier Dec 12 18:46:12.296924 systemd-networkd[1432]: cilium_host: Gained carrier Dec 12 18:46:12.432873 systemd-networkd[1432]: cilium_vxlan: Link UP Dec 12 18:46:12.432889 systemd-networkd[1432]: cilium_vxlan: Gained carrier Dec 12 18:46:12.682436 kernel: NET: Registered PF_ALG protocol family Dec 12 18:46:12.982598 systemd-networkd[1432]: cilium_host: Gained IPv6LL Dec 12 18:46:13.239132 systemd-networkd[1432]: cilium_net: Gained IPv6LL Dec 12 18:46:13.624166 systemd-networkd[1432]: cilium_vxlan: Gained IPv6LL Dec 12 18:46:13.736483 systemd-networkd[1432]: lxc_health: Link UP Dec 12 18:46:13.737375 systemd-networkd[1432]: lxc_health: Gained carrier Dec 12 18:46:14.040946 systemd-networkd[1432]: lxcebd8b8841875: Link UP Dec 12 18:46:14.049632 kernel: eth0: renamed from tmpc66fb Dec 12 18:46:14.066386 kernel: eth0: renamed from tmpe53eb Dec 12 18:46:14.070112 systemd-networkd[1432]: lxc2c70cdffffe4: Link UP Dec 12 18:46:14.070532 systemd-networkd[1432]: lxcebd8b8841875: Gained carrier Dec 12 18:46:14.070787 systemd-networkd[1432]: lxc2c70cdffffe4: Gained carrier Dec 12 18:46:15.350519 systemd-networkd[1432]: lxc_health: Gained IPv6LL Dec 12 18:46:15.606591 systemd-networkd[1432]: lxc2c70cdffffe4: Gained IPv6LL Dec 12 18:46:15.670587 systemd-networkd[1432]: lxcebd8b8841875: Gained IPv6LL Dec 12 18:46:16.301235 systemd[1]: Started sshd@7-10.0.0.119:22-10.0.0.1:47120.service - OpenSSH per-connection server daemon (10.0.0.1:47120). Dec 12 18:46:16.398715 sshd[3880]: Accepted publickey for core from 10.0.0.1 port 47120 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:16.400879 sshd-session[3880]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:16.406157 systemd-logind[1524]: New session 8 of user core. Dec 12 18:46:16.419570 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 12 18:46:16.570725 sshd[3883]: Connection closed by 10.0.0.1 port 47120 Dec 12 18:46:16.572606 sshd-session[3880]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:16.577960 systemd[1]: sshd@7-10.0.0.119:22-10.0.0.1:47120.service: Deactivated successfully. Dec 12 18:46:16.580982 systemd[1]: session-8.scope: Deactivated successfully. Dec 12 18:46:16.582191 systemd-logind[1524]: Session 8 logged out. Waiting for processes to exit. Dec 12 18:46:16.583454 systemd-logind[1524]: Removed session 8. Dec 12 18:46:18.618178 containerd[1543]: time="2025-12-12T18:46:18.618089442Z" level=info msg="connecting to shim c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329" address="unix:///run/containerd/s/1d9ad707529d9ab82ae200869795c8bf03ee3defaf7e957c6454586f4fc09b31" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:46:18.676857 systemd[1]: Started cri-containerd-c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329.scope - libcontainer container c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329. Dec 12 18:46:18.682440 containerd[1543]: time="2025-12-12T18:46:18.681374210Z" level=info msg="connecting to shim e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786" address="unix:///run/containerd/s/532d081d4f0a8c883999b52498815e4046d0c388abf1f5bdd4ba159c06261b81" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:46:18.711342 systemd-resolved[1390]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 12 18:46:18.745762 systemd[1]: Started cri-containerd-e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786.scope - libcontainer container e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786. Dec 12 18:46:18.776667 containerd[1543]: time="2025-12-12T18:46:18.774657413Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-nmp52,Uid:e7b01855-8751-4669-ab1a-27bd7e844c87,Namespace:kube-system,Attempt:0,} returns sandbox id \"c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329\"" Dec 12 18:46:18.782441 systemd-resolved[1390]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 12 18:46:18.788450 containerd[1543]: time="2025-12-12T18:46:18.788396165Z" level=info msg="CreateContainer within sandbox \"c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 12 18:46:18.817877 containerd[1543]: time="2025-12-12T18:46:18.817809230Z" level=info msg="Container 9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:18.837915 containerd[1543]: time="2025-12-12T18:46:18.837810965Z" level=info msg="CreateContainer within sandbox \"c66fbc2b0fdf1cd48f01804364ea80f97d929abbf09b6285df32b970a0450329\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1\"" Dec 12 18:46:18.843729 containerd[1543]: time="2025-12-12T18:46:18.843665659Z" level=info msg="StartContainer for \"9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1\"" Dec 12 18:46:18.853954 containerd[1543]: time="2025-12-12T18:46:18.853017118Z" level=info msg="connecting to shim 9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1" address="unix:///run/containerd/s/1d9ad707529d9ab82ae200869795c8bf03ee3defaf7e957c6454586f4fc09b31" protocol=ttrpc version=3 Dec 12 18:46:18.907971 systemd[1]: Started cri-containerd-9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1.scope - libcontainer container 9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1. Dec 12 18:46:18.921374 containerd[1543]: time="2025-12-12T18:46:18.921282014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-wfx25,Uid:d5ee204c-76a5-420b-9a30-e2349e43a98b,Namespace:kube-system,Attempt:0,} returns sandbox id \"e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786\"" Dec 12 18:46:18.936698 containerd[1543]: time="2025-12-12T18:46:18.936606337Z" level=info msg="CreateContainer within sandbox \"e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 12 18:46:18.989598 containerd[1543]: time="2025-12-12T18:46:18.989527380Z" level=info msg="Container d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:46:19.022741 containerd[1543]: time="2025-12-12T18:46:19.022581953Z" level=info msg="CreateContainer within sandbox \"e53eb53586e200cb8ce8560dd49e272a949f6b8412bea8e44f1e8cfcaa6f0786\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18\"" Dec 12 18:46:19.022920 containerd[1543]: time="2025-12-12T18:46:19.022782048Z" level=info msg="StartContainer for \"9af34e944edd5287590fa5dd655ab20c1796b8f831ba65a59bbb9323d8b9d9c1\" returns successfully" Dec 12 18:46:19.026044 containerd[1543]: time="2025-12-12T18:46:19.025774683Z" level=info msg="StartContainer for \"d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18\"" Dec 12 18:46:19.029866 containerd[1543]: time="2025-12-12T18:46:19.029499293Z" level=info msg="connecting to shim d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18" address="unix:///run/containerd/s/532d081d4f0a8c883999b52498815e4046d0c388abf1f5bdd4ba159c06261b81" protocol=ttrpc version=3 Dec 12 18:46:19.083046 systemd[1]: Started cri-containerd-d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18.scope - libcontainer container d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18. Dec 12 18:46:19.173803 containerd[1543]: time="2025-12-12T18:46:19.173563274Z" level=info msg="StartContainer for \"d4dbde15eec5f227ce94d57d5b07e4b9093eb870cca482586ad7bea1f4f86a18\" returns successfully" Dec 12 18:46:19.605800 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2850348205.mount: Deactivated successfully. Dec 12 18:46:20.070011 kubelet[2712]: I1212 18:46:20.069941 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-wfx25" podStartSLOduration=26.069042411 podStartE2EDuration="26.069042411s" podCreationTimestamp="2025-12-12 18:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:46:20.008564353 +0000 UTC m=+33.366177096" watchObservedRunningTime="2025-12-12 18:46:20.069042411 +0000 UTC m=+33.426655144" Dec 12 18:46:20.070639 kubelet[2712]: I1212 18:46:20.070292 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-nmp52" podStartSLOduration=26.070283992 podStartE2EDuration="26.070283992s" podCreationTimestamp="2025-12-12 18:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:46:20.070132916 +0000 UTC m=+33.427745659" watchObservedRunningTime="2025-12-12 18:46:20.070283992 +0000 UTC m=+33.427896735" Dec 12 18:46:21.637305 systemd[1]: Started sshd@8-10.0.0.119:22-10.0.0.1:35428.service - OpenSSH per-connection server daemon (10.0.0.1:35428). Dec 12 18:46:21.818105 sshd[4078]: Accepted publickey for core from 10.0.0.1 port 35428 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:21.829832 sshd-session[4078]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:21.850522 systemd-logind[1524]: New session 9 of user core. Dec 12 18:46:21.862703 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 12 18:46:22.153405 sshd[4081]: Connection closed by 10.0.0.1 port 35428 Dec 12 18:46:22.154314 sshd-session[4078]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:22.166774 systemd[1]: sshd@8-10.0.0.119:22-10.0.0.1:35428.service: Deactivated successfully. Dec 12 18:46:22.178337 systemd[1]: session-9.scope: Deactivated successfully. Dec 12 18:46:22.184501 systemd-logind[1524]: Session 9 logged out. Waiting for processes to exit. Dec 12 18:46:22.196950 systemd-logind[1524]: Removed session 9. Dec 12 18:46:27.173360 systemd[1]: Started sshd@9-10.0.0.119:22-10.0.0.1:35456.service - OpenSSH per-connection server daemon (10.0.0.1:35456). Dec 12 18:46:27.253644 sshd[4099]: Accepted publickey for core from 10.0.0.1 port 35456 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:27.255604 sshd-session[4099]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:27.261364 systemd-logind[1524]: New session 10 of user core. Dec 12 18:46:27.269561 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 12 18:46:27.450601 sshd[4102]: Connection closed by 10.0.0.1 port 35456 Dec 12 18:46:27.450878 sshd-session[4099]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:27.454968 systemd[1]: sshd@9-10.0.0.119:22-10.0.0.1:35456.service: Deactivated successfully. Dec 12 18:46:27.457120 systemd[1]: session-10.scope: Deactivated successfully. Dec 12 18:46:27.458044 systemd-logind[1524]: Session 10 logged out. Waiting for processes to exit. Dec 12 18:46:27.459531 systemd-logind[1524]: Removed session 10. Dec 12 18:46:32.470408 systemd[1]: Started sshd@10-10.0.0.119:22-10.0.0.1:36320.service - OpenSSH per-connection server daemon (10.0.0.1:36320). Dec 12 18:46:32.537289 sshd[4117]: Accepted publickey for core from 10.0.0.1 port 36320 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:32.539189 sshd-session[4117]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:32.543671 systemd-logind[1524]: New session 11 of user core. Dec 12 18:46:32.556682 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 12 18:46:32.677266 sshd[4120]: Connection closed by 10.0.0.1 port 36320 Dec 12 18:46:32.677670 sshd-session[4117]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:32.687206 systemd[1]: sshd@10-10.0.0.119:22-10.0.0.1:36320.service: Deactivated successfully. Dec 12 18:46:32.689626 systemd[1]: session-11.scope: Deactivated successfully. Dec 12 18:46:32.690518 systemd-logind[1524]: Session 11 logged out. Waiting for processes to exit. Dec 12 18:46:32.693669 systemd[1]: Started sshd@11-10.0.0.119:22-10.0.0.1:36332.service - OpenSSH per-connection server daemon (10.0.0.1:36332). Dec 12 18:46:32.694910 systemd-logind[1524]: Removed session 11. Dec 12 18:46:32.755321 sshd[4134]: Accepted publickey for core from 10.0.0.1 port 36332 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:32.756938 sshd-session[4134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:32.761267 systemd-logind[1524]: New session 12 of user core. Dec 12 18:46:32.771548 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 12 18:46:32.949635 sshd[4138]: Connection closed by 10.0.0.1 port 36332 Dec 12 18:46:32.951577 sshd-session[4134]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:32.966537 systemd[1]: sshd@11-10.0.0.119:22-10.0.0.1:36332.service: Deactivated successfully. Dec 12 18:46:32.968945 systemd[1]: session-12.scope: Deactivated successfully. Dec 12 18:46:32.970232 systemd-logind[1524]: Session 12 logged out. Waiting for processes to exit. Dec 12 18:46:32.976625 systemd[1]: Started sshd@12-10.0.0.119:22-10.0.0.1:36336.service - OpenSSH per-connection server daemon (10.0.0.1:36336). Dec 12 18:46:32.979561 systemd-logind[1524]: Removed session 12. Dec 12 18:46:33.034074 sshd[4150]: Accepted publickey for core from 10.0.0.1 port 36336 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:33.035726 sshd-session[4150]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:33.040292 systemd-logind[1524]: New session 13 of user core. Dec 12 18:46:33.052491 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 12 18:46:33.171957 sshd[4153]: Connection closed by 10.0.0.1 port 36336 Dec 12 18:46:33.172306 sshd-session[4150]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:33.176468 systemd[1]: sshd@12-10.0.0.119:22-10.0.0.1:36336.service: Deactivated successfully. Dec 12 18:46:33.178622 systemd[1]: session-13.scope: Deactivated successfully. Dec 12 18:46:33.180571 systemd-logind[1524]: Session 13 logged out. Waiting for processes to exit. Dec 12 18:46:33.182455 systemd-logind[1524]: Removed session 13. Dec 12 18:46:38.211522 systemd[1]: Started sshd@13-10.0.0.119:22-10.0.0.1:36348.service - OpenSSH per-connection server daemon (10.0.0.1:36348). Dec 12 18:46:38.356249 sshd[4166]: Accepted publickey for core from 10.0.0.1 port 36348 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:38.360057 sshd-session[4166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:38.374498 systemd-logind[1524]: New session 14 of user core. Dec 12 18:46:38.382803 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 12 18:46:38.617306 sshd[4169]: Connection closed by 10.0.0.1 port 36348 Dec 12 18:46:38.615492 sshd-session[4166]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:38.623199 systemd[1]: sshd@13-10.0.0.119:22-10.0.0.1:36348.service: Deactivated successfully. Dec 12 18:46:38.630704 systemd[1]: session-14.scope: Deactivated successfully. Dec 12 18:46:38.632312 systemd-logind[1524]: Session 14 logged out. Waiting for processes to exit. Dec 12 18:46:38.637418 systemd-logind[1524]: Removed session 14. Dec 12 18:46:43.629281 systemd[1]: Started sshd@14-10.0.0.119:22-10.0.0.1:42784.service - OpenSSH per-connection server daemon (10.0.0.1:42784). Dec 12 18:46:43.687034 sshd[4182]: Accepted publickey for core from 10.0.0.1 port 42784 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:43.688910 sshd-session[4182]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:43.694133 systemd-logind[1524]: New session 15 of user core. Dec 12 18:46:43.705624 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 12 18:46:43.825479 sshd[4185]: Connection closed by 10.0.0.1 port 42784 Dec 12 18:46:43.827658 sshd-session[4182]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:43.831646 systemd[1]: sshd@14-10.0.0.119:22-10.0.0.1:42784.service: Deactivated successfully. Dec 12 18:46:43.833944 systemd[1]: session-15.scope: Deactivated successfully. Dec 12 18:46:43.834876 systemd-logind[1524]: Session 15 logged out. Waiting for processes to exit. Dec 12 18:46:43.836246 systemd-logind[1524]: Removed session 15. Dec 12 18:46:48.847851 systemd[1]: Started sshd@15-10.0.0.119:22-10.0.0.1:42934.service - OpenSSH per-connection server daemon (10.0.0.1:42934). Dec 12 18:46:48.903277 sshd[4200]: Accepted publickey for core from 10.0.0.1 port 42934 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:48.905421 sshd-session[4200]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:48.910907 systemd-logind[1524]: New session 16 of user core. Dec 12 18:46:48.920558 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 12 18:46:49.045484 sshd[4203]: Connection closed by 10.0.0.1 port 42934 Dec 12 18:46:49.045938 sshd-session[4200]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:49.059576 systemd[1]: sshd@15-10.0.0.119:22-10.0.0.1:42934.service: Deactivated successfully. Dec 12 18:46:49.061561 systemd[1]: session-16.scope: Deactivated successfully. Dec 12 18:46:49.062471 systemd-logind[1524]: Session 16 logged out. Waiting for processes to exit. Dec 12 18:46:49.065052 systemd[1]: Started sshd@16-10.0.0.119:22-10.0.0.1:42964.service - OpenSSH per-connection server daemon (10.0.0.1:42964). Dec 12 18:46:49.065894 systemd-logind[1524]: Removed session 16. Dec 12 18:46:49.121678 sshd[4216]: Accepted publickey for core from 10.0.0.1 port 42964 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:49.123417 sshd-session[4216]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:49.128265 systemd-logind[1524]: New session 17 of user core. Dec 12 18:46:49.142706 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 12 18:46:49.386996 sshd[4219]: Connection closed by 10.0.0.1 port 42964 Dec 12 18:46:49.387937 sshd-session[4216]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:49.405413 systemd[1]: sshd@16-10.0.0.119:22-10.0.0.1:42964.service: Deactivated successfully. Dec 12 18:46:49.408114 systemd[1]: session-17.scope: Deactivated successfully. Dec 12 18:46:49.409149 systemd-logind[1524]: Session 17 logged out. Waiting for processes to exit. Dec 12 18:46:49.412949 systemd[1]: Started sshd@17-10.0.0.119:22-10.0.0.1:42966.service - OpenSSH per-connection server daemon (10.0.0.1:42966). Dec 12 18:46:49.413692 systemd-logind[1524]: Removed session 17. Dec 12 18:46:49.477381 sshd[4231]: Accepted publickey for core from 10.0.0.1 port 42966 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:49.479462 sshd-session[4231]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:49.484672 systemd-logind[1524]: New session 18 of user core. Dec 12 18:46:49.494575 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 12 18:46:50.483292 sshd[4234]: Connection closed by 10.0.0.1 port 42966 Dec 12 18:46:50.483803 sshd-session[4231]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:50.497472 systemd[1]: sshd@17-10.0.0.119:22-10.0.0.1:42966.service: Deactivated successfully. Dec 12 18:46:50.500101 systemd[1]: session-18.scope: Deactivated successfully. Dec 12 18:46:50.501458 systemd-logind[1524]: Session 18 logged out. Waiting for processes to exit. Dec 12 18:46:50.504918 systemd[1]: Started sshd@18-10.0.0.119:22-10.0.0.1:56234.service - OpenSSH per-connection server daemon (10.0.0.1:56234). Dec 12 18:46:50.505756 systemd-logind[1524]: Removed session 18. Dec 12 18:46:50.564777 sshd[4250]: Accepted publickey for core from 10.0.0.1 port 56234 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:50.566884 sshd-session[4250]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:50.572735 systemd-logind[1524]: New session 19 of user core. Dec 12 18:46:50.582505 systemd[1]: Started session-19.scope - Session 19 of User core. Dec 12 18:46:51.041458 sshd[4253]: Connection closed by 10.0.0.1 port 56234 Dec 12 18:46:51.040885 sshd-session[4250]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:51.055592 systemd[1]: sshd@18-10.0.0.119:22-10.0.0.1:56234.service: Deactivated successfully. Dec 12 18:46:51.058056 systemd[1]: session-19.scope: Deactivated successfully. Dec 12 18:46:51.059243 systemd-logind[1524]: Session 19 logged out. Waiting for processes to exit. Dec 12 18:46:51.062856 systemd[1]: Started sshd@19-10.0.0.119:22-10.0.0.1:56248.service - OpenSSH per-connection server daemon (10.0.0.1:56248). Dec 12 18:46:51.063664 systemd-logind[1524]: Removed session 19. Dec 12 18:46:51.117530 sshd[4265]: Accepted publickey for core from 10.0.0.1 port 56248 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:51.118923 sshd-session[4265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:51.123233 systemd-logind[1524]: New session 20 of user core. Dec 12 18:46:51.131459 systemd[1]: Started session-20.scope - Session 20 of User core. Dec 12 18:46:51.243116 sshd[4268]: Connection closed by 10.0.0.1 port 56248 Dec 12 18:46:51.243488 sshd-session[4265]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:51.247662 systemd[1]: sshd@19-10.0.0.119:22-10.0.0.1:56248.service: Deactivated successfully. Dec 12 18:46:51.249629 systemd[1]: session-20.scope: Deactivated successfully. Dec 12 18:46:51.250542 systemd-logind[1524]: Session 20 logged out. Waiting for processes to exit. Dec 12 18:46:51.251774 systemd-logind[1524]: Removed session 20. Dec 12 18:46:56.261938 systemd[1]: Started sshd@20-10.0.0.119:22-10.0.0.1:56272.service - OpenSSH per-connection server daemon (10.0.0.1:56272). Dec 12 18:46:56.323687 sshd[4285]: Accepted publickey for core from 10.0.0.1 port 56272 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:46:56.325937 sshd-session[4285]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:46:56.331893 systemd-logind[1524]: New session 21 of user core. Dec 12 18:46:56.341614 systemd[1]: Started session-21.scope - Session 21 of User core. Dec 12 18:46:56.454499 sshd[4288]: Connection closed by 10.0.0.1 port 56272 Dec 12 18:46:56.454861 sshd-session[4285]: pam_unix(sshd:session): session closed for user core Dec 12 18:46:56.459491 systemd[1]: sshd@20-10.0.0.119:22-10.0.0.1:56272.service: Deactivated successfully. Dec 12 18:46:56.461690 systemd[1]: session-21.scope: Deactivated successfully. Dec 12 18:46:56.462649 systemd-logind[1524]: Session 21 logged out. Waiting for processes to exit. Dec 12 18:46:56.463969 systemd-logind[1524]: Removed session 21. Dec 12 18:47:01.471674 systemd[1]: Started sshd@21-10.0.0.119:22-10.0.0.1:43626.service - OpenSSH per-connection server daemon (10.0.0.1:43626). Dec 12 18:47:01.543811 sshd[4303]: Accepted publickey for core from 10.0.0.1 port 43626 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:01.545416 sshd-session[4303]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:01.549818 systemd-logind[1524]: New session 22 of user core. Dec 12 18:47:01.563506 systemd[1]: Started session-22.scope - Session 22 of User core. Dec 12 18:47:01.676678 sshd[4306]: Connection closed by 10.0.0.1 port 43626 Dec 12 18:47:01.677046 sshd-session[4303]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:01.681490 systemd[1]: sshd@21-10.0.0.119:22-10.0.0.1:43626.service: Deactivated successfully. Dec 12 18:47:01.683370 systemd[1]: session-22.scope: Deactivated successfully. Dec 12 18:47:01.684237 systemd-logind[1524]: Session 22 logged out. Waiting for processes to exit. Dec 12 18:47:01.685686 systemd-logind[1524]: Removed session 22. Dec 12 18:47:06.694143 systemd[1]: Started sshd@22-10.0.0.119:22-10.0.0.1:43638.service - OpenSSH per-connection server daemon (10.0.0.1:43638). Dec 12 18:47:06.768109 sshd[4319]: Accepted publickey for core from 10.0.0.1 port 43638 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:06.768850 sshd-session[4319]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:06.780819 systemd-logind[1524]: New session 23 of user core. Dec 12 18:47:06.791738 systemd[1]: Started session-23.scope - Session 23 of User core. Dec 12 18:47:06.923976 sshd[4322]: Connection closed by 10.0.0.1 port 43638 Dec 12 18:47:06.924496 sshd-session[4319]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:06.933062 systemd[1]: sshd@22-10.0.0.119:22-10.0.0.1:43638.service: Deactivated successfully. Dec 12 18:47:06.935123 systemd[1]: session-23.scope: Deactivated successfully. Dec 12 18:47:06.936033 systemd-logind[1524]: Session 23 logged out. Waiting for processes to exit. Dec 12 18:47:06.939117 systemd[1]: Started sshd@23-10.0.0.119:22-10.0.0.1:43642.service - OpenSSH per-connection server daemon (10.0.0.1:43642). Dec 12 18:47:06.939947 systemd-logind[1524]: Removed session 23. Dec 12 18:47:07.002545 sshd[4335]: Accepted publickey for core from 10.0.0.1 port 43642 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:07.004206 sshd-session[4335]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:07.008404 systemd-logind[1524]: New session 24 of user core. Dec 12 18:47:07.020515 systemd[1]: Started session-24.scope - Session 24 of User core. Dec 12 18:47:09.013651 containerd[1543]: time="2025-12-12T18:47:09.013361679Z" level=info msg="StopContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" with timeout 30 (s)" Dec 12 18:47:09.037237 containerd[1543]: time="2025-12-12T18:47:09.037096321Z" level=info msg="Stop container \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" with signal terminated" Dec 12 18:47:09.079871 systemd[1]: cri-containerd-91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655.scope: Deactivated successfully. Dec 12 18:47:09.083975 containerd[1543]: time="2025-12-12T18:47:09.083482871Z" level=info msg="received container exit event container_id:\"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" id:\"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" pid:3268 exited_at:{seconds:1765565229 nanos:82805809}" Dec 12 18:47:09.101826 containerd[1543]: time="2025-12-12T18:47:09.101645111Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 12 18:47:09.107730 containerd[1543]: time="2025-12-12T18:47:09.107522166Z" level=info msg="StopContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" with timeout 2 (s)" Dec 12 18:47:09.107998 containerd[1543]: time="2025-12-12T18:47:09.107971841Z" level=info msg="Stop container \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" with signal terminated" Dec 12 18:47:09.126449 systemd-networkd[1432]: lxc_health: Link DOWN Dec 12 18:47:09.126461 systemd-networkd[1432]: lxc_health: Lost carrier Dec 12 18:47:09.143498 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655-rootfs.mount: Deactivated successfully. Dec 12 18:47:09.160277 systemd[1]: cri-containerd-19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71.scope: Deactivated successfully. Dec 12 18:47:09.161414 systemd[1]: cri-containerd-19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71.scope: Consumed 8.600s CPU time, 125.7M memory peak, 212K read from disk, 13.3M written to disk. Dec 12 18:47:09.164538 containerd[1543]: time="2025-12-12T18:47:09.160999771Z" level=info msg="received container exit event container_id:\"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" id:\"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" pid:3376 exited_at:{seconds:1765565229 nanos:160724624}" Dec 12 18:47:09.191392 containerd[1543]: time="2025-12-12T18:47:09.190671861Z" level=info msg="StopContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" returns successfully" Dec 12 18:47:09.196653 containerd[1543]: time="2025-12-12T18:47:09.195669092Z" level=info msg="StopPodSandbox for \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\"" Dec 12 18:47:09.197486 containerd[1543]: time="2025-12-12T18:47:09.197413631Z" level=info msg="Container to stop \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.213195 systemd[1]: cri-containerd-7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c.scope: Deactivated successfully. Dec 12 18:47:09.222471 containerd[1543]: time="2025-12-12T18:47:09.218698439Z" level=info msg="received sandbox exit event container_id:\"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" id:\"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" exit_status:137 exited_at:{seconds:1765565229 nanos:216183603}" monitor_name=podsandbox Dec 12 18:47:09.228786 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71-rootfs.mount: Deactivated successfully. Dec 12 18:47:09.269611 containerd[1543]: time="2025-12-12T18:47:09.269010014Z" level=info msg="StopContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" returns successfully" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271207054Z" level=info msg="StopPodSandbox for \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\"" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271284549Z" level=info msg="Container to stop \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271307492Z" level=info msg="Container to stop \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271319875Z" level=info msg="Container to stop \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271331126Z" level=info msg="Container to stop \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.271782 containerd[1543]: time="2025-12-12T18:47:09.271380479Z" level=info msg="Container to stop \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 12 18:47:09.276287 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c-rootfs.mount: Deactivated successfully. Dec 12 18:47:09.303503 containerd[1543]: time="2025-12-12T18:47:09.303165780Z" level=info msg="shim disconnected" id=7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c namespace=k8s.io Dec 12 18:47:09.303503 containerd[1543]: time="2025-12-12T18:47:09.303213210Z" level=warning msg="cleaning up after shim disconnected" id=7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c namespace=k8s.io Dec 12 18:47:09.303503 containerd[1543]: time="2025-12-12T18:47:09.303222988Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 12 18:47:09.309808 systemd[1]: cri-containerd-a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea.scope: Deactivated successfully. Dec 12 18:47:09.311421 containerd[1543]: time="2025-12-12T18:47:09.310515422Z" level=info msg="received sandbox exit event container_id:\"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" id:\"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" exit_status:137 exited_at:{seconds:1765565229 nanos:310279759}" monitor_name=podsandbox Dec 12 18:47:09.348907 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c-shm.mount: Deactivated successfully. Dec 12 18:47:09.357291 containerd[1543]: time="2025-12-12T18:47:09.357231080Z" level=info msg="TearDown network for sandbox \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" successfully" Dec 12 18:47:09.357291 containerd[1543]: time="2025-12-12T18:47:09.357272969Z" level=info msg="StopPodSandbox for \"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" returns successfully" Dec 12 18:47:09.366094 containerd[1543]: time="2025-12-12T18:47:09.366019315Z" level=info msg="received sandbox container exit event sandbox_id:\"7bb24a640ad6b5abaa1079e03012c5e33b1944338272984ff0cee17511d0430c\" exit_status:137 exited_at:{seconds:1765565229 nanos:216183603}" monitor_name=criService Dec 12 18:47:09.391574 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea-rootfs.mount: Deactivated successfully. Dec 12 18:47:09.443999 containerd[1543]: time="2025-12-12T18:47:09.443897654Z" level=info msg="shim disconnected" id=a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea namespace=k8s.io Dec 12 18:47:09.443999 containerd[1543]: time="2025-12-12T18:47:09.443968608Z" level=warning msg="cleaning up after shim disconnected" id=a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea namespace=k8s.io Dec 12 18:47:09.443999 containerd[1543]: time="2025-12-12T18:47:09.443978998Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 12 18:47:09.474420 containerd[1543]: time="2025-12-12T18:47:09.472415906Z" level=info msg="received sandbox container exit event sandbox_id:\"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" exit_status:137 exited_at:{seconds:1765565229 nanos:310279759}" monitor_name=criService Dec 12 18:47:09.475878 containerd[1543]: time="2025-12-12T18:47:09.475392029Z" level=info msg="TearDown network for sandbox \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" successfully" Dec 12 18:47:09.475878 containerd[1543]: time="2025-12-12T18:47:09.475442463Z" level=info msg="StopPodSandbox for \"a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea\" returns successfully" Dec 12 18:47:09.495239 kubelet[2712]: I1212 18:47:09.494535 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m44cv\" (UniqueName: \"kubernetes.io/projected/3dbdbe3f-84a9-4932-9148-cd1f33377b91-kube-api-access-m44cv\") pod \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\" (UID: \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\") " Dec 12 18:47:09.495239 kubelet[2712]: I1212 18:47:09.494585 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3dbdbe3f-84a9-4932-9148-cd1f33377b91-cilium-config-path\") pod \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\" (UID: \"3dbdbe3f-84a9-4932-9148-cd1f33377b91\") " Dec 12 18:47:09.501087 kubelet[2712]: I1212 18:47:09.500990 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dbdbe3f-84a9-4932-9148-cd1f33377b91-kube-api-access-m44cv" (OuterVolumeSpecName: "kube-api-access-m44cv") pod "3dbdbe3f-84a9-4932-9148-cd1f33377b91" (UID: "3dbdbe3f-84a9-4932-9148-cd1f33377b91"). InnerVolumeSpecName "kube-api-access-m44cv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 12 18:47:09.503213 kubelet[2712]: I1212 18:47:09.503157 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbdbe3f-84a9-4932-9148-cd1f33377b91-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3dbdbe3f-84a9-4932-9148-cd1f33377b91" (UID: "3dbdbe3f-84a9-4932-9148-cd1f33377b91"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 12 18:47:09.595963 kubelet[2712]: I1212 18:47:09.595456 2712 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3dbdbe3f-84a9-4932-9148-cd1f33377b91-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.595963 kubelet[2712]: I1212 18:47:09.595510 2712 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m44cv\" (UniqueName: \"kubernetes.io/projected/3dbdbe3f-84a9-4932-9148-cd1f33377b91-kube-api-access-m44cv\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.696300 kubelet[2712]: I1212 18:47:09.696228 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-kernel\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696300 kubelet[2712]: I1212 18:47:09.696294 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-cgroup\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696300 kubelet[2712]: I1212 18:47:09.696316 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-xtables-lock\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696339 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-bpf-maps\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696404 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvvq5\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-kube-api-access-dvvq5\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696423 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cni-path\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696441 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-config-path\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696458 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-hostproc\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696628 kubelet[2712]: I1212 18:47:09.696478 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-lib-modules\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696496 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-etc-cni-netd\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696518 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-hubble-tls\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696536 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-run\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696596 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4548a03e-d58b-4ff6-a670-4064c33ce862-clustermesh-secrets\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696617 2712 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-net\") pod \"4548a03e-d58b-4ff6-a670-4064c33ce862\" (UID: \"4548a03e-d58b-4ff6-a670-4064c33ce862\") " Dec 12 18:47:09.696849 kubelet[2712]: I1212 18:47:09.696718 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.697040 kubelet[2712]: I1212 18:47:09.696762 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.697040 kubelet[2712]: I1212 18:47:09.696783 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.697040 kubelet[2712]: I1212 18:47:09.696801 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.697040 kubelet[2712]: I1212 18:47:09.696818 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.699283 kubelet[2712]: I1212 18:47:09.697380 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.699283 kubelet[2712]: I1212 18:47:09.697437 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cni-path" (OuterVolumeSpecName: "cni-path") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.699283 kubelet[2712]: I1212 18:47:09.698849 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.699822 kubelet[2712]: I1212 18:47:09.699768 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.702097 kubelet[2712]: I1212 18:47:09.702052 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-hostproc" (OuterVolumeSpecName: "hostproc") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 12 18:47:09.702204 kubelet[2712]: I1212 18:47:09.702074 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 12 18:47:09.706990 kubelet[2712]: I1212 18:47:09.706743 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 12 18:47:09.706990 kubelet[2712]: I1212 18:47:09.706932 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4548a03e-d58b-4ff6-a670-4064c33ce862-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 12 18:47:09.707976 kubelet[2712]: I1212 18:47:09.707896 2712 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-kube-api-access-dvvq5" (OuterVolumeSpecName: "kube-api-access-dvvq5") pod "4548a03e-d58b-4ff6-a670-4064c33ce862" (UID: "4548a03e-d58b-4ff6-a670-4064c33ce862"). InnerVolumeSpecName "kube-api-access-dvvq5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797564 2712 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dvvq5\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-kube-api-access-dvvq5\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797616 2712 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cni-path\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797629 2712 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797645 2712 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-hostproc\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797657 2712 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-lib-modules\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797668 2712 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797678 2712 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/4548a03e-d58b-4ff6-a670-4064c33ce862-hubble-tls\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.797661 kubelet[2712]: I1212 18:47:09.797689 2712 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-run\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797700 2712 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/4548a03e-d58b-4ff6-a670-4064c33ce862-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797711 2712 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797722 2712 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797734 2712 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797747 2712 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-xtables-lock\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:09.798088 kubelet[2712]: I1212 18:47:09.797757 2712 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/4548a03e-d58b-4ff6-a670-4064c33ce862-bpf-maps\") on node \"localhost\" DevicePath \"\"" Dec 12 18:47:10.154636 systemd[1]: var-lib-kubelet-pods-3dbdbe3f\x2d84a9\x2d4932\x2d9148\x2dcd1f33377b91-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dm44cv.mount: Deactivated successfully. Dec 12 18:47:10.157271 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a452beef34819e20a33f298c72f8dbda076f7eeca9da3e688d3bac77366199ea-shm.mount: Deactivated successfully. Dec 12 18:47:10.164532 systemd[1]: var-lib-kubelet-pods-4548a03e\x2dd58b\x2d4ff6\x2da670\x2d4064c33ce862-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 12 18:47:10.164817 systemd[1]: var-lib-kubelet-pods-4548a03e\x2dd58b\x2d4ff6\x2da670\x2d4064c33ce862-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 12 18:47:10.169316 systemd[1]: var-lib-kubelet-pods-4548a03e\x2dd58b\x2d4ff6\x2da670\x2d4064c33ce862-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ddvvq5.mount: Deactivated successfully. Dec 12 18:47:10.187959 kubelet[2712]: I1212 18:47:10.164998 2712 scope.go:117] "RemoveContainer" containerID="19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71" Dec 12 18:47:10.188065 containerd[1543]: time="2025-12-12T18:47:10.184548963Z" level=info msg="RemoveContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\"" Dec 12 18:47:10.199972 systemd[1]: Removed slice kubepods-besteffort-pod3dbdbe3f_84a9_4932_9148_cd1f33377b91.slice - libcontainer container kubepods-besteffort-pod3dbdbe3f_84a9_4932_9148_cd1f33377b91.slice. Dec 12 18:47:10.212324 systemd[1]: Removed slice kubepods-burstable-pod4548a03e_d58b_4ff6_a670_4064c33ce862.slice - libcontainer container kubepods-burstable-pod4548a03e_d58b_4ff6_a670_4064c33ce862.slice. Dec 12 18:47:10.214404 systemd[1]: kubepods-burstable-pod4548a03e_d58b_4ff6_a670_4064c33ce862.slice: Consumed 8.745s CPU time, 126.1M memory peak, 256K read from disk, 13.3M written to disk. Dec 12 18:47:10.230967 containerd[1543]: time="2025-12-12T18:47:10.230850542Z" level=info msg="RemoveContainer for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" returns successfully" Dec 12 18:47:10.235997 kubelet[2712]: I1212 18:47:10.231277 2712 scope.go:117] "RemoveContainer" containerID="fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58" Dec 12 18:47:10.240735 containerd[1543]: time="2025-12-12T18:47:10.237791639Z" level=info msg="RemoveContainer for \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\"" Dec 12 18:47:10.255943 containerd[1543]: time="2025-12-12T18:47:10.255784295Z" level=info msg="RemoveContainer for \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" returns successfully" Dec 12 18:47:10.257488 kubelet[2712]: I1212 18:47:10.256249 2712 scope.go:117] "RemoveContainer" containerID="4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0" Dec 12 18:47:10.261452 containerd[1543]: time="2025-12-12T18:47:10.261237854Z" level=info msg="RemoveContainer for \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\"" Dec 12 18:47:10.274228 containerd[1543]: time="2025-12-12T18:47:10.273657798Z" level=info msg="RemoveContainer for \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" returns successfully" Dec 12 18:47:10.274463 kubelet[2712]: I1212 18:47:10.274033 2712 scope.go:117] "RemoveContainer" containerID="ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6" Dec 12 18:47:10.280542 containerd[1543]: time="2025-12-12T18:47:10.280444675Z" level=info msg="RemoveContainer for \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\"" Dec 12 18:47:10.289115 containerd[1543]: time="2025-12-12T18:47:10.288378599Z" level=info msg="RemoveContainer for \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" returns successfully" Dec 12 18:47:10.289284 kubelet[2712]: I1212 18:47:10.288704 2712 scope.go:117] "RemoveContainer" containerID="0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4" Dec 12 18:47:10.293383 containerd[1543]: time="2025-12-12T18:47:10.293316318Z" level=info msg="RemoveContainer for \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\"" Dec 12 18:47:10.309957 containerd[1543]: time="2025-12-12T18:47:10.309866250Z" level=info msg="RemoveContainer for \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" returns successfully" Dec 12 18:47:10.310362 kubelet[2712]: I1212 18:47:10.310271 2712 scope.go:117] "RemoveContainer" containerID="19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71" Dec 12 18:47:10.312694 containerd[1543]: time="2025-12-12T18:47:10.312599382Z" level=error msg="ContainerStatus for \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\": not found" Dec 12 18:47:10.314335 kubelet[2712]: E1212 18:47:10.314267 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\": not found" containerID="19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71" Dec 12 18:47:10.314685 kubelet[2712]: I1212 18:47:10.314331 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71"} err="failed to get container status \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\": rpc error: code = NotFound desc = an error occurred when try to find container \"19c9efd47eb41ac87337a92c9d5dbd4dabc7b42476439d25eb096f955b303b71\": not found" Dec 12 18:47:10.314762 kubelet[2712]: I1212 18:47:10.314689 2712 scope.go:117] "RemoveContainer" containerID="fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58" Dec 12 18:47:10.315106 containerd[1543]: time="2025-12-12T18:47:10.315029173Z" level=error msg="ContainerStatus for \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\": not found" Dec 12 18:47:10.315273 kubelet[2712]: E1212 18:47:10.315200 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\": not found" containerID="fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58" Dec 12 18:47:10.315273 kubelet[2712]: I1212 18:47:10.315225 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58"} err="failed to get container status \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\": rpc error: code = NotFound desc = an error occurred when try to find container \"fcf49c09ad337e474ef33ef0cd727fac6a1b40d6c111fbb58cc162d7f22fbd58\": not found" Dec 12 18:47:10.315273 kubelet[2712]: I1212 18:47:10.315240 2712 scope.go:117] "RemoveContainer" containerID="4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0" Dec 12 18:47:10.315541 containerd[1543]: time="2025-12-12T18:47:10.315495179Z" level=error msg="ContainerStatus for \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\": not found" Dec 12 18:47:10.319288 kubelet[2712]: E1212 18:47:10.319221 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\": not found" containerID="4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0" Dec 12 18:47:10.319288 kubelet[2712]: I1212 18:47:10.319277 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0"} err="failed to get container status \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\": rpc error: code = NotFound desc = an error occurred when try to find container \"4531c930d97a8b10675737f70284e56be6a05bab234e06d5b452b539c05c14f0\": not found" Dec 12 18:47:10.319548 kubelet[2712]: I1212 18:47:10.319308 2712 scope.go:117] "RemoveContainer" containerID="ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6" Dec 12 18:47:10.319813 containerd[1543]: time="2025-12-12T18:47:10.319735566Z" level=error msg="ContainerStatus for \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\": not found" Dec 12 18:47:10.320026 kubelet[2712]: E1212 18:47:10.319949 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\": not found" containerID="ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6" Dec 12 18:47:10.320026 kubelet[2712]: I1212 18:47:10.319978 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6"} err="failed to get container status \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\": rpc error: code = NotFound desc = an error occurred when try to find container \"ce29e31fee77b1b0ae16509e4dac7647d854c90ac06b8e158e0df5a4757900e6\": not found" Dec 12 18:47:10.320026 kubelet[2712]: I1212 18:47:10.319996 2712 scope.go:117] "RemoveContainer" containerID="0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4" Dec 12 18:47:10.322321 kubelet[2712]: E1212 18:47:10.320398 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\": not found" containerID="0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4" Dec 12 18:47:10.322321 kubelet[2712]: I1212 18:47:10.320443 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4"} err="failed to get container status \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\": rpc error: code = NotFound desc = an error occurred when try to find container \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\": not found" Dec 12 18:47:10.322321 kubelet[2712]: I1212 18:47:10.320476 2712 scope.go:117] "RemoveContainer" containerID="91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655" Dec 12 18:47:10.322449 containerd[1543]: time="2025-12-12T18:47:10.320173791Z" level=error msg="ContainerStatus for \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0b90a1f608c285f8f3e1cc70e8f2247c13700729ea308da7979c0c6b26a3d8f4\": not found" Dec 12 18:47:10.328424 containerd[1543]: time="2025-12-12T18:47:10.324214511Z" level=info msg="RemoveContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\"" Dec 12 18:47:10.334461 containerd[1543]: time="2025-12-12T18:47:10.330977013Z" level=info msg="RemoveContainer for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" returns successfully" Dec 12 18:47:10.334461 containerd[1543]: time="2025-12-12T18:47:10.334265679Z" level=error msg="ContainerStatus for \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\": not found" Dec 12 18:47:10.334639 kubelet[2712]: I1212 18:47:10.331278 2712 scope.go:117] "RemoveContainer" containerID="91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655" Dec 12 18:47:10.335306 kubelet[2712]: E1212 18:47:10.335269 2712 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\": not found" containerID="91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655" Dec 12 18:47:10.337016 kubelet[2712]: I1212 18:47:10.335402 2712 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655"} err="failed to get container status \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\": rpc error: code = NotFound desc = an error occurred when try to find container \"91dec422c1f5077ccf606e2602d4695fce3c19941f18d5700cf8fd517d4c7655\": not found" Dec 12 18:47:10.739963 kubelet[2712]: I1212 18:47:10.739869 2712 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dbdbe3f-84a9-4932-9148-cd1f33377b91" path="/var/lib/kubelet/pods/3dbdbe3f-84a9-4932-9148-cd1f33377b91/volumes" Dec 12 18:47:10.740596 kubelet[2712]: I1212 18:47:10.740574 2712 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4548a03e-d58b-4ff6-a670-4064c33ce862" path="/var/lib/kubelet/pods/4548a03e-d58b-4ff6-a670-4064c33ce862/volumes" Dec 12 18:47:10.895883 sshd[4338]: Connection closed by 10.0.0.1 port 43642 Dec 12 18:47:10.896771 sshd-session[4335]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:10.927024 systemd[1]: sshd@23-10.0.0.119:22-10.0.0.1:43642.service: Deactivated successfully. Dec 12 18:47:10.936627 systemd[1]: session-24.scope: Deactivated successfully. Dec 12 18:47:10.937069 systemd[1]: session-24.scope: Consumed 1.005s CPU time, 25.9M memory peak. Dec 12 18:47:10.938884 systemd-logind[1524]: Session 24 logged out. Waiting for processes to exit. Dec 12 18:47:10.948476 systemd[1]: Started sshd@24-10.0.0.119:22-10.0.0.1:38076.service - OpenSSH per-connection server daemon (10.0.0.1:38076). Dec 12 18:47:10.953180 systemd-logind[1524]: Removed session 24. Dec 12 18:47:11.043178 sshd[4479]: Accepted publickey for core from 10.0.0.1 port 38076 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:11.046046 sshd-session[4479]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:11.063642 systemd-logind[1524]: New session 25 of user core. Dec 12 18:47:11.072673 systemd[1]: Started session-25.scope - Session 25 of User core. Dec 12 18:47:11.797581 kubelet[2712]: E1212 18:47:11.797520 2712 kubelet.go:3011] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 12 18:47:11.864553 sshd[4482]: Connection closed by 10.0.0.1 port 38076 Dec 12 18:47:11.865174 sshd-session[4479]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:11.890778 systemd[1]: sshd@24-10.0.0.119:22-10.0.0.1:38076.service: Deactivated successfully. Dec 12 18:47:11.898540 systemd[1]: session-25.scope: Deactivated successfully. Dec 12 18:47:11.903862 systemd-logind[1524]: Session 25 logged out. Waiting for processes to exit. Dec 12 18:47:11.914064 systemd[1]: Started sshd@25-10.0.0.119:22-10.0.0.1:38086.service - OpenSSH per-connection server daemon (10.0.0.1:38086). Dec 12 18:47:11.915470 systemd-logind[1524]: Removed session 25. Dec 12 18:47:11.930416 systemd[1]: Created slice kubepods-burstable-pod89b93aee_6cf9_49b3_86a1_bb4f916b4eba.slice - libcontainer container kubepods-burstable-pod89b93aee_6cf9_49b3_86a1_bb4f916b4eba.slice. Dec 12 18:47:11.934484 kubelet[2712]: I1212 18:47:11.934438 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-cilium-cgroup\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934484 kubelet[2712]: I1212 18:47:11.934481 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-etc-cni-netd\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934599 kubelet[2712]: I1212 18:47:11.934504 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-cilium-ipsec-secrets\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934599 kubelet[2712]: I1212 18:47:11.934524 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-cilium-config-path\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934599 kubelet[2712]: I1212 18:47:11.934541 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-host-proc-sys-kernel\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934599 kubelet[2712]: I1212 18:47:11.934558 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-hubble-tls\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934599 kubelet[2712]: I1212 18:47:11.934575 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27mzr\" (UniqueName: \"kubernetes.io/projected/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-kube-api-access-27mzr\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934594 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-host-proc-sys-net\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934610 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-cilium-run\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934627 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-bpf-maps\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934644 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-xtables-lock\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934660 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-clustermesh-secrets\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934757 kubelet[2712]: I1212 18:47:11.934678 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-hostproc\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934930 kubelet[2712]: I1212 18:47:11.934704 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-cni-path\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:11.934930 kubelet[2712]: I1212 18:47:11.934721 2712 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/89b93aee-6cf9-49b3-86a1-bb4f916b4eba-lib-modules\") pod \"cilium-tfrsd\" (UID: \"89b93aee-6cf9-49b3-86a1-bb4f916b4eba\") " pod="kube-system/cilium-tfrsd" Dec 12 18:47:12.005551 sshd[4494]: Accepted publickey for core from 10.0.0.1 port 38086 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:12.007575 sshd-session[4494]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:12.013098 systemd-logind[1524]: New session 26 of user core. Dec 12 18:47:12.023675 systemd[1]: Started session-26.scope - Session 26 of User core. Dec 12 18:47:12.078278 sshd[4497]: Connection closed by 10.0.0.1 port 38086 Dec 12 18:47:12.078920 sshd-session[4494]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:12.089811 systemd[1]: sshd@25-10.0.0.119:22-10.0.0.1:38086.service: Deactivated successfully. Dec 12 18:47:12.092407 systemd[1]: session-26.scope: Deactivated successfully. Dec 12 18:47:12.093705 systemd-logind[1524]: Session 26 logged out. Waiting for processes to exit. Dec 12 18:47:12.097328 systemd[1]: Started sshd@26-10.0.0.119:22-10.0.0.1:38092.service - OpenSSH per-connection server daemon (10.0.0.1:38092). Dec 12 18:47:12.098154 systemd-logind[1524]: Removed session 26. Dec 12 18:47:12.160574 sshd[4508]: Accepted publickey for core from 10.0.0.1 port 38092 ssh2: RSA SHA256:P1s5gEg3hMj1tDtE6I6RWVrUOC+71cTuFOU1V+vviNE Dec 12 18:47:12.162460 sshd-session[4508]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 12 18:47:12.167387 systemd-logind[1524]: New session 27 of user core. Dec 12 18:47:12.176587 systemd[1]: Started session-27.scope - Session 27 of User core. Dec 12 18:47:12.242816 containerd[1543]: time="2025-12-12T18:47:12.242739673Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tfrsd,Uid:89b93aee-6cf9-49b3-86a1-bb4f916b4eba,Namespace:kube-system,Attempt:0,}" Dec 12 18:47:12.267257 containerd[1543]: time="2025-12-12T18:47:12.266935107Z" level=info msg="connecting to shim 1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" namespace=k8s.io protocol=ttrpc version=3 Dec 12 18:47:12.298694 systemd[1]: Started cri-containerd-1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6.scope - libcontainer container 1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6. Dec 12 18:47:12.340235 containerd[1543]: time="2025-12-12T18:47:12.339890543Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tfrsd,Uid:89b93aee-6cf9-49b3-86a1-bb4f916b4eba,Namespace:kube-system,Attempt:0,} returns sandbox id \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\"" Dec 12 18:47:12.347923 containerd[1543]: time="2025-12-12T18:47:12.347833263Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 12 18:47:12.360281 containerd[1543]: time="2025-12-12T18:47:12.360208518Z" level=info msg="Container 8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:47:12.370217 containerd[1543]: time="2025-12-12T18:47:12.370122433Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746\"" Dec 12 18:47:12.371159 containerd[1543]: time="2025-12-12T18:47:12.371114002Z" level=info msg="StartContainer for \"8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746\"" Dec 12 18:47:12.372584 containerd[1543]: time="2025-12-12T18:47:12.372550941Z" level=info msg="connecting to shim 8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" protocol=ttrpc version=3 Dec 12 18:47:12.404686 systemd[1]: Started cri-containerd-8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746.scope - libcontainer container 8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746. Dec 12 18:47:12.440672 containerd[1543]: time="2025-12-12T18:47:12.440616150Z" level=info msg="StartContainer for \"8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746\" returns successfully" Dec 12 18:47:12.450144 systemd[1]: cri-containerd-8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746.scope: Deactivated successfully. Dec 12 18:47:12.452219 containerd[1543]: time="2025-12-12T18:47:12.452177541Z" level=info msg="received container exit event container_id:\"8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746\" id:\"8823f39ebd665a2592ef80ccb908a136fa7623a132b564f2ac6813753ed33746\" pid:4576 exited_at:{seconds:1765565232 nanos:451712865}" Dec 12 18:47:13.443338 containerd[1543]: time="2025-12-12T18:47:13.443264864Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 12 18:47:13.963693 containerd[1543]: time="2025-12-12T18:47:13.963618874Z" level=info msg="Container 9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:47:13.975954 containerd[1543]: time="2025-12-12T18:47:13.975855700Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494\"" Dec 12 18:47:13.977698 containerd[1543]: time="2025-12-12T18:47:13.976494034Z" level=info msg="StartContainer for \"9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494\"" Dec 12 18:47:13.977698 containerd[1543]: time="2025-12-12T18:47:13.977618836Z" level=info msg="connecting to shim 9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" protocol=ttrpc version=3 Dec 12 18:47:14.006604 systemd[1]: Started cri-containerd-9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494.scope - libcontainer container 9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494. Dec 12 18:47:14.048538 systemd[1]: cri-containerd-9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494.scope: Deactivated successfully. Dec 12 18:47:14.050206 containerd[1543]: time="2025-12-12T18:47:14.050144410Z" level=info msg="received container exit event container_id:\"9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494\" id:\"9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494\" pid:4623 exited_at:{seconds:1765565234 nanos:49665186}" Dec 12 18:47:14.051321 containerd[1543]: time="2025-12-12T18:47:14.051295223Z" level=info msg="StartContainer for \"9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494\" returns successfully" Dec 12 18:47:14.075452 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f39cc244e2cb37fb10aa385a622f999bcfe73c2c3f6c1f6b17c978e3a28c494-rootfs.mount: Deactivated successfully. Dec 12 18:47:14.198252 containerd[1543]: time="2025-12-12T18:47:14.198207228Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 12 18:47:14.211296 containerd[1543]: time="2025-12-12T18:47:14.211252922Z" level=info msg="Container db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:47:14.224686 containerd[1543]: time="2025-12-12T18:47:14.224575639Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6\"" Dec 12 18:47:14.225116 containerd[1543]: time="2025-12-12T18:47:14.225093396Z" level=info msg="StartContainer for \"db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6\"" Dec 12 18:47:14.226755 containerd[1543]: time="2025-12-12T18:47:14.226721872Z" level=info msg="connecting to shim db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" protocol=ttrpc version=3 Dec 12 18:47:14.252585 systemd[1]: Started cri-containerd-db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6.scope - libcontainer container db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6. Dec 12 18:47:14.345325 systemd[1]: cri-containerd-db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6.scope: Deactivated successfully. Dec 12 18:47:14.345783 containerd[1543]: time="2025-12-12T18:47:14.345727562Z" level=info msg="StartContainer for \"db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6\" returns successfully" Dec 12 18:47:14.347453 containerd[1543]: time="2025-12-12T18:47:14.347420700Z" level=info msg="received container exit event container_id:\"db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6\" id:\"db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6\" pid:4666 exited_at:{seconds:1765565234 nanos:347093091}" Dec 12 18:47:15.075115 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-db0920dca40cc78f654b0541df7128db999be5da785fae17ed52f9c0e62d87b6-rootfs.mount: Deactivated successfully. Dec 12 18:47:15.392417 containerd[1543]: time="2025-12-12T18:47:15.391755693Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 12 18:47:15.410511 containerd[1543]: time="2025-12-12T18:47:15.410454862Z" level=info msg="Container 5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:47:15.423338 containerd[1543]: time="2025-12-12T18:47:15.423274950Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c\"" Dec 12 18:47:15.424077 containerd[1543]: time="2025-12-12T18:47:15.423908437Z" level=info msg="StartContainer for \"5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c\"" Dec 12 18:47:15.425075 containerd[1543]: time="2025-12-12T18:47:15.425016622Z" level=info msg="connecting to shim 5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" protocol=ttrpc version=3 Dec 12 18:47:15.446570 systemd[1]: Started cri-containerd-5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c.scope - libcontainer container 5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c. Dec 12 18:47:15.486528 systemd[1]: cri-containerd-5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c.scope: Deactivated successfully. Dec 12 18:47:15.514289 containerd[1543]: time="2025-12-12T18:47:15.514221679Z" level=info msg="received container exit event container_id:\"5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c\" id:\"5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c\" pid:4705 exited_at:{seconds:1765565235 nanos:486715886}" Dec 12 18:47:15.522955 containerd[1543]: time="2025-12-12T18:47:15.522913114Z" level=info msg="StartContainer for \"5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c\" returns successfully" Dec 12 18:47:15.537910 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5de80b24ea2fc1ce7b82f425fc78006ca878c4665457e0b1eef0b6c779d1e13c-rootfs.mount: Deactivated successfully. Dec 12 18:47:16.394182 containerd[1543]: time="2025-12-12T18:47:16.394120266Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 12 18:47:16.493385 containerd[1543]: time="2025-12-12T18:47:16.493107757Z" level=info msg="Container 36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860: CDI devices from CRI Config.CDIDevices: []" Dec 12 18:47:16.497995 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount364906216.mount: Deactivated successfully. Dec 12 18:47:16.506118 containerd[1543]: time="2025-12-12T18:47:16.506047582Z" level=info msg="CreateContainer within sandbox \"1bf61294891a5cd5bda803503873233dee08b1ace8eebe7a6183fca08252eae6\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860\"" Dec 12 18:47:16.507378 containerd[1543]: time="2025-12-12T18:47:16.506678987Z" level=info msg="StartContainer for \"36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860\"" Dec 12 18:47:16.507925 containerd[1543]: time="2025-12-12T18:47:16.507902882Z" level=info msg="connecting to shim 36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860" address="unix:///run/containerd/s/b438b9e65be1a5bb957ae2dd4c0476e5c30966e5bd98785b8341774a603044a4" protocol=ttrpc version=3 Dec 12 18:47:16.535590 systemd[1]: Started cri-containerd-36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860.scope - libcontainer container 36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860. Dec 12 18:47:16.591984 containerd[1543]: time="2025-12-12T18:47:16.591929436Z" level=info msg="StartContainer for \"36f297eb126b2fd82e2d4141d2142f7ffb118813aaf950c2ebe65b259c07d860\" returns successfully" Dec 12 18:47:17.024387 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni-avx)) Dec 12 18:47:17.228881 kubelet[2712]: I1212 18:47:17.228784 2712 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-tfrsd" podStartSLOduration=6.228763638 podStartE2EDuration="6.228763638s" podCreationTimestamp="2025-12-12 18:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 18:47:17.228433443 +0000 UTC m=+90.586046186" watchObservedRunningTime="2025-12-12 18:47:17.228763638 +0000 UTC m=+90.586376381" Dec 12 18:47:20.259282 systemd-networkd[1432]: lxc_health: Link UP Dec 12 18:47:20.263977 systemd-networkd[1432]: lxc_health: Gained carrier Dec 12 18:47:21.974540 systemd-networkd[1432]: lxc_health: Gained IPv6LL Dec 12 18:47:27.246664 sshd[4511]: Connection closed by 10.0.0.1 port 38092 Dec 12 18:47:27.251606 sshd-session[4508]: pam_unix(sshd:session): session closed for user core Dec 12 18:47:27.262842 systemd[1]: sshd@26-10.0.0.119:22-10.0.0.1:38092.service: Deactivated successfully. Dec 12 18:47:27.266102 systemd[1]: session-27.scope: Deactivated successfully. Dec 12 18:47:27.270150 systemd-logind[1524]: Session 27 logged out. Waiting for processes to exit. Dec 12 18:47:27.272961 systemd-logind[1524]: Removed session 27.