Sep 13 01:01:01.040341 kernel: Linux version 6.6.106-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 22:30:50 -00 2025 Sep 13 01:01:01.040375 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:01:01.040389 kernel: BIOS-provided physical RAM map: Sep 13 01:01:01.040405 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 13 01:01:01.040415 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 13 01:01:01.040425 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 13 01:01:01.040436 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdbfff] usable Sep 13 01:01:01.040447 kernel: BIOS-e820: [mem 0x000000007ffdc000-0x000000007fffffff] reserved Sep 13 01:01:01.040457 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Sep 13 01:01:01.040467 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Sep 13 01:01:01.040478 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 13 01:01:01.040488 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 13 01:01:01.040503 kernel: NX (Execute Disable) protection: active Sep 13 01:01:01.040514 kernel: APIC: Static calls initialized Sep 13 01:01:01.040526 kernel: SMBIOS 2.8 present. Sep 13 01:01:01.040538 kernel: DMI: Red Hat KVM/RHEL-AV, BIOS 1.13.0-2.module_el8.5.0+2608+72063365 04/01/2014 Sep 13 01:01:01.040549 kernel: Hypervisor detected: KVM Sep 13 01:01:01.040565 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 13 01:01:01.040577 kernel: kvm-clock: using sched offset of 4436058069 cycles Sep 13 01:01:01.040589 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 13 01:01:01.040616 kernel: tsc: Detected 2499.998 MHz processor Sep 13 01:01:01.040629 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 13 01:01:01.040641 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 13 01:01:01.040652 kernel: last_pfn = 0x7ffdc max_arch_pfn = 0x400000000 Sep 13 01:01:01.040664 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 13 01:01:01.040675 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 13 01:01:01.040693 kernel: Using GB pages for direct mapping Sep 13 01:01:01.040704 kernel: ACPI: Early table checksum verification disabled Sep 13 01:01:01.040715 kernel: ACPI: RSDP 0x00000000000F5AA0 000014 (v00 BOCHS ) Sep 13 01:01:01.040727 kernel: ACPI: RSDT 0x000000007FFE47A5 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040739 kernel: ACPI: FACP 0x000000007FFE438D 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040750 kernel: ACPI: DSDT 0x000000007FFDFD80 00460D (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040762 kernel: ACPI: FACS 0x000000007FFDFD40 000040 Sep 13 01:01:01.040773 kernel: ACPI: APIC 0x000000007FFE4481 0000F0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040785 kernel: ACPI: SRAT 0x000000007FFE4571 0001D0 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040800 kernel: ACPI: MCFG 0x000000007FFE4741 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040812 kernel: ACPI: WAET 0x000000007FFE477D 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 01:01:01.040824 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe438d-0x7ffe4480] Sep 13 01:01:01.040835 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffdfd80-0x7ffe438c] Sep 13 01:01:01.040847 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffdfd40-0x7ffdfd7f] Sep 13 01:01:01.040865 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe4481-0x7ffe4570] Sep 13 01:01:01.040877 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe4571-0x7ffe4740] Sep 13 01:01:01.040907 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe4741-0x7ffe477c] Sep 13 01:01:01.040920 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe477d-0x7ffe47a4] Sep 13 01:01:01.040932 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 13 01:01:01.040944 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 13 01:01:01.040956 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Sep 13 01:01:01.040967 kernel: SRAT: PXM 0 -> APIC 0x03 -> Node 0 Sep 13 01:01:01.040979 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Sep 13 01:01:01.040997 kernel: SRAT: PXM 0 -> APIC 0x05 -> Node 0 Sep 13 01:01:01.041009 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Sep 13 01:01:01.041021 kernel: SRAT: PXM 0 -> APIC 0x07 -> Node 0 Sep 13 01:01:01.041032 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Sep 13 01:01:01.041044 kernel: SRAT: PXM 0 -> APIC 0x09 -> Node 0 Sep 13 01:01:01.041056 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Sep 13 01:01:01.041068 kernel: SRAT: PXM 0 -> APIC 0x0b -> Node 0 Sep 13 01:01:01.041080 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Sep 13 01:01:01.041092 kernel: SRAT: PXM 0 -> APIC 0x0d -> Node 0 Sep 13 01:01:01.041103 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Sep 13 01:01:01.041119 kernel: SRAT: PXM 0 -> APIC 0x0f -> Node 0 Sep 13 01:01:01.041132 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Sep 13 01:01:01.041144 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Sep 13 01:01:01.041155 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x20800fffff] hotplug Sep 13 01:01:01.041168 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdbfff] -> [mem 0x00000000-0x7ffdbfff] Sep 13 01:01:01.041180 kernel: NODE_DATA(0) allocated [mem 0x7ffd6000-0x7ffdbfff] Sep 13 01:01:01.041192 kernel: Zone ranges: Sep 13 01:01:01.041204 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 13 01:01:01.041216 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdbfff] Sep 13 01:01:01.041232 kernel: Normal empty Sep 13 01:01:01.041245 kernel: Movable zone start for each node Sep 13 01:01:01.041256 kernel: Early memory node ranges Sep 13 01:01:01.041269 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 13 01:01:01.041281 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdbfff] Sep 13 01:01:01.041292 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdbfff] Sep 13 01:01:01.041304 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 13 01:01:01.041316 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 13 01:01:01.041328 kernel: On node 0, zone DMA32: 36 pages in unavailable ranges Sep 13 01:01:01.041340 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 13 01:01:01.041357 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 13 01:01:01.041369 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 13 01:01:01.041381 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 13 01:01:01.041393 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 13 01:01:01.041405 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 13 01:01:01.041417 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 13 01:01:01.041429 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 13 01:01:01.041441 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 13 01:01:01.041453 kernel: TSC deadline timer available Sep 13 01:01:01.041470 kernel: smpboot: Allowing 16 CPUs, 14 hotplug CPUs Sep 13 01:01:01.041556 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 13 01:01:01.041573 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Sep 13 01:01:01.041585 kernel: Booting paravirtualized kernel on KVM Sep 13 01:01:01.041624 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 13 01:01:01.041638 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:16 nr_cpu_ids:16 nr_node_ids:1 Sep 13 01:01:01.041650 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u262144 Sep 13 01:01:01.041662 kernel: pcpu-alloc: s197160 r8192 d32216 u262144 alloc=1*2097152 Sep 13 01:01:01.041674 kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 12 13 14 15 Sep 13 01:01:01.041693 kernel: kvm-guest: PV spinlocks enabled Sep 13 01:01:01.041705 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 13 01:01:01.041720 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:01:01.041733 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 01:01:01.041745 kernel: random: crng init done Sep 13 01:01:01.041757 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 01:01:01.041769 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 13 01:01:01.041781 kernel: Fallback order for Node 0: 0 Sep 13 01:01:01.041798 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515804 Sep 13 01:01:01.041810 kernel: Policy zone: DMA32 Sep 13 01:01:01.041822 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 01:01:01.041834 kernel: software IO TLB: area num 16. Sep 13 01:01:01.041847 kernel: Memory: 1901532K/2096616K available (12288K kernel code, 2293K rwdata, 22744K rodata, 42884K init, 2312K bss, 194824K reserved, 0K cma-reserved) Sep 13 01:01:01.041859 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=16, Nodes=1 Sep 13 01:01:01.041871 kernel: Kernel/User page tables isolation: enabled Sep 13 01:01:01.041883 kernel: ftrace: allocating 37974 entries in 149 pages Sep 13 01:01:01.041908 kernel: ftrace: allocated 149 pages with 4 groups Sep 13 01:01:01.041926 kernel: Dynamic Preempt: voluntary Sep 13 01:01:01.041938 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 13 01:01:01.041951 kernel: rcu: RCU event tracing is enabled. Sep 13 01:01:01.041963 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=16. Sep 13 01:01:01.041976 kernel: Trampoline variant of Tasks RCU enabled. Sep 13 01:01:01.042000 kernel: Rude variant of Tasks RCU enabled. Sep 13 01:01:01.042017 kernel: Tracing variant of Tasks RCU enabled. Sep 13 01:01:01.042030 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 01:01:01.042042 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=16 Sep 13 01:01:01.042055 kernel: NR_IRQS: 33024, nr_irqs: 552, preallocated irqs: 16 Sep 13 01:01:01.042068 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 13 01:01:01.042085 kernel: Console: colour VGA+ 80x25 Sep 13 01:01:01.042097 kernel: printk: console [tty0] enabled Sep 13 01:01:01.042110 kernel: printk: console [ttyS0] enabled Sep 13 01:01:01.042123 kernel: ACPI: Core revision 20230628 Sep 13 01:01:01.042135 kernel: APIC: Switch to symmetric I/O mode setup Sep 13 01:01:01.042148 kernel: x2apic enabled Sep 13 01:01:01.042165 kernel: APIC: Switched APIC routing to: physical x2apic Sep 13 01:01:01.042178 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Sep 13 01:01:01.042191 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499998) Sep 13 01:01:01.042204 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Sep 13 01:01:01.042217 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Sep 13 01:01:01.042229 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Sep 13 01:01:01.042242 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 13 01:01:01.042254 kernel: Spectre V2 : Mitigation: Retpolines Sep 13 01:01:01.042267 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 13 01:01:01.042284 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Sep 13 01:01:01.042297 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 13 01:01:01.042310 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 13 01:01:01.042335 kernel: MDS: Mitigation: Clear CPU buffers Sep 13 01:01:01.042347 kernel: MMIO Stale Data: Unknown: No mitigations Sep 13 01:01:01.042359 kernel: SRBDS: Unknown: Dependent on hypervisor status Sep 13 01:01:01.042371 kernel: active return thunk: its_return_thunk Sep 13 01:01:01.042383 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 13 01:01:01.042395 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 13 01:01:01.042407 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 13 01:01:01.042420 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 13 01:01:01.042436 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 13 01:01:01.042448 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Sep 13 01:01:01.042461 kernel: Freeing SMP alternatives memory: 32K Sep 13 01:01:01.042473 kernel: pid_max: default: 32768 minimum: 301 Sep 13 01:01:01.042485 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 13 01:01:01.042497 kernel: landlock: Up and running. Sep 13 01:01:01.042522 kernel: SELinux: Initializing. Sep 13 01:01:01.042535 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 01:01:01.042548 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 01:01:01.042560 kernel: smpboot: CPU0: Intel Xeon E3-12xx v2 (Ivy Bridge, IBRS) (family: 0x6, model: 0x3a, stepping: 0x9) Sep 13 01:01:01.042573 kernel: RCU Tasks: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:01:01.042591 kernel: RCU Tasks Rude: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:01:01.042604 kernel: RCU Tasks Trace: Setting shift to 4 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=16. Sep 13 01:01:01.042617 kernel: Performance Events: unsupported p6 CPU model 58 no PMU driver, software events only. Sep 13 01:01:01.042649 kernel: signal: max sigframe size: 1776 Sep 13 01:01:01.042662 kernel: rcu: Hierarchical SRCU implementation. Sep 13 01:01:01.042675 kernel: rcu: Max phase no-delay instances is 400. Sep 13 01:01:01.042688 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 13 01:01:01.042701 kernel: smp: Bringing up secondary CPUs ... Sep 13 01:01:01.042713 kernel: smpboot: x86: Booting SMP configuration: Sep 13 01:01:01.042732 kernel: .... node #0, CPUs: #1 Sep 13 01:01:01.042745 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Sep 13 01:01:01.042758 kernel: smp: Brought up 1 node, 2 CPUs Sep 13 01:01:01.042771 kernel: smpboot: Max logical packages: 16 Sep 13 01:01:01.042783 kernel: smpboot: Total of 2 processors activated (9999.99 BogoMIPS) Sep 13 01:01:01.042796 kernel: devtmpfs: initialized Sep 13 01:01:01.042809 kernel: x86/mm: Memory block size: 128MB Sep 13 01:01:01.042822 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 01:01:01.042835 kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear) Sep 13 01:01:01.042852 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 01:01:01.042865 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 01:01:01.042877 kernel: audit: initializing netlink subsys (disabled) Sep 13 01:01:01.042901 kernel: audit: type=2000 audit(1757725259.364:1): state=initialized audit_enabled=0 res=1 Sep 13 01:01:01.042914 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 01:01:01.042927 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 13 01:01:01.042939 kernel: cpuidle: using governor menu Sep 13 01:01:01.042952 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 01:01:01.042965 kernel: dca service started, version 1.12.1 Sep 13 01:01:01.042983 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Sep 13 01:01:01.042996 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Sep 13 01:01:01.043009 kernel: PCI: Using configuration type 1 for base access Sep 13 01:01:01.043022 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 13 01:01:01.043035 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 13 01:01:01.043047 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 13 01:01:01.043060 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 01:01:01.043073 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 13 01:01:01.043086 kernel: ACPI: Added _OSI(Module Device) Sep 13 01:01:01.043103 kernel: ACPI: Added _OSI(Processor Device) Sep 13 01:01:01.043116 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 01:01:01.043129 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 13 01:01:01.043142 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 13 01:01:01.043154 kernel: ACPI: Interpreter enabled Sep 13 01:01:01.043167 kernel: ACPI: PM: (supports S0 S5) Sep 13 01:01:01.043180 kernel: ACPI: Using IOAPIC for interrupt routing Sep 13 01:01:01.043193 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 13 01:01:01.043205 kernel: PCI: Using E820 reservations for host bridge windows Sep 13 01:01:01.043222 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Sep 13 01:01:01.043235 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 13 01:01:01.043506 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 13 01:01:01.043728 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 13 01:01:01.043923 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 13 01:01:01.043944 kernel: PCI host bridge to bus 0000:00 Sep 13 01:01:01.044133 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 13 01:01:01.044305 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 13 01:01:01.044469 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 13 01:01:01.044657 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Sep 13 01:01:01.044862 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Sep 13 01:01:01.045039 kernel: pci_bus 0000:00: root bus resource [mem 0x20c0000000-0x28bfffffff window] Sep 13 01:01:01.045202 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 13 01:01:01.045410 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Sep 13 01:01:01.046839 kernel: pci 0000:00:01.0: [1013:00b8] type 00 class 0x030000 Sep 13 01:01:01.047050 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfa000000-0xfbffffff pref] Sep 13 01:01:01.047233 kernel: pci 0000:00:01.0: reg 0x14: [mem 0xfea50000-0xfea50fff] Sep 13 01:01:01.047410 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfea40000-0xfea4ffff pref] Sep 13 01:01:01.047586 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 13 01:01:01.047823 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.048027 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfea51000-0xfea51fff] Sep 13 01:01:01.048235 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.048413 kernel: pci 0000:00:02.1: reg 0x10: [mem 0xfea52000-0xfea52fff] Sep 13 01:01:01.048627 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.048810 kernel: pci 0000:00:02.2: reg 0x10: [mem 0xfea53000-0xfea53fff] Sep 13 01:01:01.050801 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.051058 kernel: pci 0000:00:02.3: reg 0x10: [mem 0xfea54000-0xfea54fff] Sep 13 01:01:01.051277 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.051457 kernel: pci 0000:00:02.4: reg 0x10: [mem 0xfea55000-0xfea55fff] Sep 13 01:01:01.052722 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.052943 kernel: pci 0000:00:02.5: reg 0x10: [mem 0xfea56000-0xfea56fff] Sep 13 01:01:01.053157 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.053354 kernel: pci 0000:00:02.6: reg 0x10: [mem 0xfea57000-0xfea57fff] Sep 13 01:01:01.053542 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Sep 13 01:01:01.054793 kernel: pci 0000:00:02.7: reg 0x10: [mem 0xfea58000-0xfea58fff] Sep 13 01:01:01.055007 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Sep 13 01:01:01.055224 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc0c0-0xc0df] Sep 13 01:01:01.055410 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfea59000-0xfea59fff] Sep 13 01:01:01.055594 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfd000000-0xfd003fff 64bit pref] Sep 13 01:01:01.056822 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfea00000-0xfea3ffff pref] Sep 13 01:01:01.057044 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Sep 13 01:01:01.057225 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Sep 13 01:01:01.057402 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfea5a000-0xfea5afff] Sep 13 01:01:01.057578 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfd004000-0xfd007fff 64bit pref] Sep 13 01:01:01.058823 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Sep 13 01:01:01.059034 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Sep 13 01:01:01.059232 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Sep 13 01:01:01.059414 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc0e0-0xc0ff] Sep 13 01:01:01.059616 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfea5b000-0xfea5bfff] Sep 13 01:01:01.059859 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Sep 13 01:01:01.060063 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Sep 13 01:01:01.060713 kernel: pci 0000:01:00.0: [1b36:000e] type 01 class 0x060400 Sep 13 01:01:01.060929 kernel: pci 0000:01:00.0: reg 0x10: [mem 0xfda00000-0xfda000ff 64bit] Sep 13 01:01:01.061119 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Sep 13 01:01:01.061308 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Sep 13 01:01:01.061497 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Sep 13 01:01:01.063763 kernel: pci_bus 0000:02: extended config space not accessible Sep 13 01:01:01.063996 kernel: pci 0000:02:01.0: [8086:25ab] type 00 class 0x088000 Sep 13 01:01:01.064201 kernel: pci 0000:02:01.0: reg 0x10: [mem 0xfd800000-0xfd80000f] Sep 13 01:01:01.064394 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Sep 13 01:01:01.064589 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Sep 13 01:01:01.065806 kernel: pci 0000:03:00.0: [1b36:000d] type 00 class 0x0c0330 Sep 13 01:01:01.066012 kernel: pci 0000:03:00.0: reg 0x10: [mem 0xfe800000-0xfe803fff 64bit] Sep 13 01:01:01.066194 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Sep 13 01:01:01.066369 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Sep 13 01:01:01.066545 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Sep 13 01:01:01.068805 kernel: pci 0000:04:00.0: [1af4:1044] type 00 class 0x00ff00 Sep 13 01:01:01.069020 kernel: pci 0000:04:00.0: reg 0x20: [mem 0xfca00000-0xfca03fff 64bit pref] Sep 13 01:01:01.069203 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Sep 13 01:01:01.069381 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Sep 13 01:01:01.070956 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Sep 13 01:01:01.071155 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Sep 13 01:01:01.071334 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Sep 13 01:01:01.071521 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Sep 13 01:01:01.072785 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Sep 13 01:01:01.072997 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Sep 13 01:01:01.073178 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Sep 13 01:01:01.073358 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Sep 13 01:01:01.073535 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Sep 13 01:01:01.075764 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Sep 13 01:01:01.075969 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Sep 13 01:01:01.076158 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Sep 13 01:01:01.076359 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Sep 13 01:01:01.076794 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Sep 13 01:01:01.077008 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Sep 13 01:01:01.077186 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Sep 13 01:01:01.077206 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 13 01:01:01.077221 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 13 01:01:01.077234 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 13 01:01:01.077255 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 13 01:01:01.077268 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Sep 13 01:01:01.077281 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Sep 13 01:01:01.077294 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Sep 13 01:01:01.077307 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Sep 13 01:01:01.077320 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Sep 13 01:01:01.077333 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Sep 13 01:01:01.077346 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Sep 13 01:01:01.077359 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Sep 13 01:01:01.077376 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Sep 13 01:01:01.077390 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Sep 13 01:01:01.077402 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Sep 13 01:01:01.077415 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Sep 13 01:01:01.077428 kernel: iommu: Default domain type: Translated Sep 13 01:01:01.077441 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 13 01:01:01.077453 kernel: PCI: Using ACPI for IRQ routing Sep 13 01:01:01.077466 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 13 01:01:01.077479 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 13 01:01:01.077496 kernel: e820: reserve RAM buffer [mem 0x7ffdc000-0x7fffffff] Sep 13 01:01:01.081741 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Sep 13 01:01:01.081955 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Sep 13 01:01:01.082136 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 13 01:01:01.082157 kernel: vgaarb: loaded Sep 13 01:01:01.082172 kernel: clocksource: Switched to clocksource kvm-clock Sep 13 01:01:01.082185 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 01:01:01.082198 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 01:01:01.082211 kernel: pnp: PnP ACPI init Sep 13 01:01:01.082407 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Sep 13 01:01:01.082430 kernel: pnp: PnP ACPI: found 5 devices Sep 13 01:01:01.082443 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 13 01:01:01.082457 kernel: NET: Registered PF_INET protocol family Sep 13 01:01:01.082470 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 13 01:01:01.082483 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 13 01:01:01.082496 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 01:01:01.082508 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 13 01:01:01.082528 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 13 01:01:01.082542 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 13 01:01:01.082555 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 01:01:01.082568 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 01:01:01.082581 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 01:01:01.082593 kernel: NET: Registered PF_XDP protocol family Sep 13 01:01:01.082790 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01-02] add_size 1000 Sep 13 01:01:01.082985 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Sep 13 01:01:01.083171 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Sep 13 01:01:01.083350 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Sep 13 01:01:01.083528 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Sep 13 01:01:01.085741 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Sep 13 01:01:01.085966 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Sep 13 01:01:01.086146 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Sep 13 01:01:01.086334 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Sep 13 01:01:01.086514 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Sep 13 01:01:01.086711 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Sep 13 01:01:01.086902 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Sep 13 01:01:01.087085 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Sep 13 01:01:01.087266 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Sep 13 01:01:01.087445 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Sep 13 01:01:01.089733 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Sep 13 01:01:01.089975 kernel: pci 0000:01:00.0: PCI bridge to [bus 02] Sep 13 01:01:01.090172 kernel: pci 0000:01:00.0: bridge window [mem 0xfd800000-0xfd9fffff] Sep 13 01:01:01.090353 kernel: pci 0000:00:02.0: PCI bridge to [bus 01-02] Sep 13 01:01:01.090533 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Sep 13 01:01:01.090736 kernel: pci 0000:00:02.0: bridge window [mem 0xfd800000-0xfdbfffff] Sep 13 01:01:01.090933 kernel: pci 0000:00:02.0: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Sep 13 01:01:01.091115 kernel: pci 0000:00:02.1: PCI bridge to [bus 03] Sep 13 01:01:01.092700 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Sep 13 01:01:01.092923 kernel: pci 0000:00:02.1: bridge window [mem 0xfe800000-0xfe9fffff] Sep 13 01:01:01.093102 kernel: pci 0000:00:02.1: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Sep 13 01:01:01.093278 kernel: pci 0000:00:02.2: PCI bridge to [bus 04] Sep 13 01:01:01.093452 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Sep 13 01:01:01.093662 kernel: pci 0000:00:02.2: bridge window [mem 0xfe600000-0xfe7fffff] Sep 13 01:01:01.093848 kernel: pci 0000:00:02.2: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Sep 13 01:01:01.094045 kernel: pci 0000:00:02.3: PCI bridge to [bus 05] Sep 13 01:01:01.094220 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Sep 13 01:01:01.094395 kernel: pci 0000:00:02.3: bridge window [mem 0xfe400000-0xfe5fffff] Sep 13 01:01:01.094570 kernel: pci 0000:00:02.3: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Sep 13 01:01:01.095823 kernel: pci 0000:00:02.4: PCI bridge to [bus 06] Sep 13 01:01:01.096035 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Sep 13 01:01:01.096217 kernel: pci 0000:00:02.4: bridge window [mem 0xfe200000-0xfe3fffff] Sep 13 01:01:01.096394 kernel: pci 0000:00:02.4: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Sep 13 01:01:01.096570 kernel: pci 0000:00:02.5: PCI bridge to [bus 07] Sep 13 01:01:01.096780 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Sep 13 01:01:01.096973 kernel: pci 0000:00:02.5: bridge window [mem 0xfe000000-0xfe1fffff] Sep 13 01:01:01.097149 kernel: pci 0000:00:02.5: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Sep 13 01:01:01.097345 kernel: pci 0000:00:02.6: PCI bridge to [bus 08] Sep 13 01:01:01.097519 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Sep 13 01:01:01.099756 kernel: pci 0000:00:02.6: bridge window [mem 0xfde00000-0xfdffffff] Sep 13 01:01:01.099964 kernel: pci 0000:00:02.6: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Sep 13 01:01:01.100148 kernel: pci 0000:00:02.7: PCI bridge to [bus 09] Sep 13 01:01:01.100327 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Sep 13 01:01:01.100505 kernel: pci 0000:00:02.7: bridge window [mem 0xfdc00000-0xfddfffff] Sep 13 01:01:01.101735 kernel: pci 0000:00:02.7: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Sep 13 01:01:01.101926 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 13 01:01:01.102090 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 13 01:01:01.102250 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 13 01:01:01.102419 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Sep 13 01:01:01.102577 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Sep 13 01:01:01.104784 kernel: pci_bus 0000:00: resource 9 [mem 0x20c0000000-0x28bfffffff window] Sep 13 01:01:01.105002 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Sep 13 01:01:01.105179 kernel: pci_bus 0000:01: resource 1 [mem 0xfd800000-0xfdbfffff] Sep 13 01:01:01.105351 kernel: pci_bus 0000:01: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Sep 13 01:01:01.105535 kernel: pci_bus 0000:02: resource 1 [mem 0xfd800000-0xfd9fffff] Sep 13 01:01:01.105789 kernel: pci_bus 0000:03: resource 0 [io 0x2000-0x2fff] Sep 13 01:01:01.105978 kernel: pci_bus 0000:03: resource 1 [mem 0xfe800000-0xfe9fffff] Sep 13 01:01:01.106146 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Sep 13 01:01:01.106326 kernel: pci_bus 0000:04: resource 0 [io 0x3000-0x3fff] Sep 13 01:01:01.106493 kernel: pci_bus 0000:04: resource 1 [mem 0xfe600000-0xfe7fffff] Sep 13 01:01:01.107231 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Sep 13 01:01:01.107440 kernel: pci_bus 0000:05: resource 0 [io 0x4000-0x4fff] Sep 13 01:01:01.107655 kernel: pci_bus 0000:05: resource 1 [mem 0xfe400000-0xfe5fffff] Sep 13 01:01:01.107827 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Sep 13 01:01:01.108029 kernel: pci_bus 0000:06: resource 0 [io 0x5000-0x5fff] Sep 13 01:01:01.108197 kernel: pci_bus 0000:06: resource 1 [mem 0xfe200000-0xfe3fffff] Sep 13 01:01:01.108364 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Sep 13 01:01:01.108550 kernel: pci_bus 0000:07: resource 0 [io 0x6000-0x6fff] Sep 13 01:01:01.108758 kernel: pci_bus 0000:07: resource 1 [mem 0xfe000000-0xfe1fffff] Sep 13 01:01:01.108945 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Sep 13 01:01:01.109123 kernel: pci_bus 0000:08: resource 0 [io 0x7000-0x7fff] Sep 13 01:01:01.109292 kernel: pci_bus 0000:08: resource 1 [mem 0xfde00000-0xfdffffff] Sep 13 01:01:01.109460 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Sep 13 01:01:01.109684 kernel: pci_bus 0000:09: resource 0 [io 0x8000-0x8fff] Sep 13 01:01:01.109862 kernel: pci_bus 0000:09: resource 1 [mem 0xfdc00000-0xfddfffff] Sep 13 01:01:01.110050 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Sep 13 01:01:01.110073 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Sep 13 01:01:01.110087 kernel: PCI: CLS 0 bytes, default 64 Sep 13 01:01:01.110101 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 13 01:01:01.110115 kernel: software IO TLB: mapped [mem 0x0000000079800000-0x000000007d800000] (64MB) Sep 13 01:01:01.110129 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 13 01:01:01.110144 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Sep 13 01:01:01.110157 kernel: Initialise system trusted keyrings Sep 13 01:01:01.110178 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 13 01:01:01.110192 kernel: Key type asymmetric registered Sep 13 01:01:01.110205 kernel: Asymmetric key parser 'x509' registered Sep 13 01:01:01.110219 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 13 01:01:01.110233 kernel: io scheduler mq-deadline registered Sep 13 01:01:01.110246 kernel: io scheduler kyber registered Sep 13 01:01:01.110260 kernel: io scheduler bfq registered Sep 13 01:01:01.110438 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Sep 13 01:01:01.110646 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Sep 13 01:01:01.110836 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.111030 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Sep 13 01:01:01.111207 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Sep 13 01:01:01.111392 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.111589 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Sep 13 01:01:01.111817 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Sep 13 01:01:01.112017 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.112196 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Sep 13 01:01:01.112377 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Sep 13 01:01:01.112552 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.112763 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Sep 13 01:01:01.112956 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Sep 13 01:01:01.113142 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.113322 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Sep 13 01:01:01.113499 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Sep 13 01:01:01.113723 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.113918 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Sep 13 01:01:01.114096 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Sep 13 01:01:01.114279 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.114466 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Sep 13 01:01:01.114669 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Sep 13 01:01:01.114848 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Sep 13 01:01:01.114870 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 13 01:01:01.114896 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Sep 13 01:01:01.114920 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Sep 13 01:01:01.114934 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 01:01:01.114948 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 13 01:01:01.114962 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 13 01:01:01.114976 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 13 01:01:01.114990 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 13 01:01:01.115004 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 13 01:01:01.115200 kernel: rtc_cmos 00:03: RTC can wake from S4 Sep 13 01:01:01.115373 kernel: rtc_cmos 00:03: registered as rtc0 Sep 13 01:01:01.115548 kernel: rtc_cmos 00:03: setting system clock to 2025-09-13T01:01:00 UTC (1757725260) Sep 13 01:01:01.115765 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Sep 13 01:01:01.115786 kernel: intel_pstate: CPU model not supported Sep 13 01:01:01.115800 kernel: NET: Registered PF_INET6 protocol family Sep 13 01:01:01.115814 kernel: Segment Routing with IPv6 Sep 13 01:01:01.115827 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 01:01:01.115841 kernel: NET: Registered PF_PACKET protocol family Sep 13 01:01:01.115854 kernel: Key type dns_resolver registered Sep 13 01:01:01.115878 kernel: IPI shorthand broadcast: enabled Sep 13 01:01:01.115906 kernel: sched_clock: Marking stable (1242003381, 231769073)->(1602518259, -128745805) Sep 13 01:01:01.115920 kernel: registered taskstats version 1 Sep 13 01:01:01.115934 kernel: Loading compiled-in X.509 certificates Sep 13 01:01:01.115948 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.106-flatcar: 1274e0c573ac8d09163d6bc6d1ee1445fb2f8cc6' Sep 13 01:01:01.115961 kernel: Key type .fscrypt registered Sep 13 01:01:01.115974 kernel: Key type fscrypt-provisioning registered Sep 13 01:01:01.115987 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 13 01:01:01.116001 kernel: ima: Allocated hash algorithm: sha1 Sep 13 01:01:01.116020 kernel: ima: No architecture policies found Sep 13 01:01:01.116034 kernel: clk: Disabling unused clocks Sep 13 01:01:01.116047 kernel: Freeing unused kernel image (initmem) memory: 42884K Sep 13 01:01:01.116061 kernel: Write protecting the kernel read-only data: 36864k Sep 13 01:01:01.116074 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Sep 13 01:01:01.116088 kernel: Run /init as init process Sep 13 01:01:01.116101 kernel: with arguments: Sep 13 01:01:01.116114 kernel: /init Sep 13 01:01:01.116127 kernel: with environment: Sep 13 01:01:01.116145 kernel: HOME=/ Sep 13 01:01:01.116158 kernel: TERM=linux Sep 13 01:01:01.116172 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 01:01:01.116188 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 13 01:01:01.116205 systemd[1]: Detected virtualization kvm. Sep 13 01:01:01.116219 systemd[1]: Detected architecture x86-64. Sep 13 01:01:01.116233 systemd[1]: Running in initrd. Sep 13 01:01:01.116246 systemd[1]: No hostname configured, using default hostname. Sep 13 01:01:01.116265 systemd[1]: Hostname set to . Sep 13 01:01:01.116280 systemd[1]: Initializing machine ID from VM UUID. Sep 13 01:01:01.116294 systemd[1]: Queued start job for default target initrd.target. Sep 13 01:01:01.116308 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:01:01.116322 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:01:01.116337 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 13 01:01:01.116351 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 13 01:01:01.116366 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 13 01:01:01.116386 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 13 01:01:01.116402 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 13 01:01:01.116417 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 13 01:01:01.116431 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:01:01.116445 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:01:01.116459 systemd[1]: Reached target paths.target - Path Units. Sep 13 01:01:01.116478 systemd[1]: Reached target slices.target - Slice Units. Sep 13 01:01:01.116493 systemd[1]: Reached target swap.target - Swaps. Sep 13 01:01:01.116508 systemd[1]: Reached target timers.target - Timer Units. Sep 13 01:01:01.116522 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:01:01.116536 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:01:01.116551 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 13 01:01:01.116565 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 13 01:01:01.116579 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:01:01.116593 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 13 01:01:01.116642 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:01:01.116657 systemd[1]: Reached target sockets.target - Socket Units. Sep 13 01:01:01.116672 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 13 01:01:01.116686 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 13 01:01:01.116700 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 13 01:01:01.116715 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 01:01:01.116729 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 13 01:01:01.116743 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 13 01:01:01.116757 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:01:01.116818 systemd-journald[201]: Collecting audit messages is disabled. Sep 13 01:01:01.116852 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 13 01:01:01.116867 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:01:01.116899 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 01:01:01.116916 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 13 01:01:01.116930 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 01:01:01.116944 kernel: Bridge firewalling registered Sep 13 01:01:01.116958 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 13 01:01:01.116979 systemd-journald[201]: Journal started Sep 13 01:01:01.117010 systemd-journald[201]: Runtime Journal (/run/log/journal/a357e97034d54349a33834319d8eb5e4) is 4.7M, max 38.0M, 33.2M free. Sep 13 01:01:01.049350 systemd-modules-load[202]: Inserted module 'overlay' Sep 13 01:01:01.160374 systemd[1]: Started systemd-journald.service - Journal Service. Sep 13 01:01:01.092790 systemd-modules-load[202]: Inserted module 'br_netfilter' Sep 13 01:01:01.162617 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:01:01.163840 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 13 01:01:01.177785 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:01:01.179791 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:01:01.185781 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 13 01:01:01.188411 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 13 01:01:01.206166 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:01:01.209192 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:01:01.218242 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:01:01.227847 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 13 01:01:01.228970 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:01:01.232811 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 13 01:01:01.253260 dracut-cmdline[236]: dracut-dracut-053 Sep 13 01:01:01.257720 dracut-cmdline[236]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=openstack flatcar.autologin verity.usrhash=2945e6465d436b7d1da8a9350a0544af0bd9aec821cd06987451d5e1d3071534 Sep 13 01:01:01.266730 systemd-resolved[235]: Positive Trust Anchors: Sep 13 01:01:01.266751 systemd-resolved[235]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 01:01:01.266807 systemd-resolved[235]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 13 01:01:01.271931 systemd-resolved[235]: Defaulting to hostname 'linux'. Sep 13 01:01:01.273661 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 13 01:01:01.275570 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:01:01.368729 kernel: SCSI subsystem initialized Sep 13 01:01:01.380635 kernel: Loading iSCSI transport class v2.0-870. Sep 13 01:01:01.394618 kernel: iscsi: registered transport (tcp) Sep 13 01:01:01.420969 kernel: iscsi: registered transport (qla4xxx) Sep 13 01:01:01.421035 kernel: QLogic iSCSI HBA Driver Sep 13 01:01:01.476413 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 13 01:01:01.483856 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 13 01:01:01.522020 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 01:01:01.522115 kernel: device-mapper: uevent: version 1.0.3 Sep 13 01:01:01.525097 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 13 01:01:01.576706 kernel: raid6: sse2x4 gen() 14031 MB/s Sep 13 01:01:01.592642 kernel: raid6: sse2x2 gen() 9736 MB/s Sep 13 01:01:01.611256 kernel: raid6: sse2x1 gen() 10286 MB/s Sep 13 01:01:01.611340 kernel: raid6: using algorithm sse2x4 gen() 14031 MB/s Sep 13 01:01:01.630301 kernel: raid6: .... xor() 7645 MB/s, rmw enabled Sep 13 01:01:01.630393 kernel: raid6: using ssse3x2 recovery algorithm Sep 13 01:01:01.657673 kernel: xor: automatically using best checksumming function avx Sep 13 01:01:01.850643 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 13 01:01:01.867011 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:01:01.873909 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:01:01.903490 systemd-udevd[419]: Using default interface naming scheme 'v255'. Sep 13 01:01:01.911131 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:01:01.919845 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 13 01:01:01.941726 dracut-pre-trigger[427]: rd.md=0: removing MD RAID activation Sep 13 01:01:01.987550 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:01:01.994848 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 13 01:01:02.122047 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:01:02.128789 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 13 01:01:02.158294 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 13 01:01:02.160899 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:01:02.163686 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:01:02.165075 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 13 01:01:02.173823 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 13 01:01:02.206289 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:01:02.255647 kernel: virtio_blk virtio1: 2/0/0 default/read/poll queues Sep 13 01:01:02.276776 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 01:01:02.281624 kernel: virtio_blk virtio1: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Sep 13 01:01:02.315626 kernel: libata version 3.00 loaded. Sep 13 01:01:02.314835 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:01:02.333315 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 01:01:02.333366 kernel: GPT:17805311 != 125829119 Sep 13 01:01:02.333385 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 01:01:02.333404 kernel: GPT:17805311 != 125829119 Sep 13 01:01:02.333430 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 01:01:02.333448 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 01:01:02.333466 kernel: ahci 0000:00:1f.2: version 3.0 Sep 13 01:01:02.333779 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Sep 13 01:01:02.315050 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:01:02.333077 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:01:02.359725 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Sep 13 01:01:02.360048 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Sep 13 01:01:02.360316 kernel: AVX version of gcm_enc/dec engaged. Sep 13 01:01:02.360349 kernel: AES CTR mode by8 optimization enabled Sep 13 01:01:02.360376 kernel: scsi host0: ahci Sep 13 01:01:02.361656 kernel: scsi host1: ahci Sep 13 01:01:02.362992 kernel: scsi host2: ahci Sep 13 01:01:02.363238 kernel: scsi host3: ahci Sep 13 01:01:02.363471 kernel: scsi host4: ahci Sep 13 01:01:02.334087 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:01:02.334375 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:01:02.336251 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:01:02.363437 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:01:02.376628 kernel: ACPI: bus type USB registered Sep 13 01:01:02.380635 kernel: scsi host5: ahci Sep 13 01:01:02.382647 kernel: usbcore: registered new interface driver usbfs Sep 13 01:01:02.389627 kernel: usbcore: registered new interface driver hub Sep 13 01:01:02.411114 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b100 irq 38 Sep 13 01:01:02.411191 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b180 irq 38 Sep 13 01:01:02.411212 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b200 irq 38 Sep 13 01:01:02.411247 kernel: usbcore: registered new device driver usb Sep 13 01:01:02.411266 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b280 irq 38 Sep 13 01:01:02.411283 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b300 irq 38 Sep 13 01:01:02.411301 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (475) Sep 13 01:01:02.411320 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea5b000 port 0xfea5b380 irq 38 Sep 13 01:01:02.411337 kernel: BTRFS: device fsid fa70a3b0-3d47-4508-bba0-9fa4607626aa devid 1 transid 36 /dev/vda3 scanned by (udev-worker) (466) Sep 13 01:01:02.451061 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 13 01:01:02.507243 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:01:02.520628 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 13 01:01:02.533098 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 13 01:01:02.539176 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 13 01:01:02.540027 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 13 01:01:02.558948 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 13 01:01:02.562797 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 13 01:01:02.570675 disk-uuid[558]: Primary Header is updated. Sep 13 01:01:02.570675 disk-uuid[558]: Secondary Entries is updated. Sep 13 01:01:02.570675 disk-uuid[558]: Secondary Header is updated. Sep 13 01:01:02.576655 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 01:01:02.585451 kernel: GPT:disk_guids don't match. Sep 13 01:01:02.585541 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 01:01:02.585563 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 01:01:02.598647 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 01:01:02.602822 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:01:02.722926 kernel: ata4: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.722995 kernel: ata1: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.723634 kernel: ata6: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.726433 kernel: ata2: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.733631 kernel: ata3: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.736619 kernel: ata5: SATA link down (SStatus 0 SControl 300) Sep 13 01:01:02.798629 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Sep 13 01:01:02.803635 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 1 Sep 13 01:01:02.808623 kernel: xhci_hcd 0000:03:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Sep 13 01:01:02.815313 kernel: xhci_hcd 0000:03:00.0: xHCI Host Controller Sep 13 01:01:02.815565 kernel: xhci_hcd 0000:03:00.0: new USB bus registered, assigned bus number 2 Sep 13 01:01:02.817620 kernel: xhci_hcd 0000:03:00.0: Host supports USB 3.0 SuperSpeed Sep 13 01:01:02.823618 kernel: hub 1-0:1.0: USB hub found Sep 13 01:01:02.825622 kernel: hub 1-0:1.0: 4 ports detected Sep 13 01:01:02.831642 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Sep 13 01:01:02.834357 kernel: hub 2-0:1.0: USB hub found Sep 13 01:01:02.834638 kernel: hub 2-0:1.0: 4 ports detected Sep 13 01:01:03.069685 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Sep 13 01:01:03.212947 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 13 01:01:03.221081 kernel: usbcore: registered new interface driver usbhid Sep 13 01:01:03.221143 kernel: usbhid: USB HID core driver Sep 13 01:01:03.230007 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:03:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input2 Sep 13 01:01:03.230065 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:03:00.0-1/input0 Sep 13 01:01:03.594660 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 01:01:03.596004 disk-uuid[559]: The operation has completed successfully. Sep 13 01:01:03.647157 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 01:01:03.647393 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 13 01:01:03.676961 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 13 01:01:03.681137 sh[585]: Success Sep 13 01:01:03.699655 kernel: device-mapper: verity: sha256 using implementation "sha256-avx" Sep 13 01:01:03.768078 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 13 01:01:03.772129 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 13 01:01:03.774628 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 13 01:01:03.795641 kernel: BTRFS info (device dm-0): first mount of filesystem fa70a3b0-3d47-4508-bba0-9fa4607626aa Sep 13 01:01:03.798693 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:01:03.798743 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 13 01:01:03.800959 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 13 01:01:03.802774 kernel: BTRFS info (device dm-0): using free space tree Sep 13 01:01:03.813695 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 13 01:01:03.815204 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 13 01:01:03.826832 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 13 01:01:03.830547 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 13 01:01:03.845853 kernel: BTRFS info (device vda6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:01:03.850250 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:01:03.850303 kernel: BTRFS info (device vda6): using free space tree Sep 13 01:01:03.857620 kernel: BTRFS info (device vda6): auto enabling async discard Sep 13 01:01:03.873129 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 13 01:01:03.876462 kernel: BTRFS info (device vda6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:01:03.883201 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 13 01:01:03.892156 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 13 01:01:04.011733 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:01:04.022942 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 13 01:01:04.047508 ignition[681]: Ignition 2.19.0 Sep 13 01:01:04.047535 ignition[681]: Stage: fetch-offline Sep 13 01:01:04.047653 ignition[681]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:04.047680 ignition[681]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:04.051504 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:01:04.047955 ignition[681]: parsed url from cmdline: "" Sep 13 01:01:04.047962 ignition[681]: no config URL provided Sep 13 01:01:04.047972 ignition[681]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 01:01:04.047990 ignition[681]: no config at "/usr/lib/ignition/user.ign" Sep 13 01:01:04.048000 ignition[681]: failed to fetch config: resource requires networking Sep 13 01:01:04.048378 ignition[681]: Ignition finished successfully Sep 13 01:01:04.066851 systemd-networkd[771]: lo: Link UP Sep 13 01:01:04.066867 systemd-networkd[771]: lo: Gained carrier Sep 13 01:01:04.069776 systemd-networkd[771]: Enumeration completed Sep 13 01:01:04.070737 systemd-networkd[771]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 13 01:01:04.070743 systemd-networkd[771]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:01:04.072435 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 13 01:01:04.073005 systemd-networkd[771]: eth0: Link UP Sep 13 01:01:04.073012 systemd-networkd[771]: eth0: Gained carrier Sep 13 01:01:04.073029 systemd-networkd[771]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 13 01:01:04.076133 systemd[1]: Reached target network.target - Network. Sep 13 01:01:04.081008 systemd-networkd[771]: eth0: DHCPv4 address 10.230.35.122/30, gateway 10.230.35.121 acquired from 10.230.35.121 Sep 13 01:01:04.084928 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 13 01:01:04.109215 ignition[775]: Ignition 2.19.0 Sep 13 01:01:04.109238 ignition[775]: Stage: fetch Sep 13 01:01:04.109510 ignition[775]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:04.109531 ignition[775]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:04.109707 ignition[775]: parsed url from cmdline: "" Sep 13 01:01:04.109714 ignition[775]: no config URL provided Sep 13 01:01:04.109724 ignition[775]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 01:01:04.109740 ignition[775]: no config at "/usr/lib/ignition/user.ign" Sep 13 01:01:04.109967 ignition[775]: GET http://169.254.169.254/openstack/latest/user_data: attempt #1 Sep 13 01:01:04.110088 ignition[775]: config drive ("/dev/disk/by-label/config-2") not found. Waiting... Sep 13 01:01:04.110138 ignition[775]: config drive ("/dev/disk/by-label/CONFIG-2") not found. Waiting... Sep 13 01:01:04.126001 ignition[775]: GET result: OK Sep 13 01:01:04.126614 ignition[775]: parsing config with SHA512: 848c71351d25ef96efa21eb6a18bc242b7ce4d7702aee31994d13965955156383d068fa937505061af1368aef44b3d5c9f12c189800a10b50996938750c6af03 Sep 13 01:01:04.133249 unknown[775]: fetched base config from "system" Sep 13 01:01:04.133270 unknown[775]: fetched base config from "system" Sep 13 01:01:04.134128 ignition[775]: fetch: fetch complete Sep 13 01:01:04.133285 unknown[775]: fetched user config from "openstack" Sep 13 01:01:04.134137 ignition[775]: fetch: fetch passed Sep 13 01:01:04.136103 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 13 01:01:04.134205 ignition[775]: Ignition finished successfully Sep 13 01:01:04.159757 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 13 01:01:04.178877 ignition[782]: Ignition 2.19.0 Sep 13 01:01:04.178900 ignition[782]: Stage: kargs Sep 13 01:01:04.179146 ignition[782]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:04.179167 ignition[782]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:04.182530 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 13 01:01:04.180392 ignition[782]: kargs: kargs passed Sep 13 01:01:04.180471 ignition[782]: Ignition finished successfully Sep 13 01:01:04.188871 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 13 01:01:04.210222 ignition[788]: Ignition 2.19.0 Sep 13 01:01:04.210245 ignition[788]: Stage: disks Sep 13 01:01:04.210513 ignition[788]: no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:04.213071 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 13 01:01:04.210534 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:04.211835 ignition[788]: disks: disks passed Sep 13 01:01:04.211907 ignition[788]: Ignition finished successfully Sep 13 01:01:04.217290 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 13 01:01:04.218105 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 13 01:01:04.219774 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 13 01:01:04.221324 systemd[1]: Reached target sysinit.target - System Initialization. Sep 13 01:01:04.222719 systemd[1]: Reached target basic.target - Basic System. Sep 13 01:01:04.229911 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 13 01:01:04.249987 systemd-fsck[796]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Sep 13 01:01:04.254943 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 13 01:01:04.262781 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 13 01:01:04.388647 kernel: EXT4-fs (vda9): mounted filesystem 3a3ecd49-b269-4fcb-bb61-e2994e1868ee r/w with ordered data mode. Quota mode: none. Sep 13 01:01:04.388756 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 13 01:01:04.390212 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 13 01:01:04.396707 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:01:04.399647 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 13 01:01:04.402103 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 13 01:01:04.408791 systemd[1]: Starting flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent... Sep 13 01:01:04.420819 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (804) Sep 13 01:01:04.420866 kernel: BTRFS info (device vda6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:01:04.420886 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:01:04.420905 kernel: BTRFS info (device vda6): using free space tree Sep 13 01:01:04.420923 kernel: BTRFS info (device vda6): auto enabling async discard Sep 13 01:01:04.410743 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 01:01:04.410784 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:01:04.424160 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 13 01:01:04.429470 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:01:04.440518 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 13 01:01:04.505690 initrd-setup-root[832]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 01:01:04.514807 initrd-setup-root[839]: cut: /sysroot/etc/group: No such file or directory Sep 13 01:01:04.522751 initrd-setup-root[846]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 01:01:04.530508 initrd-setup-root[853]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 01:01:04.641913 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 13 01:01:04.648722 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 13 01:01:04.667773 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 13 01:01:04.680652 kernel: BTRFS info (device vda6): last unmount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:01:04.698675 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 13 01:01:04.712535 ignition[922]: INFO : Ignition 2.19.0 Sep 13 01:01:04.712535 ignition[922]: INFO : Stage: mount Sep 13 01:01:04.715017 ignition[922]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:04.715017 ignition[922]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:04.715017 ignition[922]: INFO : mount: mount passed Sep 13 01:01:04.715017 ignition[922]: INFO : Ignition finished successfully Sep 13 01:01:04.716935 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 13 01:01:04.794490 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 13 01:01:05.970029 systemd-networkd[771]: eth0: Gained IPv6LL Sep 13 01:01:07.477009 systemd-networkd[771]: eth0: Ignoring DHCPv6 address 2a02:1348:179:88de:24:19ff:fee6:237a/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:88de:24:19ff:fee6:237a/64 assigned by NDisc. Sep 13 01:01:07.477031 systemd-networkd[771]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Sep 13 01:01:11.574812 coreos-metadata[806]: Sep 13 01:01:11.574 WARN failed to locate config-drive, using the metadata service API instead Sep 13 01:01:11.599866 coreos-metadata[806]: Sep 13 01:01:11.599 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Sep 13 01:01:11.616747 coreos-metadata[806]: Sep 13 01:01:11.616 INFO Fetch successful Sep 13 01:01:11.617679 coreos-metadata[806]: Sep 13 01:01:11.617 INFO wrote hostname srv-ngi99.gb1.brightbox.com to /sysroot/etc/hostname Sep 13 01:01:11.619357 systemd[1]: flatcar-openstack-hostname.service: Deactivated successfully. Sep 13 01:01:11.619516 systemd[1]: Finished flatcar-openstack-hostname.service - Flatcar OpenStack Metadata Hostname Agent. Sep 13 01:01:11.628742 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 13 01:01:11.638735 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 13 01:01:11.667669 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (937) Sep 13 01:01:11.673758 kernel: BTRFS info (device vda6): first mount of filesystem 94088f30-ba7d-4694-bba6-875359d7b417 Sep 13 01:01:11.673841 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 13 01:01:11.673864 kernel: BTRFS info (device vda6): using free space tree Sep 13 01:01:11.679648 kernel: BTRFS info (device vda6): auto enabling async discard Sep 13 01:01:11.683458 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 13 01:01:11.723276 ignition[954]: INFO : Ignition 2.19.0 Sep 13 01:01:11.727276 ignition[954]: INFO : Stage: files Sep 13 01:01:11.727276 ignition[954]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:11.727276 ignition[954]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:11.730069 ignition[954]: DEBUG : files: compiled without relabeling support, skipping Sep 13 01:01:11.731132 ignition[954]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 01:01:11.731132 ignition[954]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 01:01:11.750095 ignition[954]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 01:01:11.751284 ignition[954]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 01:01:11.751284 ignition[954]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 01:01:11.750969 unknown[954]: wrote ssh authorized keys file for user: core Sep 13 01:01:11.755118 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 13 01:01:11.755118 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 13 01:01:11.755118 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 13 01:01:11.755118 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 13 01:01:11.984799 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 13 01:01:12.273760 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 13 01:01:12.273760 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 01:01:12.277733 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 13 01:01:12.524324 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 01:01:12.792231 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 01:01:12.802831 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Sep 13 01:01:13.075754 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Sep 13 01:01:14.211359 ignition[954]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 01:01:14.211359 ignition[954]: INFO : files: op(d): [started] processing unit "containerd.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(d): [finished] processing unit "containerd.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 01:01:14.214448 ignition[954]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:01:14.214448 ignition[954]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 01:01:14.214448 ignition[954]: INFO : files: files passed Sep 13 01:01:14.214448 ignition[954]: INFO : Ignition finished successfully Sep 13 01:01:14.216389 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 13 01:01:14.231784 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 13 01:01:14.234983 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 13 01:01:14.248151 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 01:01:14.248380 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 13 01:01:14.260534 initrd-setup-root-after-ignition[984]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:01:14.260534 initrd-setup-root-after-ignition[984]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:01:14.263903 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 01:01:14.266806 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:01:14.267900 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 13 01:01:14.280353 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 13 01:01:14.326898 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 01:01:14.327100 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 13 01:01:14.330449 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 13 01:01:14.332429 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 13 01:01:14.334081 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 13 01:01:14.335827 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 13 01:01:14.361823 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:01:14.368838 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 13 01:01:14.385387 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:01:14.386416 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:01:14.388121 systemd[1]: Stopped target timers.target - Timer Units. Sep 13 01:01:14.389704 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 01:01:14.389908 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 13 01:01:14.391719 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 13 01:01:14.392715 systemd[1]: Stopped target basic.target - Basic System. Sep 13 01:01:14.394135 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 13 01:01:14.395695 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 13 01:01:14.397122 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 13 01:01:14.398731 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 13 01:01:14.400315 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 13 01:01:14.401941 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 13 01:01:14.403478 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 13 01:01:14.405056 systemd[1]: Stopped target swap.target - Swaps. Sep 13 01:01:14.406674 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 01:01:14.406952 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 13 01:01:14.408488 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:01:14.409503 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:01:14.411074 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 13 01:01:14.411267 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:01:14.412869 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 01:01:14.413119 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 13 01:01:14.415047 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 01:01:14.415222 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 13 01:01:14.416901 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 01:01:14.417075 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 13 01:01:14.434480 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 13 01:01:14.435313 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 01:01:14.435643 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:01:14.438895 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 13 01:01:14.440310 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 01:01:14.440798 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:01:14.443893 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 01:01:14.444170 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 13 01:01:14.452634 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 01:01:14.452871 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 13 01:01:14.467194 ignition[1008]: INFO : Ignition 2.19.0 Sep 13 01:01:14.468533 ignition[1008]: INFO : Stage: umount Sep 13 01:01:14.470660 ignition[1008]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 13 01:01:14.470660 ignition[1008]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/openstack" Sep 13 01:01:14.473046 ignition[1008]: INFO : umount: umount passed Sep 13 01:01:14.473046 ignition[1008]: INFO : Ignition finished successfully Sep 13 01:01:14.474865 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 01:01:14.475083 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 13 01:01:14.476848 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 01:01:14.476976 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 13 01:01:14.478206 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 01:01:14.478276 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 13 01:01:14.479649 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 13 01:01:14.479740 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 13 01:01:14.483325 systemd[1]: Stopped target network.target - Network. Sep 13 01:01:14.483947 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 01:01:14.484029 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 13 01:01:14.484840 systemd[1]: Stopped target paths.target - Path Units. Sep 13 01:01:14.486221 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 01:01:14.490103 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:01:14.491251 systemd[1]: Stopped target slices.target - Slice Units. Sep 13 01:01:14.492836 systemd[1]: Stopped target sockets.target - Socket Units. Sep 13 01:01:14.494259 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 01:01:14.494341 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 13 01:01:14.495699 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 01:01:14.495802 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 13 01:01:14.497298 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 01:01:14.497374 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 13 01:01:14.498772 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 13 01:01:14.498846 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 13 01:01:14.500425 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 13 01:01:14.502006 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 13 01:01:14.505258 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 01:01:14.505729 systemd-networkd[771]: eth0: DHCPv6 lease lost Sep 13 01:01:14.507693 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 01:01:14.507846 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 13 01:01:14.510238 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 01:01:14.510388 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 13 01:01:14.511923 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 01:01:14.512226 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 13 01:01:14.521652 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 01:01:14.522264 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:01:14.523392 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 01:01:14.523516 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 13 01:01:14.531786 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 13 01:01:14.532516 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 01:01:14.532633 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 13 01:01:14.534206 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 01:01:14.534291 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:01:14.537033 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 01:01:14.537110 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 13 01:01:14.538826 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 13 01:01:14.538910 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:01:14.542213 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:01:14.554239 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 01:01:14.555509 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:01:14.561730 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 01:01:14.561907 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 13 01:01:14.562753 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 01:01:14.562814 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:01:14.563643 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 01:01:14.563719 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 13 01:01:14.565284 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 01:01:14.565361 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 13 01:01:14.566830 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 01:01:14.566920 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 13 01:01:14.570818 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 13 01:01:14.572151 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 13 01:01:14.572236 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:01:14.574716 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 01:01:14.574797 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:01:14.576734 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 01:01:14.578177 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 13 01:01:14.594504 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 01:01:14.594750 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 13 01:01:14.596590 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 13 01:01:14.608816 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 13 01:01:14.634068 systemd[1]: Switching root. Sep 13 01:01:14.679615 systemd-journald[201]: Received SIGTERM from PID 1 (systemd). Sep 13 01:01:14.679775 systemd-journald[201]: Journal stopped Sep 13 01:01:16.377821 kernel: SELinux: policy capability network_peer_controls=1 Sep 13 01:01:16.377926 kernel: SELinux: policy capability open_perms=1 Sep 13 01:01:16.377970 kernel: SELinux: policy capability extended_socket_class=1 Sep 13 01:01:16.379673 kernel: SELinux: policy capability always_check_network=0 Sep 13 01:01:16.379702 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 13 01:01:16.379742 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 13 01:01:16.379775 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 13 01:01:16.379796 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 13 01:01:16.379823 kernel: audit: type=1403 audit(1757725275.175:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 13 01:01:16.384241 systemd[1]: Successfully loaded SELinux policy in 49.625ms. Sep 13 01:01:16.384304 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 22.598ms. Sep 13 01:01:16.384344 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 13 01:01:16.384368 systemd[1]: Detected virtualization kvm. Sep 13 01:01:16.384389 systemd[1]: Detected architecture x86-64. Sep 13 01:01:16.384410 systemd[1]: Detected first boot. Sep 13 01:01:16.384438 systemd[1]: Hostname set to . Sep 13 01:01:16.384460 systemd[1]: Initializing machine ID from VM UUID. Sep 13 01:01:16.384481 zram_generator::config[1073]: No configuration found. Sep 13 01:01:16.384520 systemd[1]: Populated /etc with preset unit settings. Sep 13 01:01:16.384556 systemd[1]: Queued start job for default target multi-user.target. Sep 13 01:01:16.384579 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 13 01:01:16.384619 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 13 01:01:16.384643 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 13 01:01:16.384665 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 13 01:01:16.384700 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 13 01:01:16.384724 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 13 01:01:16.384745 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 13 01:01:16.384779 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 13 01:01:16.384808 systemd[1]: Created slice user.slice - User and Session Slice. Sep 13 01:01:16.384836 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 13 01:01:16.384859 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 13 01:01:16.384880 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 13 01:01:16.384900 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 13 01:01:16.384922 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 13 01:01:16.384956 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 13 01:01:16.384980 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 13 01:01:16.385014 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 13 01:01:16.385037 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 13 01:01:16.385059 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 13 01:01:16.385109 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 13 01:01:16.385133 systemd[1]: Reached target slices.target - Slice Units. Sep 13 01:01:16.385177 systemd[1]: Reached target swap.target - Swaps. Sep 13 01:01:16.385200 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 13 01:01:16.385222 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 13 01:01:16.385243 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 13 01:01:16.385270 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 13 01:01:16.385292 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 13 01:01:16.385313 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 13 01:01:16.385339 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 13 01:01:16.385372 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 13 01:01:16.385395 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 13 01:01:16.385422 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 13 01:01:16.385444 systemd[1]: Mounting media.mount - External Media Directory... Sep 13 01:01:16.385465 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:16.385486 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 13 01:01:16.385520 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 13 01:01:16.385542 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 13 01:01:16.385579 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 13 01:01:16.392056 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:01:16.392103 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 13 01:01:16.392134 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 13 01:01:16.392156 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:01:16.392178 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 13 01:01:16.392210 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:01:16.392238 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 13 01:01:16.392260 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:01:16.392298 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 01:01:16.392328 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Sep 13 01:01:16.392351 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Sep 13 01:01:16.392372 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 13 01:01:16.392394 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 13 01:01:16.392415 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 13 01:01:16.392436 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 13 01:01:16.392472 kernel: fuse: init (API version 7.39) Sep 13 01:01:16.392544 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 13 01:01:16.392586 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:16.398800 kernel: loop: module loaded Sep 13 01:01:16.398876 systemd-journald[1177]: Collecting audit messages is disabled. Sep 13 01:01:16.398939 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 13 01:01:16.398964 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 13 01:01:16.399002 systemd[1]: Mounted media.mount - External Media Directory. Sep 13 01:01:16.399026 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 13 01:01:16.399047 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 13 01:01:16.399068 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 13 01:01:16.399089 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 13 01:01:16.399110 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 13 01:01:16.399130 kernel: ACPI: bus type drm_connector registered Sep 13 01:01:16.399150 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 13 01:01:16.399171 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 13 01:01:16.399204 systemd-journald[1177]: Journal started Sep 13 01:01:16.399239 systemd-journald[1177]: Runtime Journal (/run/log/journal/a357e97034d54349a33834319d8eb5e4) is 4.7M, max 38.0M, 33.2M free. Sep 13 01:01:16.404900 systemd[1]: Started systemd-journald.service - Journal Service. Sep 13 01:01:16.407588 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:01:16.408271 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:01:16.409668 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 01:01:16.410021 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 13 01:01:16.411331 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:01:16.411748 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:01:16.413211 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 13 01:01:16.413579 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 13 01:01:16.415086 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:01:16.415518 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:01:16.416939 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 13 01:01:16.418396 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 13 01:01:16.419883 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 13 01:01:16.435222 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 13 01:01:16.442729 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 13 01:01:16.451736 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 13 01:01:16.455236 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 01:01:16.467844 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 13 01:01:16.472443 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 13 01:01:16.474398 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 01:01:16.482646 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 13 01:01:16.488768 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 13 01:01:16.495112 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:01:16.509743 systemd-journald[1177]: Time spent on flushing to /var/log/journal/a357e97034d54349a33834319d8eb5e4 is 52.707ms for 1130 entries. Sep 13 01:01:16.509743 systemd-journald[1177]: System Journal (/var/log/journal/a357e97034d54349a33834319d8eb5e4) is 8.0M, max 584.8M, 576.8M free. Sep 13 01:01:16.615945 systemd-journald[1177]: Received client request to flush runtime journal. Sep 13 01:01:16.511583 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 13 01:01:16.516702 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 13 01:01:16.520938 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 13 01:01:16.541224 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 13 01:01:16.544318 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 13 01:01:16.563120 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:01:16.599477 systemd-tmpfiles[1226]: ACLs are not supported, ignoring. Sep 13 01:01:16.599510 systemd-tmpfiles[1226]: ACLs are not supported, ignoring. Sep 13 01:01:16.611911 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 13 01:01:16.618861 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 13 01:01:16.621435 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 13 01:01:16.665835 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 13 01:01:16.673831 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 13 01:01:16.704200 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 13 01:01:16.715925 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 13 01:01:16.730912 systemd-tmpfiles[1245]: ACLs are not supported, ignoring. Sep 13 01:01:16.731454 systemd-tmpfiles[1245]: ACLs are not supported, ignoring. Sep 13 01:01:16.741570 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 13 01:01:16.748885 udevadm[1249]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 13 01:01:17.290991 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 13 01:01:17.300851 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 13 01:01:17.337487 systemd-udevd[1254]: Using default interface naming scheme 'v255'. Sep 13 01:01:17.366398 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 13 01:01:17.377636 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 13 01:01:17.408832 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 13 01:01:17.454404 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Sep 13 01:01:17.522209 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 13 01:01:17.570637 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1257) Sep 13 01:01:17.604630 kernel: mousedev: PS/2 mouse device common for all mice Sep 13 01:01:17.637685 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 13 01:01:17.638182 systemd-networkd[1259]: lo: Link UP Sep 13 01:01:17.638195 systemd-networkd[1259]: lo: Gained carrier Sep 13 01:01:17.641133 systemd-networkd[1259]: Enumeration completed Sep 13 01:01:17.641802 systemd-networkd[1259]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 13 01:01:17.641814 systemd-networkd[1259]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 01:01:17.643199 systemd-networkd[1259]: eth0: Link UP Sep 13 01:01:17.643213 systemd-networkd[1259]: eth0: Gained carrier Sep 13 01:01:17.643231 systemd-networkd[1259]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 13 01:01:17.649194 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 13 01:01:17.664707 systemd-networkd[1259]: eth0: DHCPv4 address 10.230.35.122/30, gateway 10.230.35.121 acquired from 10.230.35.121 Sep 13 01:01:17.669470 systemd-networkd[1259]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 13 01:01:17.672772 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 13 01:01:17.676625 kernel: ACPI: button: Power Button [PWRF] Sep 13 01:01:17.718852 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 13 01:01:17.764623 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Sep 13 01:01:17.771215 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Sep 13 01:01:17.771705 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Sep 13 01:01:17.775619 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Sep 13 01:01:17.811983 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 13 01:01:17.987538 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 13 01:01:18.030271 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 13 01:01:18.040894 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 13 01:01:18.058624 lvm[1294]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 01:01:18.095855 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 13 01:01:18.097330 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 13 01:01:18.102811 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 13 01:01:18.126371 lvm[1297]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 01:01:18.161039 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 13 01:01:18.163213 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 13 01:01:18.164193 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 13 01:01:18.164390 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 13 01:01:18.165239 systemd[1]: Reached target machines.target - Containers. Sep 13 01:01:18.167983 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 13 01:01:18.175799 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 13 01:01:18.178787 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 13 01:01:18.180866 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:01:18.189803 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 13 01:01:18.194785 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 13 01:01:18.200823 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 13 01:01:18.213908 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 13 01:01:18.218904 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 13 01:01:18.237638 kernel: loop0: detected capacity change from 0 to 140768 Sep 13 01:01:18.261837 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 13 01:01:18.264889 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 13 01:01:18.283867 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 13 01:01:18.306666 kernel: loop1: detected capacity change from 0 to 8 Sep 13 01:01:18.329023 kernel: loop2: detected capacity change from 0 to 142488 Sep 13 01:01:18.368649 kernel: loop3: detected capacity change from 0 to 221472 Sep 13 01:01:18.431037 kernel: loop4: detected capacity change from 0 to 140768 Sep 13 01:01:18.457641 kernel: loop5: detected capacity change from 0 to 8 Sep 13 01:01:18.463629 kernel: loop6: detected capacity change from 0 to 142488 Sep 13 01:01:18.490638 kernel: loop7: detected capacity change from 0 to 221472 Sep 13 01:01:18.503117 (sd-merge)[1318]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-openstack'. Sep 13 01:01:18.504002 (sd-merge)[1318]: Merged extensions into '/usr'. Sep 13 01:01:18.510346 systemd[1]: Reloading requested from client PID 1305 ('systemd-sysext') (unit systemd-sysext.service)... Sep 13 01:01:18.510387 systemd[1]: Reloading... Sep 13 01:01:18.635637 zram_generator::config[1345]: No configuration found. Sep 13 01:01:18.780499 ldconfig[1302]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 13 01:01:18.851706 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:01:18.939025 systemd[1]: Reloading finished in 427 ms. Sep 13 01:01:18.961837 systemd-networkd[1259]: eth0: Gained IPv6LL Sep 13 01:01:18.965845 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 13 01:01:18.967459 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 13 01:01:18.968947 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 13 01:01:18.987829 systemd[1]: Starting ensure-sysext.service... Sep 13 01:01:18.990982 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 13 01:01:18.998473 systemd[1]: Reloading requested from client PID 1411 ('systemctl') (unit ensure-sysext.service)... Sep 13 01:01:18.998496 systemd[1]: Reloading... Sep 13 01:01:19.037943 systemd-tmpfiles[1412]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 13 01:01:19.038587 systemd-tmpfiles[1412]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 13 01:01:19.040187 systemd-tmpfiles[1412]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 13 01:01:19.040681 systemd-tmpfiles[1412]: ACLs are not supported, ignoring. Sep 13 01:01:19.040796 systemd-tmpfiles[1412]: ACLs are not supported, ignoring. Sep 13 01:01:19.047159 systemd-tmpfiles[1412]: Detected autofs mount point /boot during canonicalization of boot. Sep 13 01:01:19.047177 systemd-tmpfiles[1412]: Skipping /boot Sep 13 01:01:19.065089 systemd-tmpfiles[1412]: Detected autofs mount point /boot during canonicalization of boot. Sep 13 01:01:19.065111 systemd-tmpfiles[1412]: Skipping /boot Sep 13 01:01:19.105635 zram_generator::config[1439]: No configuration found. Sep 13 01:01:19.293091 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:01:19.386661 systemd[1]: Reloading finished in 387 ms. Sep 13 01:01:19.415384 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 13 01:01:19.421790 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 13 01:01:19.428189 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 13 01:01:19.438834 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 13 01:01:19.444075 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 13 01:01:19.455887 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 13 01:01:19.472274 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:19.473053 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:01:19.478229 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:01:19.493720 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:01:19.513347 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:01:19.516300 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:01:19.516989 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:19.522343 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 13 01:01:19.527571 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:01:19.529931 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:01:19.536316 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:01:19.536593 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:01:19.544404 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 01:01:19.556447 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 13 01:01:19.563438 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 13 01:01:19.573290 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:01:19.574675 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:01:19.591225 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:19.592144 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 13 01:01:19.595892 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 13 01:01:19.606995 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 13 01:01:19.610981 augenrules[1542]: No rules Sep 13 01:01:19.613697 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 13 01:01:19.631018 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 13 01:01:19.632952 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 13 01:01:19.633148 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 01:01:19.637690 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 13 01:01:19.643425 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 13 01:01:19.646431 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 13 01:01:19.649323 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 01:01:19.649583 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 13 01:01:19.653871 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 01:01:19.654138 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 13 01:01:19.655922 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 01:01:19.656165 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 13 01:01:19.660773 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 01:01:19.661153 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 13 01:01:19.671997 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 01:01:19.672263 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 13 01:01:19.672317 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 01:01:19.678129 systemd[1]: Finished ensure-sysext.service. Sep 13 01:01:19.691778 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 13 01:01:19.702096 systemd-resolved[1507]: Positive Trust Anchors: Sep 13 01:01:19.702141 systemd-resolved[1507]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 01:01:19.702191 systemd-resolved[1507]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 13 01:01:19.709500 systemd-resolved[1507]: Using system hostname 'srv-ngi99.gb1.brightbox.com'. Sep 13 01:01:19.712879 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 13 01:01:19.714086 systemd[1]: Reached target network.target - Network. Sep 13 01:01:19.715003 systemd[1]: Reached target network-online.target - Network is Online. Sep 13 01:01:19.715785 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 13 01:01:19.774616 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 13 01:01:19.776348 systemd[1]: Reached target sysinit.target - System Initialization. Sep 13 01:01:19.777539 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 13 01:01:19.778528 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 13 01:01:19.779350 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 13 01:01:19.780159 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 13 01:01:19.780219 systemd[1]: Reached target paths.target - Path Units. Sep 13 01:01:19.780881 systemd[1]: Reached target time-set.target - System Time Set. Sep 13 01:01:19.781853 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 13 01:01:19.782719 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 13 01:01:19.783490 systemd[1]: Reached target timers.target - Timer Units. Sep 13 01:01:19.785494 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 13 01:01:19.788817 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 13 01:01:19.792305 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 13 01:01:19.792747 systemd-networkd[1259]: eth0: Ignoring DHCPv6 address 2a02:1348:179:88de:24:19ff:fee6:237a/128 (valid for 59min 59s, preferred for 59min 59s) which conflicts with 2a02:1348:179:88de:24:19ff:fee6:237a/64 assigned by NDisc. Sep 13 01:01:19.792758 systemd-networkd[1259]: eth0: Hint: use IPv6Token= setting to change the address generated by NDisc or set UseAutonomousPrefix=no. Sep 13 01:01:19.794872 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 13 01:01:19.795644 systemd[1]: Reached target sockets.target - Socket Units. Sep 13 01:01:19.796302 systemd[1]: Reached target basic.target - Basic System. Sep 13 01:01:19.803009 systemd[1]: System is tainted: cgroupsv1 Sep 13 01:01:19.803109 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 13 01:01:19.803153 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 13 01:01:19.808800 systemd[1]: Starting containerd.service - containerd container runtime... Sep 13 01:01:19.813778 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 13 01:01:19.817861 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 13 01:01:19.822469 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 13 01:01:19.834143 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 13 01:01:19.835905 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 13 01:01:19.846629 jq[1573]: false Sep 13 01:01:19.849768 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:01:19.873150 dbus-daemon[1572]: [system] SELinux support is enabled Sep 13 01:01:19.875953 dbus-daemon[1572]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.2' (uid=244 pid=1259 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Sep 13 01:01:19.874977 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 13 01:01:19.883814 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 13 01:01:19.898754 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 13 01:01:19.907675 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 13 01:01:19.913785 extend-filesystems[1574]: Found loop4 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found loop5 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found loop6 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found loop7 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda1 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda2 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda3 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found usr Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda4 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda6 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda7 Sep 13 01:01:19.922859 extend-filesystems[1574]: Found vda9 Sep 13 01:01:19.922859 extend-filesystems[1574]: Checking size of /dev/vda9 Sep 13 01:01:19.926856 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 13 01:01:19.936928 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 13 01:01:19.939468 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 13 01:01:19.950981 systemd[1]: Starting update-engine.service - Update Engine... Sep 13 01:01:19.967726 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 13 01:01:19.972082 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 13 01:01:19.992713 extend-filesystems[1574]: Resized partition /dev/vda9 Sep 13 01:01:19.997059 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 13 01:01:19.997417 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 13 01:01:20.000962 systemd[1]: motdgen.service: Deactivated successfully. Sep 13 01:01:20.011843 extend-filesystems[1613]: resize2fs 1.47.1 (20-May-2024) Sep 13 01:01:20.038860 kernel: EXT4-fs (vda9): resizing filesystem from 1617920 to 15121403 blocks Sep 13 01:01:20.002678 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 13 01:01:20.039097 update_engine[1598]: I20250913 01:01:20.034307 1598 main.cc:92] Flatcar Update Engine starting Sep 13 01:01:20.039477 jq[1604]: true Sep 13 01:01:20.009177 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 13 01:01:20.010640 systemd-timesyncd[1565]: Contacted time server 217.114.59.66:123 (0.flatcar.pool.ntp.org). Sep 13 01:01:20.010985 systemd-timesyncd[1565]: Initial clock synchronization to Sat 2025-09-13 01:01:20.360734 UTC. Sep 13 01:01:20.022385 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 13 01:01:20.027355 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 13 01:01:20.062735 update_engine[1598]: I20250913 01:01:20.060523 1598 update_check_scheduler.cc:74] Next update check in 7m6s Sep 13 01:01:20.077655 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1255) Sep 13 01:01:20.071750 dbus-daemon[1572]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 13 01:01:20.080713 jq[1617]: true Sep 13 01:01:20.081504 (ntainerd)[1618]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 13 01:01:20.088008 systemd[1]: Started update-engine.service - Update Engine. Sep 13 01:01:20.090949 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 13 01:01:20.090996 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 13 01:01:20.104840 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Sep 13 01:01:20.107729 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 13 01:01:20.107779 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 13 01:01:20.113709 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 13 01:01:20.119809 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 13 01:01:20.142361 tar[1615]: linux-amd64/helm Sep 13 01:01:20.390946 dbus-daemon[1572]: [system] Successfully activated service 'org.freedesktop.hostname1' Sep 13 01:01:20.391413 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Sep 13 01:01:20.394767 dbus-daemon[1572]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1630 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Sep 13 01:01:20.404022 systemd-logind[1597]: Watching system buttons on /dev/input/event2 (Power Button) Sep 13 01:01:20.404804 systemd-logind[1597]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 13 01:01:20.406980 systemd-logind[1597]: New seat seat0. Sep 13 01:01:20.408791 systemd[1]: Starting polkit.service - Authorization Manager... Sep 13 01:01:20.425167 systemd[1]: Started systemd-logind.service - User Login Management. Sep 13 01:01:20.475569 bash[1650]: Updated "/home/core/.ssh/authorized_keys" Sep 13 01:01:20.477120 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 13 01:01:20.492977 systemd[1]: Starting sshkeys.service... Sep 13 01:01:20.496114 polkitd[1651]: Started polkitd version 121 Sep 13 01:01:20.531329 polkitd[1651]: Loading rules from directory /etc/polkit-1/rules.d Sep 13 01:01:20.531455 polkitd[1651]: Loading rules from directory /usr/share/polkit-1/rules.d Sep 13 01:01:20.552273 polkitd[1651]: Finished loading, compiling and executing 2 rules Sep 13 01:01:20.562087 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 13 01:01:20.594827 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Sep 13 01:01:20.564092 dbus-daemon[1572]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Sep 13 01:01:20.565406 locksmithd[1631]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 13 01:01:20.575039 polkitd[1651]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Sep 13 01:01:20.571000 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 13 01:01:20.572192 systemd[1]: Started polkit.service - Authorization Manager. Sep 13 01:01:20.612859 extend-filesystems[1613]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 13 01:01:20.612859 extend-filesystems[1613]: old_desc_blocks = 1, new_desc_blocks = 8 Sep 13 01:01:20.612859 extend-filesystems[1613]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Sep 13 01:01:20.605796 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 13 01:01:20.627353 extend-filesystems[1574]: Resized filesystem in /dev/vda9 Sep 13 01:01:20.606217 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 13 01:01:20.658490 systemd-hostnamed[1630]: Hostname set to (static) Sep 13 01:01:20.784620 containerd[1618]: time="2025-09-13T01:01:20.782960265Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Sep 13 01:01:20.875355 containerd[1618]: time="2025-09-13T01:01:20.873456156Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.879998 containerd[1618]: time="2025-09-13T01:01:20.879932611Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.106-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:01:20.879998 containerd[1618]: time="2025-09-13T01:01:20.879995152Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 13 01:01:20.880124 containerd[1618]: time="2025-09-13T01:01:20.880024919Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.883879556Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.883927903Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.884049503Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.884073657Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.884359440Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884415 containerd[1618]: time="2025-09-13T01:01:20.884400553Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884691 containerd[1618]: time="2025-09-13T01:01:20.884422780Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884691 containerd[1618]: time="2025-09-13T01:01:20.884440793Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.884691 containerd[1618]: time="2025-09-13T01:01:20.884562568Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.885363 containerd[1618]: time="2025-09-13T01:01:20.884978057Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 13 01:01:20.885363 containerd[1618]: time="2025-09-13T01:01:20.885158445Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 01:01:20.885363 containerd[1618]: time="2025-09-13T01:01:20.885182810Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 13 01:01:20.885363 containerd[1618]: time="2025-09-13T01:01:20.885316260Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 13 01:01:20.885537 containerd[1618]: time="2025-09-13T01:01:20.885409971Z" level=info msg="metadata content store policy set" policy=shared Sep 13 01:01:20.903344 containerd[1618]: time="2025-09-13T01:01:20.903285905Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 13 01:01:20.903609 containerd[1618]: time="2025-09-13T01:01:20.903399073Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 13 01:01:20.903609 containerd[1618]: time="2025-09-13T01:01:20.903428639Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 13 01:01:20.903609 containerd[1618]: time="2025-09-13T01:01:20.903452326Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 13 01:01:20.903609 containerd[1618]: time="2025-09-13T01:01:20.903480130Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 13 01:01:20.904216 containerd[1618]: time="2025-09-13T01:01:20.904187111Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 13 01:01:20.907551 containerd[1618]: time="2025-09-13T01:01:20.906694767Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 13 01:01:20.907833 containerd[1618]: time="2025-09-13T01:01:20.907803347Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 13 01:01:20.907884 containerd[1618]: time="2025-09-13T01:01:20.907839050Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 13 01:01:20.907942 containerd[1618]: time="2025-09-13T01:01:20.907881217Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 13 01:01:20.907942 containerd[1618]: time="2025-09-13T01:01:20.907918086Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910516 containerd[1618]: time="2025-09-13T01:01:20.910435444Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910516 containerd[1618]: time="2025-09-13T01:01:20.910501209Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910661 containerd[1618]: time="2025-09-13T01:01:20.910528492Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910661 containerd[1618]: time="2025-09-13T01:01:20.910550962Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910661 containerd[1618]: time="2025-09-13T01:01:20.910590651Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910661 containerd[1618]: time="2025-09-13T01:01:20.910639252Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910802 containerd[1618]: time="2025-09-13T01:01:20.910679526Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 13 01:01:20.910802 containerd[1618]: time="2025-09-13T01:01:20.910754461Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.910879 containerd[1618]: time="2025-09-13T01:01:20.910801217Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.910879 containerd[1618]: time="2025-09-13T01:01:20.910823009Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.910879 containerd[1618]: time="2025-09-13T01:01:20.910844091Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911013 containerd[1618]: time="2025-09-13T01:01:20.910886908Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911013 containerd[1618]: time="2025-09-13T01:01:20.910908204Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911013 containerd[1618]: time="2025-09-13T01:01:20.910926177Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911013 containerd[1618]: time="2025-09-13T01:01:20.910965946Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911013 containerd[1618]: time="2025-09-13T01:01:20.910986125Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911194 containerd[1618]: time="2025-09-13T01:01:20.911033463Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911194 containerd[1618]: time="2025-09-13T01:01:20.911060733Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.911194 containerd[1618]: time="2025-09-13T01:01:20.911080253Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.912137 containerd[1618]: time="2025-09-13T01:01:20.911801884Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.912137 containerd[1618]: time="2025-09-13T01:01:20.911838970Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 13 01:01:20.912137 containerd[1618]: time="2025-09-13T01:01:20.911901485Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.912137 containerd[1618]: time="2025-09-13T01:01:20.911926767Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.912137 containerd[1618]: time="2025-09-13T01:01:20.912103906Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913676586Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913743434Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913784806Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913816542Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913834697Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.913890 containerd[1618]: time="2025-09-13T01:01:20.913882317Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 13 01:01:20.914165 containerd[1618]: time="2025-09-13T01:01:20.913909222Z" level=info msg="NRI interface is disabled by configuration." Sep 13 01:01:20.914165 containerd[1618]: time="2025-09-13T01:01:20.913926667Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 13 01:01:20.917007 containerd[1618]: time="2025-09-13T01:01:20.916270917Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 13 01:01:20.917007 containerd[1618]: time="2025-09-13T01:01:20.916391927Z" level=info msg="Connect containerd service" Sep 13 01:01:20.917007 containerd[1618]: time="2025-09-13T01:01:20.916468905Z" level=info msg="using legacy CRI server" Sep 13 01:01:20.917007 containerd[1618]: time="2025-09-13T01:01:20.916505436Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 13 01:01:20.917007 containerd[1618]: time="2025-09-13T01:01:20.916754249Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927152307Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927392050Z" level=info msg="Start subscribing containerd event" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927497795Z" level=info msg="Start recovering state" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927691130Z" level=info msg="Start event monitor" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927722471Z" level=info msg="Start snapshots syncer" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927741780Z" level=info msg="Start cni network conf syncer for default" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.927754489Z" level=info msg="Start streaming server" Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.928627920Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.928720641Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 13 01:01:20.930643 containerd[1618]: time="2025-09-13T01:01:20.928820718Z" level=info msg="containerd successfully booted in 0.147021s" Sep 13 01:01:20.929421 systemd[1]: Started containerd.service - containerd container runtime. Sep 13 01:01:21.263073 sshd_keygen[1610]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 13 01:01:21.316770 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 13 01:01:21.330312 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 13 01:01:21.354512 systemd[1]: issuegen.service: Deactivated successfully. Sep 13 01:01:21.357470 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 13 01:01:21.368914 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 13 01:01:21.380700 tar[1615]: linux-amd64/LICENSE Sep 13 01:01:21.380700 tar[1615]: linux-amd64/README.md Sep 13 01:01:21.405287 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 13 01:01:21.415415 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 13 01:01:21.420878 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 13 01:01:21.423140 systemd[1]: Reached target getty.target - Login Prompts. Sep 13 01:01:21.427306 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 13 01:01:21.915918 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:01:21.931502 (kubelet)[1728]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:01:22.607286 kubelet[1728]: E0913 01:01:22.607177 1728 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:01:22.610977 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:01:22.611336 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:01:23.952133 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 13 01:01:23.960167 systemd[1]: Started sshd@0-10.230.35.122:22-139.178.68.195:45876.service - OpenSSH per-connection server daemon (139.178.68.195:45876). Sep 13 01:01:24.912724 sshd[1738]: Accepted publickey for core from 139.178.68.195 port 45876 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:24.915568 sshd[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:24.933836 systemd-logind[1597]: New session 1 of user core. Sep 13 01:01:24.936583 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 13 01:01:24.945292 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 13 01:01:24.988231 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 13 01:01:25.000336 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 13 01:01:25.009713 (systemd)[1744]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 13 01:01:25.166823 systemd[1744]: Queued start job for default target default.target. Sep 13 01:01:25.168173 systemd[1744]: Created slice app.slice - User Application Slice. Sep 13 01:01:25.168213 systemd[1744]: Reached target paths.target - Paths. Sep 13 01:01:25.168238 systemd[1744]: Reached target timers.target - Timers. Sep 13 01:01:25.179883 systemd[1744]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 13 01:01:25.191258 systemd[1744]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 13 01:01:25.191605 systemd[1744]: Reached target sockets.target - Sockets. Sep 13 01:01:25.192016 systemd[1744]: Reached target basic.target - Basic System. Sep 13 01:01:25.192245 systemd[1744]: Reached target default.target - Main User Target. Sep 13 01:01:25.192324 systemd[1744]: Startup finished in 173ms. Sep 13 01:01:25.192753 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 13 01:01:25.202295 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 13 01:01:25.866222 systemd[1]: Started sshd@1-10.230.35.122:22-139.178.68.195:45884.service - OpenSSH per-connection server daemon (139.178.68.195:45884). Sep 13 01:01:26.502638 login[1717]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 13 01:01:26.505369 login[1715]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 13 01:01:26.512964 systemd-logind[1597]: New session 2 of user core. Sep 13 01:01:26.521601 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 13 01:01:26.528886 systemd-logind[1597]: New session 3 of user core. Sep 13 01:01:26.537320 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 13 01:01:26.851774 sshd[1756]: Accepted publickey for core from 139.178.68.195 port 45884 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:26.854007 sshd[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:26.860911 systemd-logind[1597]: New session 4 of user core. Sep 13 01:01:26.872360 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 13 01:01:26.936160 coreos-metadata[1570]: Sep 13 01:01:26.936 WARN failed to locate config-drive, using the metadata service API instead Sep 13 01:01:26.965814 coreos-metadata[1570]: Sep 13 01:01:26.965 INFO Fetching http://169.254.169.254/openstack/2012-08-10/meta_data.json: Attempt #1 Sep 13 01:01:26.977046 coreos-metadata[1570]: Sep 13 01:01:26.976 INFO Fetch failed with 404: resource not found Sep 13 01:01:26.977046 coreos-metadata[1570]: Sep 13 01:01:26.977 INFO Fetching http://169.254.169.254/latest/meta-data/hostname: Attempt #1 Sep 13 01:01:26.977667 coreos-metadata[1570]: Sep 13 01:01:26.977 INFO Fetch successful Sep 13 01:01:26.977916 coreos-metadata[1570]: Sep 13 01:01:26.977 INFO Fetching http://169.254.169.254/latest/meta-data/instance-id: Attempt #1 Sep 13 01:01:26.992032 coreos-metadata[1570]: Sep 13 01:01:26.991 INFO Fetch successful Sep 13 01:01:26.992252 coreos-metadata[1570]: Sep 13 01:01:26.992 INFO Fetching http://169.254.169.254/latest/meta-data/instance-type: Attempt #1 Sep 13 01:01:27.007409 coreos-metadata[1570]: Sep 13 01:01:27.007 INFO Fetch successful Sep 13 01:01:27.007409 coreos-metadata[1570]: Sep 13 01:01:27.007 INFO Fetching http://169.254.169.254/latest/meta-data/local-ipv4: Attempt #1 Sep 13 01:01:27.029485 coreos-metadata[1570]: Sep 13 01:01:27.029 INFO Fetch successful Sep 13 01:01:27.029810 coreos-metadata[1570]: Sep 13 01:01:27.029 INFO Fetching http://169.254.169.254/latest/meta-data/public-ipv4: Attempt #1 Sep 13 01:01:27.054036 coreos-metadata[1570]: Sep 13 01:01:27.053 INFO Fetch successful Sep 13 01:01:27.084293 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 13 01:01:27.085461 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 13 01:01:27.528099 sshd[1756]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:27.532384 systemd[1]: sshd@1-10.230.35.122:22-139.178.68.195:45884.service: Deactivated successfully. Sep 13 01:01:27.537571 systemd-logind[1597]: Session 4 logged out. Waiting for processes to exit. Sep 13 01:01:27.539231 systemd[1]: session-4.scope: Deactivated successfully. Sep 13 01:01:27.541949 systemd-logind[1597]: Removed session 4. Sep 13 01:01:27.686327 systemd[1]: Started sshd@2-10.230.35.122:22-139.178.68.195:45892.service - OpenSSH per-connection server daemon (139.178.68.195:45892). Sep 13 01:01:27.816809 coreos-metadata[1669]: Sep 13 01:01:27.816 WARN failed to locate config-drive, using the metadata service API instead Sep 13 01:01:27.840069 coreos-metadata[1669]: Sep 13 01:01:27.839 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys: Attempt #1 Sep 13 01:01:27.865877 coreos-metadata[1669]: Sep 13 01:01:27.865 INFO Fetch successful Sep 13 01:01:27.865877 coreos-metadata[1669]: Sep 13 01:01:27.865 INFO Fetching http://169.254.169.254/latest/meta-data/public-keys/0/openssh-key: Attempt #1 Sep 13 01:01:27.896238 coreos-metadata[1669]: Sep 13 01:01:27.896 INFO Fetch successful Sep 13 01:01:27.899721 unknown[1669]: wrote ssh authorized keys file for user: core Sep 13 01:01:27.923822 update-ssh-keys[1806]: Updated "/home/core/.ssh/authorized_keys" Sep 13 01:01:27.924857 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 13 01:01:27.933476 systemd[1]: Finished sshkeys.service. Sep 13 01:01:27.937874 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 13 01:01:27.938590 systemd[1]: Startup finished in 15.844s (kernel) + 12.811s (userspace) = 28.655s. Sep 13 01:01:28.583483 sshd[1802]: Accepted publickey for core from 139.178.68.195 port 45892 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:28.585819 sshd[1802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:28.593775 systemd-logind[1597]: New session 5 of user core. Sep 13 01:01:28.600108 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 13 01:01:29.216002 sshd[1802]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:29.222289 systemd[1]: sshd@2-10.230.35.122:22-139.178.68.195:45892.service: Deactivated successfully. Sep 13 01:01:29.225519 systemd-logind[1597]: Session 5 logged out. Waiting for processes to exit. Sep 13 01:01:29.226521 systemd[1]: session-5.scope: Deactivated successfully. Sep 13 01:01:29.228566 systemd-logind[1597]: Removed session 5. Sep 13 01:01:32.630077 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 13 01:01:32.638909 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:01:32.889993 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:01:32.902289 (kubelet)[1832]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:01:32.972538 kubelet[1832]: E0913 01:01:32.972423 1832 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:01:32.977163 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:01:32.977717 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:01:39.460975 systemd[1]: Started sshd@3-10.230.35.122:22-139.178.68.195:57212.service - OpenSSH per-connection server daemon (139.178.68.195:57212). Sep 13 01:01:40.359026 sshd[1840]: Accepted publickey for core from 139.178.68.195 port 57212 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:40.361640 sshd[1840]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:40.369517 systemd-logind[1597]: New session 6 of user core. Sep 13 01:01:40.377068 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 13 01:01:40.990904 sshd[1840]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:40.997905 systemd[1]: sshd@3-10.230.35.122:22-139.178.68.195:57212.service: Deactivated successfully. Sep 13 01:01:41.002753 systemd-logind[1597]: Session 6 logged out. Waiting for processes to exit. Sep 13 01:01:41.004714 systemd[1]: session-6.scope: Deactivated successfully. Sep 13 01:01:41.006597 systemd-logind[1597]: Removed session 6. Sep 13 01:01:41.138950 systemd[1]: Started sshd@4-10.230.35.122:22-139.178.68.195:50600.service - OpenSSH per-connection server daemon (139.178.68.195:50600). Sep 13 01:01:42.031950 sshd[1848]: Accepted publickey for core from 139.178.68.195 port 50600 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:42.034101 sshd[1848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:42.041067 systemd-logind[1597]: New session 7 of user core. Sep 13 01:01:42.050142 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 13 01:01:42.644047 sshd[1848]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:42.649076 systemd[1]: sshd@4-10.230.35.122:22-139.178.68.195:50600.service: Deactivated successfully. Sep 13 01:01:42.652883 systemd-logind[1597]: Session 7 logged out. Waiting for processes to exit. Sep 13 01:01:42.653758 systemd[1]: session-7.scope: Deactivated successfully. Sep 13 01:01:42.655132 systemd-logind[1597]: Removed session 7. Sep 13 01:01:42.801159 systemd[1]: Started sshd@5-10.230.35.122:22-139.178.68.195:50616.service - OpenSSH per-connection server daemon (139.178.68.195:50616). Sep 13 01:01:43.129063 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 13 01:01:43.136878 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:01:43.314822 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:01:43.327180 (kubelet)[1869]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:01:43.436350 kubelet[1869]: E0913 01:01:43.435872 1869 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:01:43.439088 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:01:43.439482 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:01:43.686829 sshd[1856]: Accepted publickey for core from 139.178.68.195 port 50616 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:43.688699 sshd[1856]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:43.695122 systemd-logind[1597]: New session 8 of user core. Sep 13 01:01:43.704029 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 13 01:01:44.312132 sshd[1856]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:44.318042 systemd[1]: sshd@5-10.230.35.122:22-139.178.68.195:50616.service: Deactivated successfully. Sep 13 01:01:44.321916 systemd[1]: session-8.scope: Deactivated successfully. Sep 13 01:01:44.322074 systemd-logind[1597]: Session 8 logged out. Waiting for processes to exit. Sep 13 01:01:44.324405 systemd-logind[1597]: Removed session 8. Sep 13 01:01:44.468042 systemd[1]: Started sshd@6-10.230.35.122:22-139.178.68.195:50628.service - OpenSSH per-connection server daemon (139.178.68.195:50628). Sep 13 01:01:45.364485 sshd[1885]: Accepted publickey for core from 139.178.68.195 port 50628 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:45.366477 sshd[1885]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:45.372851 systemd-logind[1597]: New session 9 of user core. Sep 13 01:01:45.385161 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 13 01:01:45.860689 sudo[1889]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 13 01:01:45.861185 sudo[1889]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:01:45.876508 sudo[1889]: pam_unix(sudo:session): session closed for user root Sep 13 01:01:46.023038 sshd[1885]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:46.028022 systemd[1]: sshd@6-10.230.35.122:22-139.178.68.195:50628.service: Deactivated successfully. Sep 13 01:01:46.032226 systemd-logind[1597]: Session 9 logged out. Waiting for processes to exit. Sep 13 01:01:46.033182 systemd[1]: session-9.scope: Deactivated successfully. Sep 13 01:01:46.034709 systemd-logind[1597]: Removed session 9. Sep 13 01:01:46.179092 systemd[1]: Started sshd@7-10.230.35.122:22-139.178.68.195:50644.service - OpenSSH per-connection server daemon (139.178.68.195:50644). Sep 13 01:01:47.067558 sshd[1894]: Accepted publickey for core from 139.178.68.195 port 50644 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:47.069586 sshd[1894]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:47.077178 systemd-logind[1597]: New session 10 of user core. Sep 13 01:01:47.085754 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 13 01:01:47.547762 sudo[1899]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 13 01:01:47.548250 sudo[1899]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:01:47.553743 sudo[1899]: pam_unix(sudo:session): session closed for user root Sep 13 01:01:47.561818 sudo[1898]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 13 01:01:47.562291 sudo[1898]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:01:47.587035 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 13 01:01:47.589350 auditctl[1902]: No rules Sep 13 01:01:47.589949 systemd[1]: audit-rules.service: Deactivated successfully. Sep 13 01:01:47.590345 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 13 01:01:47.606299 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 13 01:01:47.644120 augenrules[1921]: No rules Sep 13 01:01:47.645804 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 13 01:01:47.647463 sudo[1898]: pam_unix(sudo:session): session closed for user root Sep 13 01:01:47.794982 sshd[1894]: pam_unix(sshd:session): session closed for user core Sep 13 01:01:47.799758 systemd[1]: sshd@7-10.230.35.122:22-139.178.68.195:50644.service: Deactivated successfully. Sep 13 01:01:47.803220 systemd-logind[1597]: Session 10 logged out. Waiting for processes to exit. Sep 13 01:01:47.803943 systemd[1]: session-10.scope: Deactivated successfully. Sep 13 01:01:47.807319 systemd-logind[1597]: Removed session 10. Sep 13 01:01:47.962900 systemd[1]: Started sshd@8-10.230.35.122:22-139.178.68.195:50646.service - OpenSSH per-connection server daemon (139.178.68.195:50646). Sep 13 01:01:48.856336 sshd[1930]: Accepted publickey for core from 139.178.68.195 port 50646 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:01:48.858337 sshd[1930]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:01:48.864472 systemd-logind[1597]: New session 11 of user core. Sep 13 01:01:48.876055 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 13 01:01:49.337771 sudo[1934]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 13 01:01:49.338246 sudo[1934]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 13 01:01:49.796981 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 13 01:01:49.797338 (dockerd)[1950]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 13 01:01:50.223037 dockerd[1950]: time="2025-09-13T01:01:50.222892600Z" level=info msg="Starting up" Sep 13 01:01:50.453062 systemd[1]: Started sshd@9-10.230.35.122:22-101.126.71.100:52868.service - OpenSSH per-connection server daemon (101.126.71.100:52868). Sep 13 01:01:50.478783 dockerd[1950]: time="2025-09-13T01:01:50.478072293Z" level=info msg="Loading containers: start." Sep 13 01:01:50.613642 kernel: Initializing XFRM netlink socket Sep 13 01:01:50.695055 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Sep 13 01:01:50.730368 systemd-networkd[1259]: docker0: Link UP Sep 13 01:01:50.749470 dockerd[1950]: time="2025-09-13T01:01:50.749423285Z" level=info msg="Loading containers: done." Sep 13 01:01:50.783968 dockerd[1950]: time="2025-09-13T01:01:50.783761680Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 13 01:01:50.784844 dockerd[1950]: time="2025-09-13T01:01:50.784054646Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 13 01:01:50.784844 dockerd[1950]: time="2025-09-13T01:01:50.784245238Z" level=info msg="Daemon has completed initialization" Sep 13 01:01:50.828473 dockerd[1950]: time="2025-09-13T01:01:50.828369228Z" level=info msg="API listen on /run/docker.sock" Sep 13 01:01:50.828885 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 13 01:01:52.129835 containerd[1618]: time="2025-09-13T01:01:52.129705124Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\"" Sep 13 01:01:53.078275 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1963170821.mount: Deactivated successfully. Sep 13 01:01:53.629715 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Sep 13 01:01:53.642100 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:01:53.859330 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:01:53.868686 (kubelet)[2163]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:01:53.937533 kubelet[2163]: E0913 01:01:53.937300 2163 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:01:53.940159 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:01:53.940499 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:01:56.021654 containerd[1618]: time="2025-09-13T01:01:56.021533957Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:56.023391 containerd[1618]: time="2025-09-13T01:01:56.023332189Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.13: active requests=0, bytes read=28117132" Sep 13 01:01:56.025625 containerd[1618]: time="2025-09-13T01:01:56.023804132Z" level=info msg="ImageCreate event name:\"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:56.028118 containerd[1618]: time="2025-09-13T01:01:56.028077751Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:56.029871 containerd[1618]: time="2025-09-13T01:01:56.029830890Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.13\" with image id \"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.13\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e\", size \"28113723\" in 3.899998134s" Sep 13 01:01:56.030123 containerd[1618]: time="2025-09-13T01:01:56.030088954Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\" returns image reference \"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\"" Sep 13 01:01:56.032279 containerd[1618]: time="2025-09-13T01:01:56.032226626Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\"" Sep 13 01:01:58.248681 containerd[1618]: time="2025-09-13T01:01:58.248026129Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:58.250523 containerd[1618]: time="2025-09-13T01:01:58.250185735Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.13: active requests=0, bytes read=24716640" Sep 13 01:01:58.251625 containerd[1618]: time="2025-09-13T01:01:58.251292147Z" level=info msg="ImageCreate event name:\"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:58.255404 containerd[1618]: time="2025-09-13T01:01:58.255367993Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:01:58.257047 containerd[1618]: time="2025-09-13T01:01:58.257002868Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.13\" with image id \"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.13\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09\", size \"26351311\" in 2.22472s" Sep 13 01:01:58.257138 containerd[1618]: time="2025-09-13T01:01:58.257054472Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\" returns image reference \"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\"" Sep 13 01:01:58.258365 containerd[1618]: time="2025-09-13T01:01:58.258321759Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\"" Sep 13 01:02:00.030111 containerd[1618]: time="2025-09-13T01:02:00.029979738Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:00.032411 containerd[1618]: time="2025-09-13T01:02:00.032341392Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.13: active requests=0, bytes read=18787706" Sep 13 01:02:00.033550 containerd[1618]: time="2025-09-13T01:02:00.033488483Z" level=info msg="ImageCreate event name:\"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:00.038109 containerd[1618]: time="2025-09-13T01:02:00.038074633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:00.040633 containerd[1618]: time="2025-09-13T01:02:00.039692955Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.13\" with image id \"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.13\", repo digest \"registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241\", size \"20422395\" in 1.781318578s" Sep 13 01:02:00.040633 containerd[1618]: time="2025-09-13T01:02:00.039763228Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\" returns image reference \"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\"" Sep 13 01:02:00.041740 containerd[1618]: time="2025-09-13T01:02:00.041710233Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\"" Sep 13 01:02:02.638737 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2343135071.mount: Deactivated successfully. Sep 13 01:02:03.382488 containerd[1618]: time="2025-09-13T01:02:03.381425444Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:03.384423 containerd[1618]: time="2025-09-13T01:02:03.384093541Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.13: active requests=0, bytes read=30410260" Sep 13 01:02:03.385873 containerd[1618]: time="2025-09-13T01:02:03.385788009Z" level=info msg="ImageCreate event name:\"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:03.389293 containerd[1618]: time="2025-09-13T01:02:03.389243450Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:03.390882 containerd[1618]: time="2025-09-13T01:02:03.390739266Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.13\" with image id \"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\", repo tag \"registry.k8s.io/kube-proxy:v1.31.13\", repo digest \"registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae\", size \"30409271\" in 3.348980396s" Sep 13 01:02:03.390882 containerd[1618]: time="2025-09-13T01:02:03.390814018Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\" returns image reference \"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\"" Sep 13 01:02:03.392575 containerd[1618]: time="2025-09-13T01:02:03.392442560Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 13 01:02:04.065458 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Sep 13 01:02:04.076580 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:04.102822 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2600528739.mount: Deactivated successfully. Sep 13 01:02:04.390428 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:04.407342 (kubelet)[2211]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 13 01:02:04.485278 kubelet[2211]: E0913 01:02:04.484827 2211 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 01:02:04.488883 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 01:02:04.489176 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 01:02:05.298137 update_engine[1598]: I20250913 01:02:05.296870 1598 update_attempter.cc:509] Updating boot flags... Sep 13 01:02:05.378093 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2270) Sep 13 01:02:05.484655 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2274) Sep 13 01:02:05.577673 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2274) Sep 13 01:02:05.948453 containerd[1618]: time="2025-09-13T01:02:05.948350123Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:05.951349 containerd[1618]: time="2025-09-13T01:02:05.950938591Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565249" Sep 13 01:02:05.953927 containerd[1618]: time="2025-09-13T01:02:05.952063605Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:05.961124 containerd[1618]: time="2025-09-13T01:02:05.960691413Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:05.963738 containerd[1618]: time="2025-09-13T01:02:05.963697090Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.571019418s" Sep 13 01:02:05.963890 containerd[1618]: time="2025-09-13T01:02:05.963860987Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 13 01:02:05.966078 containerd[1618]: time="2025-09-13T01:02:05.966029868Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 13 01:02:06.617515 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3265904217.mount: Deactivated successfully. Sep 13 01:02:06.624646 containerd[1618]: time="2025-09-13T01:02:06.623665205Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:06.627255 containerd[1618]: time="2025-09-13T01:02:06.627209578Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321146" Sep 13 01:02:06.629816 containerd[1618]: time="2025-09-13T01:02:06.629781422Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:06.634921 containerd[1618]: time="2025-09-13T01:02:06.634869187Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:06.636070 containerd[1618]: time="2025-09-13T01:02:06.636033893Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 669.952805ms" Sep 13 01:02:06.636269 containerd[1618]: time="2025-09-13T01:02:06.636238087Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 13 01:02:06.637650 containerd[1618]: time="2025-09-13T01:02:06.637588929Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 13 01:02:07.366394 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3422236171.mount: Deactivated successfully. Sep 13 01:02:10.521186 containerd[1618]: time="2025-09-13T01:02:10.520974774Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:10.523460 containerd[1618]: time="2025-09-13T01:02:10.523112746Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56910717" Sep 13 01:02:10.524643 containerd[1618]: time="2025-09-13T01:02:10.524582471Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:10.531881 containerd[1618]: time="2025-09-13T01:02:10.531040160Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:10.532803 containerd[1618]: time="2025-09-13T01:02:10.532749078Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 3.895093388s" Sep 13 01:02:10.532896 containerd[1618]: time="2025-09-13T01:02:10.532814010Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Sep 13 01:02:14.587946 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Sep 13 01:02:14.600948 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:14.614976 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 13 01:02:14.615149 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 13 01:02:14.615842 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:14.629903 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:14.664468 systemd[1]: Reloading requested from client PID 2368 ('systemctl') (unit session-11.scope)... Sep 13 01:02:14.664755 systemd[1]: Reloading... Sep 13 01:02:14.883308 zram_generator::config[2406]: No configuration found. Sep 13 01:02:15.072901 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:02:15.182627 systemd[1]: Reloading finished in 517 ms. Sep 13 01:02:15.243357 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 13 01:02:15.243500 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 13 01:02:15.244092 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:15.252973 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:15.440004 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:15.440417 (kubelet)[2486]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 13 01:02:15.525494 kubelet[2486]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:02:15.527723 kubelet[2486]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 01:02:15.527723 kubelet[2486]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:02:15.527723 kubelet[2486]: I0913 01:02:15.527418 2486 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 01:02:16.212630 kubelet[2486]: I0913 01:02:16.210876 2486 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 01:02:16.212630 kubelet[2486]: I0913 01:02:16.210934 2486 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 01:02:16.212630 kubelet[2486]: I0913 01:02:16.211265 2486 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 01:02:16.246886 kubelet[2486]: I0913 01:02:16.246815 2486 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 01:02:16.247819 kubelet[2486]: E0913 01:02:16.247777 2486 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.230.35.122:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:16.257586 kubelet[2486]: E0913 01:02:16.257522 2486 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 01:02:16.257803 kubelet[2486]: I0913 01:02:16.257782 2486 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 01:02:16.266079 kubelet[2486]: I0913 01:02:16.266050 2486 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 01:02:16.282561 kubelet[2486]: I0913 01:02:16.282513 2486 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 01:02:16.283188 kubelet[2486]: I0913 01:02:16.283122 2486 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 01:02:16.283649 kubelet[2486]: I0913 01:02:16.283285 2486 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-ngi99.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Sep 13 01:02:16.284034 kubelet[2486]: I0913 01:02:16.284007 2486 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 01:02:16.284156 kubelet[2486]: I0913 01:02:16.284137 2486 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 01:02:16.284452 kubelet[2486]: I0913 01:02:16.284427 2486 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:02:16.289538 kubelet[2486]: I0913 01:02:16.289513 2486 kubelet.go:408] "Attempting to sync node with API server" Sep 13 01:02:16.289694 kubelet[2486]: I0913 01:02:16.289673 2486 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 01:02:16.289898 kubelet[2486]: I0913 01:02:16.289876 2486 kubelet.go:314] "Adding apiserver pod source" Sep 13 01:02:16.290052 kubelet[2486]: I0913 01:02:16.290032 2486 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 01:02:16.296908 kubelet[2486]: W0913 01:02:16.296770 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.230.35.122:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-ngi99.gb1.brightbox.com&limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:16.297023 kubelet[2486]: E0913 01:02:16.296931 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.230.35.122:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-ngi99.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:16.298641 kubelet[2486]: W0913 01:02:16.298271 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.230.35.122:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:16.298641 kubelet[2486]: E0913 01:02:16.298332 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.230.35.122:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:16.299218 kubelet[2486]: I0913 01:02:16.299104 2486 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 13 01:02:16.303368 kubelet[2486]: I0913 01:02:16.303338 2486 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 01:02:16.305564 kubelet[2486]: W0913 01:02:16.304127 2486 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 13 01:02:16.306276 kubelet[2486]: I0913 01:02:16.306253 2486 server.go:1274] "Started kubelet" Sep 13 01:02:16.308579 kubelet[2486]: I0913 01:02:16.308515 2486 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 01:02:16.310069 kubelet[2486]: I0913 01:02:16.310016 2486 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 01:02:16.310805 kubelet[2486]: I0913 01:02:16.310781 2486 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 01:02:16.311027 kubelet[2486]: I0913 01:02:16.310049 2486 server.go:449] "Adding debug handlers to kubelet server" Sep 13 01:02:16.313869 kubelet[2486]: I0913 01:02:16.313828 2486 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 01:02:16.317461 kubelet[2486]: E0913 01:02:16.315037 2486 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.230.35.122:6443/api/v1/namespaces/default/events\": dial tcp 10.230.35.122:6443: connect: connection refused" event="&Event{ObjectMeta:{srv-ngi99.gb1.brightbox.com.1864b1e16b3c7d55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:srv-ngi99.gb1.brightbox.com,UID:srv-ngi99.gb1.brightbox.com,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:srv-ngi99.gb1.brightbox.com,},FirstTimestamp:2025-09-13 01:02:16.306220373 +0000 UTC m=+0.855848108,LastTimestamp:2025-09-13 01:02:16.306220373 +0000 UTC m=+0.855848108,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:srv-ngi99.gb1.brightbox.com,}" Sep 13 01:02:16.321017 kubelet[2486]: I0913 01:02:16.320989 2486 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 01:02:16.324913 kubelet[2486]: I0913 01:02:16.324881 2486 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 01:02:16.325407 kubelet[2486]: E0913 01:02:16.325354 2486 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"srv-ngi99.gb1.brightbox.com\" not found" Sep 13 01:02:16.328577 kubelet[2486]: E0913 01:02:16.328517 2486 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.35.122:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-ngi99.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.35.122:6443: connect: connection refused" interval="200ms" Sep 13 01:02:16.329163 kubelet[2486]: I0913 01:02:16.329136 2486 factory.go:221] Registration of the systemd container factory successfully Sep 13 01:02:16.329399 kubelet[2486]: I0913 01:02:16.329372 2486 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 01:02:16.329981 kubelet[2486]: I0913 01:02:16.329955 2486 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 01:02:16.330114 kubelet[2486]: I0913 01:02:16.330087 2486 reconciler.go:26] "Reconciler: start to sync state" Sep 13 01:02:16.332652 kubelet[2486]: I0913 01:02:16.332629 2486 factory.go:221] Registration of the containerd container factory successfully Sep 13 01:02:16.343651 kubelet[2486]: I0913 01:02:16.342772 2486 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 01:02:16.344327 kubelet[2486]: I0913 01:02:16.344298 2486 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 01:02:16.344411 kubelet[2486]: I0913 01:02:16.344351 2486 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 01:02:16.344411 kubelet[2486]: I0913 01:02:16.344389 2486 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 01:02:16.344495 kubelet[2486]: E0913 01:02:16.344469 2486 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 01:02:16.357468 kubelet[2486]: W0913 01:02:16.357405 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.230.35.122:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:16.357658 kubelet[2486]: E0913 01:02:16.357480 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.230.35.122:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:16.363183 kubelet[2486]: W0913 01:02:16.362755 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.230.35.122:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:16.363183 kubelet[2486]: E0913 01:02:16.362846 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.230.35.122:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:16.425617 kubelet[2486]: E0913 01:02:16.425536 2486 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"srv-ngi99.gb1.brightbox.com\" not found" Sep 13 01:02:16.430881 kubelet[2486]: I0913 01:02:16.430814 2486 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 01:02:16.430881 kubelet[2486]: I0913 01:02:16.430881 2486 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 01:02:16.431106 kubelet[2486]: I0913 01:02:16.430939 2486 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:02:16.445125 kubelet[2486]: E0913 01:02:16.445045 2486 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 01:02:16.469146 kubelet[2486]: I0913 01:02:16.468986 2486 policy_none.go:49] "None policy: Start" Sep 13 01:02:16.472213 kubelet[2486]: I0913 01:02:16.472126 2486 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 01:02:16.472312 kubelet[2486]: I0913 01:02:16.472253 2486 state_mem.go:35] "Initializing new in-memory state store" Sep 13 01:02:16.480468 kubelet[2486]: I0913 01:02:16.480415 2486 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 01:02:16.480816 kubelet[2486]: I0913 01:02:16.480784 2486 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 01:02:16.480914 kubelet[2486]: I0913 01:02:16.480822 2486 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 01:02:16.483090 kubelet[2486]: I0913 01:02:16.483055 2486 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 01:02:16.489238 kubelet[2486]: E0913 01:02:16.489203 2486 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"srv-ngi99.gb1.brightbox.com\" not found" Sep 13 01:02:16.529694 kubelet[2486]: E0913 01:02:16.529621 2486 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.35.122:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-ngi99.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.35.122:6443: connect: connection refused" interval="400ms" Sep 13 01:02:16.585213 kubelet[2486]: I0913 01:02:16.584649 2486 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.585213 kubelet[2486]: E0913 01:02:16.585131 2486 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.230.35.122:6443/api/v1/nodes\": dial tcp 10.230.35.122:6443: connect: connection refused" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.732844 kubelet[2486]: I0913 01:02:16.732642 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-flexvolume-dir\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.732844 kubelet[2486]: I0913 01:02:16.732715 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-k8s-certs\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.732844 kubelet[2486]: I0913 01:02:16.732753 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-kubeconfig\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.732844 kubelet[2486]: I0913 01:02:16.732786 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.732844 kubelet[2486]: I0913 01:02:16.732821 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a53d47bed871bd0d7b3b0a521943be9d-kubeconfig\") pod \"kube-scheduler-srv-ngi99.gb1.brightbox.com\" (UID: \"a53d47bed871bd0d7b3b0a521943be9d\") " pod="kube-system/kube-scheduler-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.733223 kubelet[2486]: I0913 01:02:16.732857 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-ca-certs\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.733223 kubelet[2486]: I0913 01:02:16.732887 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-k8s-certs\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.733223 kubelet[2486]: I0913 01:02:16.732914 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-usr-share-ca-certificates\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.733223 kubelet[2486]: I0913 01:02:16.732944 2486 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-ca-certs\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.789510 kubelet[2486]: I0913 01:02:16.789109 2486 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.789656 kubelet[2486]: E0913 01:02:16.789550 2486 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.230.35.122:6443/api/v1/nodes\": dial tcp 10.230.35.122:6443: connect: connection refused" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:16.930787 kubelet[2486]: E0913 01:02:16.930727 2486 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.35.122:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-ngi99.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.35.122:6443: connect: connection refused" interval="800ms" Sep 13 01:02:16.962342 containerd[1618]: time="2025-09-13T01:02:16.962211643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-ngi99.gb1.brightbox.com,Uid:a2ea146fa73da3c3fd2833b24aba2394,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:16.963054 containerd[1618]: time="2025-09-13T01:02:16.962266574Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-ngi99.gb1.brightbox.com,Uid:a53d47bed871bd0d7b3b0a521943be9d,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:16.968679 containerd[1618]: time="2025-09-13T01:02:16.968371257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-ngi99.gb1.brightbox.com,Uid:5914b13a5433a0adb709f585296b69a1,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:17.192028 kubelet[2486]: I0913 01:02:17.191975 2486 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:17.192520 kubelet[2486]: E0913 01:02:17.192477 2486 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.230.35.122:6443/api/v1/nodes\": dial tcp 10.230.35.122:6443: connect: connection refused" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:17.417472 kubelet[2486]: W0913 01:02:17.417328 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.230.35.122:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-ngi99.gb1.brightbox.com&limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:17.417472 kubelet[2486]: E0913 01:02:17.417424 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.230.35.122:6443/api/v1/nodes?fieldSelector=metadata.name%3Dsrv-ngi99.gb1.brightbox.com&limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:17.528254 kubelet[2486]: W0913 01:02:17.527997 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.230.35.122:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:17.528254 kubelet[2486]: E0913 01:02:17.528089 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.230.35.122:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:17.619950 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2125054856.mount: Deactivated successfully. Sep 13 01:02:17.638513 containerd[1618]: time="2025-09-13T01:02:17.638428875Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:02:17.640351 containerd[1618]: time="2025-09-13T01:02:17.640277897Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Sep 13 01:02:17.641036 containerd[1618]: time="2025-09-13T01:02:17.640990906Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:02:17.644629 containerd[1618]: time="2025-09-13T01:02:17.642665852Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:02:17.645424 containerd[1618]: time="2025-09-13T01:02:17.645364857Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 13 01:02:17.649136 containerd[1618]: time="2025-09-13T01:02:17.649088139Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:02:17.649306 containerd[1618]: time="2025-09-13T01:02:17.649182634Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 13 01:02:17.655020 containerd[1618]: time="2025-09-13T01:02:17.654942330Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 13 01:02:17.656443 containerd[1618]: time="2025-09-13T01:02:17.656397602Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 693.366282ms" Sep 13 01:02:17.659881 containerd[1618]: time="2025-09-13T01:02:17.659830421Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 691.372897ms" Sep 13 01:02:17.664083 containerd[1618]: time="2025-09-13T01:02:17.664037308Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 701.59802ms" Sep 13 01:02:17.733966 kubelet[2486]: E0913 01:02:17.733856 2486 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.230.35.122:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/srv-ngi99.gb1.brightbox.com?timeout=10s\": dial tcp 10.230.35.122:6443: connect: connection refused" interval="1.6s" Sep 13 01:02:17.868651 containerd[1618]: time="2025-09-13T01:02:17.868330840Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.868430807Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.869136007Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.868795949Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.868871505Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.868895731Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.869384 containerd[1618]: time="2025-09-13T01:02:17.869058599Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.870680 containerd[1618]: time="2025-09-13T01:02:17.869320569Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.872397 containerd[1618]: time="2025-09-13T01:02:17.871806462Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:17.872397 containerd[1618]: time="2025-09-13T01:02:17.871865763Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:17.872397 containerd[1618]: time="2025-09-13T01:02:17.871883351Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.872397 containerd[1618]: time="2025-09-13T01:02:17.871993424Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:17.899773 kubelet[2486]: W0913 01:02:17.899489 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.230.35.122:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:17.899773 kubelet[2486]: E0913 01:02:17.899655 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.230.35.122:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:17.972975 kubelet[2486]: W0913 01:02:17.972868 2486 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.230.35.122:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.230.35.122:6443: connect: connection refused Sep 13 01:02:17.972975 kubelet[2486]: E0913 01:02:17.972969 2486 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.230.35.122:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:18.006539 kubelet[2486]: I0913 01:02:18.006491 2486 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:18.007357 containerd[1618]: time="2025-09-13T01:02:18.007160119Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-srv-ngi99.gb1.brightbox.com,Uid:5914b13a5433a0adb709f585296b69a1,Namespace:kube-system,Attempt:0,} returns sandbox id \"747936c56f4c93ce472a290bf1fe7bcd3e2c7a04961b92b05049da65e2fbc338\"" Sep 13 01:02:18.007855 kubelet[2486]: E0913 01:02:18.007555 2486 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.230.35.122:6443/api/v1/nodes\": dial tcp 10.230.35.122:6443: connect: connection refused" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:18.027482 containerd[1618]: time="2025-09-13T01:02:18.026090189Z" level=info msg="CreateContainer within sandbox \"747936c56f4c93ce472a290bf1fe7bcd3e2c7a04961b92b05049da65e2fbc338\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 13 01:02:18.045685 containerd[1618]: time="2025-09-13T01:02:18.045635417Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-srv-ngi99.gb1.brightbox.com,Uid:a53d47bed871bd0d7b3b0a521943be9d,Namespace:kube-system,Attempt:0,} returns sandbox id \"e3fd538322ac5a9f28fed9d446ab8b301f02816bf8c1c8aee91b30294dd9a76e\"" Sep 13 01:02:18.051840 containerd[1618]: time="2025-09-13T01:02:18.051542349Z" level=info msg="CreateContainer within sandbox \"e3fd538322ac5a9f28fed9d446ab8b301f02816bf8c1c8aee91b30294dd9a76e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 13 01:02:18.055087 containerd[1618]: time="2025-09-13T01:02:18.055048628Z" level=info msg="CreateContainer within sandbox \"747936c56f4c93ce472a290bf1fe7bcd3e2c7a04961b92b05049da65e2fbc338\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"404f474bddce681b639fce23b23ddf7213bf11e812c3ec92851ae20b2af98ae0\"" Sep 13 01:02:18.055537 containerd[1618]: time="2025-09-13T01:02:18.055482153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-srv-ngi99.gb1.brightbox.com,Uid:a2ea146fa73da3c3fd2833b24aba2394,Namespace:kube-system,Attempt:0,} returns sandbox id \"e8a870a4e125a3edb48c07b69b15884f2265672c8425b484729a09719cd3a3e4\"" Sep 13 01:02:18.057379 containerd[1618]: time="2025-09-13T01:02:18.056732578Z" level=info msg="StartContainer for \"404f474bddce681b639fce23b23ddf7213bf11e812c3ec92851ae20b2af98ae0\"" Sep 13 01:02:18.061531 containerd[1618]: time="2025-09-13T01:02:18.061487691Z" level=info msg="CreateContainer within sandbox \"e8a870a4e125a3edb48c07b69b15884f2265672c8425b484729a09719cd3a3e4\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 13 01:02:18.097698 containerd[1618]: time="2025-09-13T01:02:18.097633459Z" level=info msg="CreateContainer within sandbox \"e3fd538322ac5a9f28fed9d446ab8b301f02816bf8c1c8aee91b30294dd9a76e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ebd1bf733ad66ffc5f3a15c3b0ee80867c576b5b72d1d846e5f5b9850046474b\"" Sep 13 01:02:18.099400 containerd[1618]: time="2025-09-13T01:02:18.098150341Z" level=info msg="StartContainer for \"ebd1bf733ad66ffc5f3a15c3b0ee80867c576b5b72d1d846e5f5b9850046474b\"" Sep 13 01:02:18.103460 containerd[1618]: time="2025-09-13T01:02:18.103410622Z" level=info msg="CreateContainer within sandbox \"e8a870a4e125a3edb48c07b69b15884f2265672c8425b484729a09719cd3a3e4\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"e4ebb8f542799db95c398b95a56c82fd7c02cc11ca8a93f0f73a4b201db5dd36\"" Sep 13 01:02:18.104383 containerd[1618]: time="2025-09-13T01:02:18.104287102Z" level=info msg="StartContainer for \"e4ebb8f542799db95c398b95a56c82fd7c02cc11ca8a93f0f73a4b201db5dd36\"" Sep 13 01:02:18.188791 containerd[1618]: time="2025-09-13T01:02:18.188733279Z" level=info msg="StartContainer for \"404f474bddce681b639fce23b23ddf7213bf11e812c3ec92851ae20b2af98ae0\" returns successfully" Sep 13 01:02:18.278316 kubelet[2486]: E0913 01:02:18.278240 2486 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.230.35.122:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.230.35.122:6443: connect: connection refused" logger="UnhandledError" Sep 13 01:02:18.278576 containerd[1618]: time="2025-09-13T01:02:18.278523231Z" level=info msg="StartContainer for \"e4ebb8f542799db95c398b95a56c82fd7c02cc11ca8a93f0f73a4b201db5dd36\" returns successfully" Sep 13 01:02:18.362246 containerd[1618]: time="2025-09-13T01:02:18.361932687Z" level=info msg="StartContainer for \"ebd1bf733ad66ffc5f3a15c3b0ee80867c576b5b72d1d846e5f5b9850046474b\" returns successfully" Sep 13 01:02:19.614275 kubelet[2486]: I0913 01:02:19.614233 2486 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:21.138104 kubelet[2486]: E0913 01:02:21.137907 2486 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"srv-ngi99.gb1.brightbox.com\" not found" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:21.187196 kubelet[2486]: I0913 01:02:21.186715 2486 kubelet_node_status.go:75] "Successfully registered node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:21.187196 kubelet[2486]: E0913 01:02:21.186775 2486 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"srv-ngi99.gb1.brightbox.com\": node \"srv-ngi99.gb1.brightbox.com\" not found" Sep 13 01:02:21.227463 kubelet[2486]: E0913 01:02:21.227257 2486 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{srv-ngi99.gb1.brightbox.com.1864b1e16b3c7d55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:srv-ngi99.gb1.brightbox.com,UID:srv-ngi99.gb1.brightbox.com,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:srv-ngi99.gb1.brightbox.com,},FirstTimestamp:2025-09-13 01:02:16.306220373 +0000 UTC m=+0.855848108,LastTimestamp:2025-09-13 01:02:16.306220373 +0000 UTC m=+0.855848108,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:srv-ngi99.gb1.brightbox.com,}" Sep 13 01:02:21.300647 kubelet[2486]: I0913 01:02:21.300539 2486 apiserver.go:52] "Watching apiserver" Sep 13 01:02:21.330814 kubelet[2486]: I0913 01:02:21.330747 2486 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 01:02:22.167343 kubelet[2486]: W0913 01:02:22.167236 2486 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:02:25.034565 systemd[1]: Reloading requested from client PID 2761 ('systemctl') (unit session-11.scope)... Sep 13 01:02:25.034648 systemd[1]: Reloading... Sep 13 01:02:25.151650 zram_generator::config[2798]: No configuration found. Sep 13 01:02:25.399752 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 01:02:25.523142 systemd[1]: Reloading finished in 487 ms. Sep 13 01:02:25.580727 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:25.597373 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 01:02:25.599072 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:25.608144 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 13 01:02:25.915917 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 13 01:02:25.917647 (kubelet)[2876]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 13 01:02:26.012451 kubelet[2876]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:02:26.012451 kubelet[2876]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 01:02:26.012451 kubelet[2876]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 01:02:26.012451 kubelet[2876]: I0913 01:02:26.010982 2876 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 01:02:26.023943 kubelet[2876]: I0913 01:02:26.023732 2876 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 01:02:26.023943 kubelet[2876]: I0913 01:02:26.023776 2876 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 01:02:26.024177 kubelet[2876]: I0913 01:02:26.024103 2876 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 01:02:26.026282 kubelet[2876]: I0913 01:02:26.026231 2876 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 13 01:02:26.034897 kubelet[2876]: I0913 01:02:26.034773 2876 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 01:02:26.044477 kubelet[2876]: E0913 01:02:26.044428 2876 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 01:02:26.044477 kubelet[2876]: I0913 01:02:26.044477 2876 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 01:02:26.050296 kubelet[2876]: I0913 01:02:26.050266 2876 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 01:02:26.050953 kubelet[2876]: I0913 01:02:26.050780 2876 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 01:02:26.051157 kubelet[2876]: I0913 01:02:26.050981 2876 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 01:02:26.051312 kubelet[2876]: I0913 01:02:26.051024 2876 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"srv-ngi99.gb1.brightbox.com","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Sep 13 01:02:26.053079 kubelet[2876]: I0913 01:02:26.053028 2876 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 01:02:26.053079 kubelet[2876]: I0913 01:02:26.053073 2876 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 01:02:26.053223 kubelet[2876]: I0913 01:02:26.053150 2876 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:02:26.057005 kubelet[2876]: I0913 01:02:26.056331 2876 kubelet.go:408] "Attempting to sync node with API server" Sep 13 01:02:26.057005 kubelet[2876]: I0913 01:02:26.057003 2876 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 01:02:26.057165 kubelet[2876]: I0913 01:02:26.057133 2876 kubelet.go:314] "Adding apiserver pod source" Sep 13 01:02:26.059083 kubelet[2876]: I0913 01:02:26.057173 2876 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 01:02:26.065815 kubelet[2876]: I0913 01:02:26.065778 2876 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Sep 13 01:02:26.066432 kubelet[2876]: I0913 01:02:26.066406 2876 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 01:02:26.074115 kubelet[2876]: I0913 01:02:26.072724 2876 server.go:1274] "Started kubelet" Sep 13 01:02:26.086167 kubelet[2876]: I0913 01:02:26.085907 2876 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 01:02:26.087741 kubelet[2876]: I0913 01:02:26.086773 2876 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 01:02:26.092412 kubelet[2876]: I0913 01:02:26.092367 2876 server.go:449] "Adding debug handlers to kubelet server" Sep 13 01:02:26.102448 kubelet[2876]: I0913 01:02:26.101807 2876 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 01:02:26.111643 kubelet[2876]: I0913 01:02:26.108986 2876 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 01:02:26.113753 kubelet[2876]: I0913 01:02:26.113557 2876 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 01:02:26.123154 kubelet[2876]: I0913 01:02:26.122465 2876 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 01:02:26.123154 kubelet[2876]: I0913 01:02:26.122647 2876 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 01:02:26.123154 kubelet[2876]: I0913 01:02:26.122826 2876 reconciler.go:26] "Reconciler: start to sync state" Sep 13 01:02:26.127087 kubelet[2876]: I0913 01:02:26.126549 2876 factory.go:221] Registration of the systemd container factory successfully Sep 13 01:02:26.127087 kubelet[2876]: I0913 01:02:26.126686 2876 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 01:02:26.128964 kubelet[2876]: I0913 01:02:26.128776 2876 factory.go:221] Registration of the containerd container factory successfully Sep 13 01:02:26.139052 kubelet[2876]: I0913 01:02:26.139003 2876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 01:02:26.139372 kubelet[2876]: E0913 01:02:26.139339 2876 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 01:02:26.141114 kubelet[2876]: I0913 01:02:26.141090 2876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 01:02:26.141276 kubelet[2876]: I0913 01:02:26.141257 2876 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 01:02:26.141421 kubelet[2876]: I0913 01:02:26.141400 2876 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 01:02:26.141623 kubelet[2876]: E0913 01:02:26.141571 2876 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 01:02:26.199486 sudo[2909]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 13 01:02:26.200443 sudo[2909]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 13 01:02:26.243637 kubelet[2876]: E0913 01:02:26.242025 2876 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 01:02:26.286134 kubelet[2876]: I0913 01:02:26.286086 2876 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 01:02:26.286134 kubelet[2876]: I0913 01:02:26.286127 2876 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 01:02:26.286465 kubelet[2876]: I0913 01:02:26.286163 2876 state_mem.go:36] "Initialized new in-memory state store" Sep 13 01:02:26.286465 kubelet[2876]: I0913 01:02:26.286396 2876 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 13 01:02:26.286465 kubelet[2876]: I0913 01:02:26.286417 2876 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 13 01:02:26.286465 kubelet[2876]: I0913 01:02:26.286455 2876 policy_none.go:49] "None policy: Start" Sep 13 01:02:26.288368 kubelet[2876]: I0913 01:02:26.288303 2876 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 01:02:26.288477 kubelet[2876]: I0913 01:02:26.288454 2876 state_mem.go:35] "Initializing new in-memory state store" Sep 13 01:02:26.288773 kubelet[2876]: I0913 01:02:26.288748 2876 state_mem.go:75] "Updated machine memory state" Sep 13 01:02:26.296066 kubelet[2876]: I0913 01:02:26.294133 2876 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 01:02:26.296066 kubelet[2876]: I0913 01:02:26.294526 2876 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 01:02:26.296066 kubelet[2876]: I0913 01:02:26.294577 2876 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 01:02:26.297114 kubelet[2876]: I0913 01:02:26.297076 2876 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 01:02:26.417134 kubelet[2876]: I0913 01:02:26.417076 2876 kubelet_node_status.go:72] "Attempting to register node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.453682 kubelet[2876]: I0913 01:02:26.453523 2876 kubelet_node_status.go:111] "Node was previously registered" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.453682 kubelet[2876]: I0913 01:02:26.453650 2876 kubelet_node_status.go:75] "Successfully registered node" node="srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.469453 kubelet[2876]: W0913 01:02:26.469409 2876 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:02:26.472856 kubelet[2876]: W0913 01:02:26.471433 2876 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:02:26.477812 kubelet[2876]: W0913 01:02:26.477780 2876 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 13 01:02:26.477936 kubelet[2876]: E0913 01:02:26.477858 2876 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" already exists" pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624059 kubelet[2876]: I0913 01:02:26.623957 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-ca-certs\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624059 kubelet[2876]: I0913 01:02:26.624022 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-k8s-certs\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624059 kubelet[2876]: I0913 01:02:26.624055 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-ca-certs\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624337 kubelet[2876]: I0913 01:02:26.624100 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-k8s-certs\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624337 kubelet[2876]: I0913 01:02:26.624155 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-kubeconfig\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624337 kubelet[2876]: I0913 01:02:26.624202 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-usr-share-ca-certificates\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624337 kubelet[2876]: I0913 01:02:26.624233 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a53d47bed871bd0d7b3b0a521943be9d-kubeconfig\") pod \"kube-scheduler-srv-ngi99.gb1.brightbox.com\" (UID: \"a53d47bed871bd0d7b3b0a521943be9d\") " pod="kube-system/kube-scheduler-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624337 kubelet[2876]: I0913 01:02:26.624262 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a2ea146fa73da3c3fd2833b24aba2394-usr-share-ca-certificates\") pod \"kube-apiserver-srv-ngi99.gb1.brightbox.com\" (UID: \"a2ea146fa73da3c3fd2833b24aba2394\") " pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:26.624545 kubelet[2876]: I0913 01:02:26.624289 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5914b13a5433a0adb709f585296b69a1-flexvolume-dir\") pod \"kube-controller-manager-srv-ngi99.gb1.brightbox.com\" (UID: \"5914b13a5433a0adb709f585296b69a1\") " pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" Sep 13 01:02:27.043755 sudo[2909]: pam_unix(sudo:session): session closed for user root Sep 13 01:02:27.062888 kubelet[2876]: I0913 01:02:27.062716 2876 apiserver.go:52] "Watching apiserver" Sep 13 01:02:27.122855 kubelet[2876]: I0913 01:02:27.122784 2876 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 01:02:27.269446 kubelet[2876]: I0913 01:02:27.269349 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-srv-ngi99.gb1.brightbox.com" podStartSLOduration=1.26931652 podStartE2EDuration="1.26931652s" podCreationTimestamp="2025-09-13 01:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:27.249882051 +0000 UTC m=+1.320137092" watchObservedRunningTime="2025-09-13 01:02:27.26931652 +0000 UTC m=+1.339571542" Sep 13 01:02:27.295361 kubelet[2876]: I0913 01:02:27.295010 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-srv-ngi99.gb1.brightbox.com" podStartSLOduration=5.29498094 podStartE2EDuration="5.29498094s" podCreationTimestamp="2025-09-13 01:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:27.271774339 +0000 UTC m=+1.342029390" watchObservedRunningTime="2025-09-13 01:02:27.29498094 +0000 UTC m=+1.365235971" Sep 13 01:02:27.295361 kubelet[2876]: I0913 01:02:27.295125 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-srv-ngi99.gb1.brightbox.com" podStartSLOduration=1.295117015 podStartE2EDuration="1.295117015s" podCreationTimestamp="2025-09-13 01:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:27.295092355 +0000 UTC m=+1.365347395" watchObservedRunningTime="2025-09-13 01:02:27.295117015 +0000 UTC m=+1.365372051" Sep 13 01:02:28.944135 sudo[1934]: pam_unix(sudo:session): session closed for user root Sep 13 01:02:29.092079 sshd[1930]: pam_unix(sshd:session): session closed for user core Sep 13 01:02:29.097056 systemd[1]: sshd@8-10.230.35.122:22-139.178.68.195:50646.service: Deactivated successfully. Sep 13 01:02:29.104527 systemd-logind[1597]: Session 11 logged out. Waiting for processes to exit. Sep 13 01:02:29.106266 systemd[1]: session-11.scope: Deactivated successfully. Sep 13 01:02:29.108072 systemd-logind[1597]: Removed session 11. Sep 13 01:02:29.344685 kubelet[2876]: I0913 01:02:29.344190 2876 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 13 01:02:29.346220 containerd[1618]: time="2025-09-13T01:02:29.346037657Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 13 01:02:29.348427 kubelet[2876]: I0913 01:02:29.346358 2876 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 13 01:02:29.446591 kubelet[2876]: I0913 01:02:29.442351 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3217f83c-673c-4c95-9ca5-2b2c1c79958d-xtables-lock\") pod \"kube-proxy-bh5tq\" (UID: \"3217f83c-673c-4c95-9ca5-2b2c1c79958d\") " pod="kube-system/kube-proxy-bh5tq" Sep 13 01:02:29.446591 kubelet[2876]: I0913 01:02:29.442863 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3217f83c-673c-4c95-9ca5-2b2c1c79958d-lib-modules\") pod \"kube-proxy-bh5tq\" (UID: \"3217f83c-673c-4c95-9ca5-2b2c1c79958d\") " pod="kube-system/kube-proxy-bh5tq" Sep 13 01:02:29.446591 kubelet[2876]: I0913 01:02:29.442893 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgsdx\" (UniqueName: \"kubernetes.io/projected/3217f83c-673c-4c95-9ca5-2b2c1c79958d-kube-api-access-mgsdx\") pod \"kube-proxy-bh5tq\" (UID: \"3217f83c-673c-4c95-9ca5-2b2c1c79958d\") " pod="kube-system/kube-proxy-bh5tq" Sep 13 01:02:29.446591 kubelet[2876]: I0913 01:02:29.442927 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3217f83c-673c-4c95-9ca5-2b2c1c79958d-kube-proxy\") pod \"kube-proxy-bh5tq\" (UID: \"3217f83c-673c-4c95-9ca5-2b2c1c79958d\") " pod="kube-system/kube-proxy-bh5tq" Sep 13 01:02:29.552625 kubelet[2876]: I0913 01:02:29.547197 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-hostproc\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.553982 kubelet[2876]: I0913 01:02:29.552996 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-etc-cni-netd\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.553982 kubelet[2876]: I0913 01:02:29.553080 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w497g\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-kube-api-access-w497g\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.553982 kubelet[2876]: I0913 01:02:29.553204 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/48712337-4ca7-4808-8468-3fcbea712daa-cilium-config-path\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557813 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-run\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557860 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-kernel\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557906 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-cgroup\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557934 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-net\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557961 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-hubble-tls\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559620 kubelet[2876]: I0913 01:02:29.557988 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cni-path\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559980 kubelet[2876]: I0913 01:02:29.558027 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-bpf-maps\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559980 kubelet[2876]: I0913 01:02:29.558052 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-lib-modules\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559980 kubelet[2876]: I0913 01:02:29.558078 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-xtables-lock\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.559980 kubelet[2876]: I0913 01:02:29.558103 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/48712337-4ca7-4808-8468-3fcbea712daa-clustermesh-secrets\") pod \"cilium-t292t\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " pod="kube-system/cilium-t292t" Sep 13 01:02:29.755783 containerd[1618]: time="2025-09-13T01:02:29.755730040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bh5tq,Uid:3217f83c-673c-4c95-9ca5-2b2c1c79958d,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:29.779308 containerd[1618]: time="2025-09-13T01:02:29.779005243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t292t,Uid:48712337-4ca7-4808-8468-3fcbea712daa,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:29.808574 containerd[1618]: time="2025-09-13T01:02:29.807995165Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:29.814358 containerd[1618]: time="2025-09-13T01:02:29.814020718Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:29.814358 containerd[1618]: time="2025-09-13T01:02:29.814062846Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:29.815577 containerd[1618]: time="2025-09-13T01:02:29.815526737Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:29.836161 containerd[1618]: time="2025-09-13T01:02:29.836020784Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:29.838516 containerd[1618]: time="2025-09-13T01:02:29.838266546Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:29.838516 containerd[1618]: time="2025-09-13T01:02:29.838297554Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:29.838516 containerd[1618]: time="2025-09-13T01:02:29.838420637Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:29.922854 containerd[1618]: time="2025-09-13T01:02:29.922682991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bh5tq,Uid:3217f83c-673c-4c95-9ca5-2b2c1c79958d,Namespace:kube-system,Attempt:0,} returns sandbox id \"8801d6eb8772647a69d84d64b3465d37b2475eeefdddea4183b3455e03e5cb06\"" Sep 13 01:02:29.932788 containerd[1618]: time="2025-09-13T01:02:29.932692895Z" level=info msg="CreateContainer within sandbox \"8801d6eb8772647a69d84d64b3465d37b2475eeefdddea4183b3455e03e5cb06\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 13 01:02:29.937585 containerd[1618]: time="2025-09-13T01:02:29.937289893Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t292t,Uid:48712337-4ca7-4808-8468-3fcbea712daa,Namespace:kube-system,Attempt:0,} returns sandbox id \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\"" Sep 13 01:02:29.939904 containerd[1618]: time="2025-09-13T01:02:29.939559758Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 13 01:02:29.968519 containerd[1618]: time="2025-09-13T01:02:29.968466808Z" level=info msg="CreateContainer within sandbox \"8801d6eb8772647a69d84d64b3465d37b2475eeefdddea4183b3455e03e5cb06\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"38597a89db20da0136dff96ec7986486e7e39c967f4739165b62f52e9d333e26\"" Sep 13 01:02:29.970802 containerd[1618]: time="2025-09-13T01:02:29.970639905Z" level=info msg="StartContainer for \"38597a89db20da0136dff96ec7986486e7e39c967f4739165b62f52e9d333e26\"" Sep 13 01:02:30.088147 containerd[1618]: time="2025-09-13T01:02:30.087069104Z" level=info msg="StartContainer for \"38597a89db20da0136dff96ec7986486e7e39c967f4739165b62f52e9d333e26\" returns successfully" Sep 13 01:02:30.523886 kubelet[2876]: I0913 01:02:30.520785 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-bh5tq" podStartSLOduration=1.5207420630000001 podStartE2EDuration="1.520742063s" podCreationTimestamp="2025-09-13 01:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:30.284046952 +0000 UTC m=+4.354301991" watchObservedRunningTime="2025-09-13 01:02:30.520742063 +0000 UTC m=+4.590997091" Sep 13 01:02:30.564121 kubelet[2876]: I0913 01:02:30.564057 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7d6b\" (UniqueName: \"kubernetes.io/projected/cd73d6ce-5f91-43b9-8990-ced1697df22a-kube-api-access-k7d6b\") pod \"cilium-operator-5d85765b45-k4pxv\" (UID: \"cd73d6ce-5f91-43b9-8990-ced1697df22a\") " pod="kube-system/cilium-operator-5d85765b45-k4pxv" Sep 13 01:02:30.564420 kubelet[2876]: I0913 01:02:30.564394 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd73d6ce-5f91-43b9-8990-ced1697df22a-cilium-config-path\") pod \"cilium-operator-5d85765b45-k4pxv\" (UID: \"cd73d6ce-5f91-43b9-8990-ced1697df22a\") " pod="kube-system/cilium-operator-5d85765b45-k4pxv" Sep 13 01:02:30.831643 containerd[1618]: time="2025-09-13T01:02:30.830556426Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-k4pxv,Uid:cd73d6ce-5f91-43b9-8990-ced1697df22a,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:30.899525 containerd[1618]: time="2025-09-13T01:02:30.897203658Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:30.899525 containerd[1618]: time="2025-09-13T01:02:30.899109339Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:30.899525 containerd[1618]: time="2025-09-13T01:02:30.899132976Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:30.899525 containerd[1618]: time="2025-09-13T01:02:30.899285691Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:31.020281 containerd[1618]: time="2025-09-13T01:02:31.020206122Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-k4pxv,Uid:cd73d6ce-5f91-43b9-8990-ced1697df22a,Namespace:kube-system,Attempt:0,} returns sandbox id \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\"" Sep 13 01:02:37.053060 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2497470196.mount: Deactivated successfully. Sep 13 01:02:40.258678 containerd[1618]: time="2025-09-13T01:02:40.258467476Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:40.261100 containerd[1618]: time="2025-09-13T01:02:40.261040958Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 13 01:02:40.261680 containerd[1618]: time="2025-09-13T01:02:40.261380002Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:40.265182 containerd[1618]: time="2025-09-13T01:02:40.265125785Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 10.325493153s" Sep 13 01:02:40.265300 containerd[1618]: time="2025-09-13T01:02:40.265220026Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 13 01:02:40.268868 containerd[1618]: time="2025-09-13T01:02:40.268293650Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 13 01:02:40.269926 containerd[1618]: time="2025-09-13T01:02:40.269890467Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 01:02:40.360168 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2156079656.mount: Deactivated successfully. Sep 13 01:02:40.365013 containerd[1618]: time="2025-09-13T01:02:40.364846449Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\"" Sep 13 01:02:40.367350 containerd[1618]: time="2025-09-13T01:02:40.367317574Z" level=info msg="StartContainer for \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\"" Sep 13 01:02:40.535543 containerd[1618]: time="2025-09-13T01:02:40.535377340Z" level=info msg="StartContainer for \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\" returns successfully" Sep 13 01:02:40.867078 containerd[1618]: time="2025-09-13T01:02:40.855865285Z" level=info msg="shim disconnected" id=529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c namespace=k8s.io Sep 13 01:02:40.867078 containerd[1618]: time="2025-09-13T01:02:40.866185400Z" level=warning msg="cleaning up after shim disconnected" id=529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c namespace=k8s.io Sep 13 01:02:40.867078 containerd[1618]: time="2025-09-13T01:02:40.866224650Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:02:41.290546 containerd[1618]: time="2025-09-13T01:02:41.289554457Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 01:02:41.320948 containerd[1618]: time="2025-09-13T01:02:41.320882380Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\"" Sep 13 01:02:41.325923 containerd[1618]: time="2025-09-13T01:02:41.323788823Z" level=info msg="StartContainer for \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\"" Sep 13 01:02:41.362386 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c-rootfs.mount: Deactivated successfully. Sep 13 01:02:41.404744 systemd[1]: run-containerd-runc-k8s.io-6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c-runc.fsRfw1.mount: Deactivated successfully. Sep 13 01:02:41.461009 containerd[1618]: time="2025-09-13T01:02:41.460785945Z" level=info msg="StartContainer for \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\" returns successfully" Sep 13 01:02:41.470223 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 01:02:41.470819 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:02:41.470981 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:02:41.481247 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 13 01:02:41.521349 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c-rootfs.mount: Deactivated successfully. Sep 13 01:02:41.528471 containerd[1618]: time="2025-09-13T01:02:41.528197554Z" level=info msg="shim disconnected" id=6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c namespace=k8s.io Sep 13 01:02:41.528471 containerd[1618]: time="2025-09-13T01:02:41.528301489Z" level=warning msg="cleaning up after shim disconnected" id=6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c namespace=k8s.io Sep 13 01:02:41.528471 containerd[1618]: time="2025-09-13T01:02:41.528317783Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:02:41.541950 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 13 01:02:41.558631 containerd[1618]: time="2025-09-13T01:02:41.558527523Z" level=warning msg="cleanup warnings time=\"2025-09-13T01:02:41Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 13 01:02:42.294119 containerd[1618]: time="2025-09-13T01:02:42.294054456Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 01:02:42.356917 containerd[1618]: time="2025-09-13T01:02:42.356196590Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\"" Sep 13 01:02:42.358349 containerd[1618]: time="2025-09-13T01:02:42.357693165Z" level=info msg="StartContainer for \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\"" Sep 13 01:02:42.446428 containerd[1618]: time="2025-09-13T01:02:42.446376025Z" level=info msg="StartContainer for \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\" returns successfully" Sep 13 01:02:42.490491 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a-rootfs.mount: Deactivated successfully. Sep 13 01:02:42.504185 containerd[1618]: time="2025-09-13T01:02:42.504110506Z" level=info msg="shim disconnected" id=1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a namespace=k8s.io Sep 13 01:02:42.504613 containerd[1618]: time="2025-09-13T01:02:42.504185875Z" level=warning msg="cleaning up after shim disconnected" id=1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a namespace=k8s.io Sep 13 01:02:42.504613 containerd[1618]: time="2025-09-13T01:02:42.504203305Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:02:43.304632 containerd[1618]: time="2025-09-13T01:02:43.304195548Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 01:02:43.348892 containerd[1618]: time="2025-09-13T01:02:43.348838476Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\"" Sep 13 01:02:43.355585 containerd[1618]: time="2025-09-13T01:02:43.355509286Z" level=info msg="StartContainer for \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\"" Sep 13 01:02:43.357498 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3720186955.mount: Deactivated successfully. Sep 13 01:02:43.481626 containerd[1618]: time="2025-09-13T01:02:43.481331533Z" level=info msg="StartContainer for \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\" returns successfully" Sep 13 01:02:43.539146 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf-rootfs.mount: Deactivated successfully. Sep 13 01:02:43.629981 containerd[1618]: time="2025-09-13T01:02:43.629903639Z" level=info msg="shim disconnected" id=10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf namespace=k8s.io Sep 13 01:02:43.630317 containerd[1618]: time="2025-09-13T01:02:43.630282531Z" level=warning msg="cleaning up after shim disconnected" id=10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf namespace=k8s.io Sep 13 01:02:43.630432 containerd[1618]: time="2025-09-13T01:02:43.630407542Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:02:43.654079 containerd[1618]: time="2025-09-13T01:02:43.654019664Z" level=warning msg="cleanup warnings time=\"2025-09-13T01:02:43Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 13 01:02:43.933917 containerd[1618]: time="2025-09-13T01:02:43.933768674Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:43.935045 containerd[1618]: time="2025-09-13T01:02:43.934985696Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 13 01:02:43.936017 containerd[1618]: time="2025-09-13T01:02:43.935536631Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 13 01:02:43.937829 containerd[1618]: time="2025-09-13T01:02:43.937785538Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.6694096s" Sep 13 01:02:43.937985 containerd[1618]: time="2025-09-13T01:02:43.937952490Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 13 01:02:43.942411 containerd[1618]: time="2025-09-13T01:02:43.942376120Z" level=info msg="CreateContainer within sandbox \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 13 01:02:43.965803 containerd[1618]: time="2025-09-13T01:02:43.965737184Z" level=info msg="CreateContainer within sandbox \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\"" Sep 13 01:02:43.967839 containerd[1618]: time="2025-09-13T01:02:43.966795345Z" level=info msg="StartContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\"" Sep 13 01:02:44.083563 containerd[1618]: time="2025-09-13T01:02:44.083491481Z" level=info msg="StartContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" returns successfully" Sep 13 01:02:44.323727 containerd[1618]: time="2025-09-13T01:02:44.322676302Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 01:02:44.377638 containerd[1618]: time="2025-09-13T01:02:44.373885131Z" level=info msg="CreateContainer within sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\"" Sep 13 01:02:44.383624 containerd[1618]: time="2025-09-13T01:02:44.383554569Z" level=info msg="StartContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\"" Sep 13 01:02:44.599212 containerd[1618]: time="2025-09-13T01:02:44.599085107Z" level=info msg="StartContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" returns successfully" Sep 13 01:02:44.607066 kubelet[2876]: I0913 01:02:44.606929 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-k4pxv" podStartSLOduration=1.690928964 podStartE2EDuration="14.606844481s" podCreationTimestamp="2025-09-13 01:02:30 +0000 UTC" firstStartedPulling="2025-09-13 01:02:31.022993088 +0000 UTC m=+5.093248116" lastFinishedPulling="2025-09-13 01:02:43.938908598 +0000 UTC m=+18.009163633" observedRunningTime="2025-09-13 01:02:44.606797335 +0000 UTC m=+18.677052368" watchObservedRunningTime="2025-09-13 01:02:44.606844481 +0000 UTC m=+18.677099517" Sep 13 01:02:44.995664 kubelet[2876]: I0913 01:02:44.993932 2876 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 13 01:02:45.179358 kubelet[2876]: I0913 01:02:45.179307 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rvqz\" (UniqueName: \"kubernetes.io/projected/b1aac901-55a3-4373-a4f9-188104c8542d-kube-api-access-2rvqz\") pod \"coredns-7c65d6cfc9-27lw6\" (UID: \"b1aac901-55a3-4373-a4f9-188104c8542d\") " pod="kube-system/coredns-7c65d6cfc9-27lw6" Sep 13 01:02:45.180071 kubelet[2876]: I0913 01:02:45.179362 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1aac901-55a3-4373-a4f9-188104c8542d-config-volume\") pod \"coredns-7c65d6cfc9-27lw6\" (UID: \"b1aac901-55a3-4373-a4f9-188104c8542d\") " pod="kube-system/coredns-7c65d6cfc9-27lw6" Sep 13 01:02:45.282244 kubelet[2876]: I0913 01:02:45.281278 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k87b\" (UniqueName: \"kubernetes.io/projected/4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27-kube-api-access-5k87b\") pod \"coredns-7c65d6cfc9-q2rfb\" (UID: \"4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27\") " pod="kube-system/coredns-7c65d6cfc9-q2rfb" Sep 13 01:02:45.282244 kubelet[2876]: I0913 01:02:45.281369 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27-config-volume\") pod \"coredns-7c65d6cfc9-q2rfb\" (UID: \"4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27\") " pod="kube-system/coredns-7c65d6cfc9-q2rfb" Sep 13 01:02:45.466687 containerd[1618]: time="2025-09-13T01:02:45.466394317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-27lw6,Uid:b1aac901-55a3-4373-a4f9-188104c8542d,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:45.510937 systemd-journald[1177]: Under memory pressure, flushing caches. Sep 13 01:02:45.492696 systemd-resolved[1507]: Under memory pressure, flushing caches. Sep 13 01:02:45.492772 systemd-resolved[1507]: Flushed all caches. Sep 13 01:02:45.521669 kubelet[2876]: I0913 01:02:45.520334 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-t292t" podStartSLOduration=6.192295723 podStartE2EDuration="16.520312831s" podCreationTimestamp="2025-09-13 01:02:29 +0000 UTC" firstStartedPulling="2025-09-13 01:02:29.938759395 +0000 UTC m=+4.009014422" lastFinishedPulling="2025-09-13 01:02:40.266776494 +0000 UTC m=+14.337031530" observedRunningTime="2025-09-13 01:02:45.488514218 +0000 UTC m=+19.558769259" watchObservedRunningTime="2025-09-13 01:02:45.520312831 +0000 UTC m=+19.590567866" Sep 13 01:02:45.788274 containerd[1618]: time="2025-09-13T01:02:45.788117888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-q2rfb,Uid:4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27,Namespace:kube-system,Attempt:0,}" Sep 13 01:02:48.025229 systemd-networkd[1259]: cilium_host: Link UP Sep 13 01:02:48.025546 systemd-networkd[1259]: cilium_net: Link UP Sep 13 01:02:48.025552 systemd-networkd[1259]: cilium_net: Gained carrier Sep 13 01:02:48.028706 systemd-networkd[1259]: cilium_host: Gained carrier Sep 13 01:02:48.215508 systemd-networkd[1259]: cilium_vxlan: Link UP Sep 13 01:02:48.215519 systemd-networkd[1259]: cilium_vxlan: Gained carrier Sep 13 01:02:48.241963 systemd-networkd[1259]: cilium_net: Gained IPv6LL Sep 13 01:02:48.690566 systemd-networkd[1259]: cilium_host: Gained IPv6LL Sep 13 01:02:48.784319 kernel: NET: Registered PF_ALG protocol family Sep 13 01:02:49.265932 systemd-networkd[1259]: cilium_vxlan: Gained IPv6LL Sep 13 01:02:49.856238 systemd-networkd[1259]: lxc_health: Link UP Sep 13 01:02:49.871846 systemd-networkd[1259]: lxc_health: Gained carrier Sep 13 01:02:50.174579 systemd-networkd[1259]: lxcd6348518a361: Link UP Sep 13 01:02:50.179624 kernel: eth0: renamed from tmp3040e Sep 13 01:02:50.187796 systemd-networkd[1259]: lxcd6348518a361: Gained carrier Sep 13 01:02:50.400420 systemd-networkd[1259]: lxcb46be66ab488: Link UP Sep 13 01:02:50.414776 kernel: eth0: renamed from tmpe4360 Sep 13 01:02:50.421041 systemd-networkd[1259]: lxcb46be66ab488: Gained carrier Sep 13 01:02:51.249848 systemd-networkd[1259]: lxc_health: Gained IPv6LL Sep 13 01:02:51.314839 systemd-networkd[1259]: lxcd6348518a361: Gained IPv6LL Sep 13 01:02:52.209885 systemd-networkd[1259]: lxcb46be66ab488: Gained IPv6LL Sep 13 01:02:56.058326 containerd[1618]: time="2025-09-13T01:02:56.048004784Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:56.058326 containerd[1618]: time="2025-09-13T01:02:56.048150744Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:56.058326 containerd[1618]: time="2025-09-13T01:02:56.048183529Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:56.058326 containerd[1618]: time="2025-09-13T01:02:56.048363840Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:56.084764 containerd[1618]: time="2025-09-13T01:02:56.057399588Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:02:56.084764 containerd[1618]: time="2025-09-13T01:02:56.059937591Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:02:56.084764 containerd[1618]: time="2025-09-13T01:02:56.059961290Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:56.084764 containerd[1618]: time="2025-09-13T01:02:56.060111851Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:02:56.122106 systemd[1]: run-containerd-runc-k8s.io-3040ec1802f236c1f6acae940297f4aa82d45c155fcee622c09f82f24ca0712e-runc.8YSE9u.mount: Deactivated successfully. Sep 13 01:02:56.249216 containerd[1618]: time="2025-09-13T01:02:56.248936559Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-q2rfb,Uid:4981b7b8-bb8b-4ff5-adb4-12b6fbadbe27,Namespace:kube-system,Attempt:0,} returns sandbox id \"e43602aca9208af6b975bd008a6795336f98b286d343694f2e8a1b490b90abfe\"" Sep 13 01:02:56.264307 containerd[1618]: time="2025-09-13T01:02:56.263798036Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-27lw6,Uid:b1aac901-55a3-4373-a4f9-188104c8542d,Namespace:kube-system,Attempt:0,} returns sandbox id \"3040ec1802f236c1f6acae940297f4aa82d45c155fcee622c09f82f24ca0712e\"" Sep 13 01:02:56.269632 containerd[1618]: time="2025-09-13T01:02:56.269224889Z" level=info msg="CreateContainer within sandbox \"e43602aca9208af6b975bd008a6795336f98b286d343694f2e8a1b490b90abfe\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 01:02:56.272561 containerd[1618]: time="2025-09-13T01:02:56.271895736Z" level=info msg="CreateContainer within sandbox \"3040ec1802f236c1f6acae940297f4aa82d45c155fcee622c09f82f24ca0712e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 01:02:56.301702 containerd[1618]: time="2025-09-13T01:02:56.301643435Z" level=info msg="CreateContainer within sandbox \"3040ec1802f236c1f6acae940297f4aa82d45c155fcee622c09f82f24ca0712e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9cea2ef0e240af5873f7c04f7bdb590d6157882c18f4caf317dd166b6da1c8b5\"" Sep 13 01:02:56.301964 containerd[1618]: time="2025-09-13T01:02:56.301925815Z" level=info msg="CreateContainer within sandbox \"e43602aca9208af6b975bd008a6795336f98b286d343694f2e8a1b490b90abfe\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"25f1ca840cfa05256410db2dd8fb1f8e7dc89478fdc94744366fe31976ea8c9d\"" Sep 13 01:02:56.303104 containerd[1618]: time="2025-09-13T01:02:56.303064916Z" level=info msg="StartContainer for \"9cea2ef0e240af5873f7c04f7bdb590d6157882c18f4caf317dd166b6da1c8b5\"" Sep 13 01:02:56.304719 containerd[1618]: time="2025-09-13T01:02:56.303826008Z" level=info msg="StartContainer for \"25f1ca840cfa05256410db2dd8fb1f8e7dc89478fdc94744366fe31976ea8c9d\"" Sep 13 01:02:56.421639 containerd[1618]: time="2025-09-13T01:02:56.421499893Z" level=info msg="StartContainer for \"25f1ca840cfa05256410db2dd8fb1f8e7dc89478fdc94744366fe31976ea8c9d\" returns successfully" Sep 13 01:02:56.431476 containerd[1618]: time="2025-09-13T01:02:56.431437765Z" level=info msg="StartContainer for \"9cea2ef0e240af5873f7c04f7bdb590d6157882c18f4caf317dd166b6da1c8b5\" returns successfully" Sep 13 01:02:57.420746 kubelet[2876]: I0913 01:02:57.420095 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-27lw6" podStartSLOduration=27.41998669 podStartE2EDuration="27.41998669s" podCreationTimestamp="2025-09-13 01:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:57.415324485 +0000 UTC m=+31.485579528" watchObservedRunningTime="2025-09-13 01:02:57.41998669 +0000 UTC m=+31.490241728" Sep 13 01:02:57.467159 kubelet[2876]: I0913 01:02:57.466839 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-q2rfb" podStartSLOduration=27.466691586 podStartE2EDuration="27.466691586s" podCreationTimestamp="2025-09-13 01:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:02:57.464557661 +0000 UTC m=+31.534812700" watchObservedRunningTime="2025-09-13 01:02:57.466691586 +0000 UTC m=+31.536946627" Sep 13 01:03:42.202242 systemd[1]: Started sshd@10-10.230.35.122:22-139.178.68.195:48452.service - OpenSSH per-connection server daemon (139.178.68.195:48452). Sep 13 01:03:43.141211 sshd[4256]: Accepted publickey for core from 139.178.68.195 port 48452 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:03:43.144366 sshd[4256]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:03:43.157882 systemd-logind[1597]: New session 12 of user core. Sep 13 01:03:43.166579 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 13 01:03:44.379459 sshd[4256]: pam_unix(sshd:session): session closed for user core Sep 13 01:03:44.390175 systemd[1]: sshd@10-10.230.35.122:22-139.178.68.195:48452.service: Deactivated successfully. Sep 13 01:03:44.397285 systemd[1]: session-12.scope: Deactivated successfully. Sep 13 01:03:44.400425 systemd-logind[1597]: Session 12 logged out. Waiting for processes to exit. Sep 13 01:03:44.406346 systemd-logind[1597]: Removed session 12. Sep 13 01:03:49.533094 systemd[1]: Started sshd@11-10.230.35.122:22-139.178.68.195:48464.service - OpenSSH per-connection server daemon (139.178.68.195:48464). Sep 13 01:03:50.427522 sshd[4273]: Accepted publickey for core from 139.178.68.195 port 48464 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:03:50.430733 sshd[4273]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:03:50.438764 systemd-logind[1597]: New session 13 of user core. Sep 13 01:03:50.449203 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 13 01:03:50.471476 systemd[1]: sshd@9-10.230.35.122:22-101.126.71.100:52868.service: Deactivated successfully. Sep 13 01:03:51.216977 sshd[4273]: pam_unix(sshd:session): session closed for user core Sep 13 01:03:51.222517 systemd[1]: sshd@11-10.230.35.122:22-139.178.68.195:48464.service: Deactivated successfully. Sep 13 01:03:51.228830 systemd-logind[1597]: Session 13 logged out. Waiting for processes to exit. Sep 13 01:03:51.230138 systemd[1]: session-13.scope: Deactivated successfully. Sep 13 01:03:51.232431 systemd-logind[1597]: Removed session 13. Sep 13 01:03:56.368999 systemd[1]: Started sshd@12-10.230.35.122:22-139.178.68.195:59286.service - OpenSSH per-connection server daemon (139.178.68.195:59286). Sep 13 01:03:57.272816 sshd[4290]: Accepted publickey for core from 139.178.68.195 port 59286 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:03:57.275019 sshd[4290]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:03:57.282875 systemd-logind[1597]: New session 14 of user core. Sep 13 01:03:57.292067 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 13 01:03:57.987214 sshd[4290]: pam_unix(sshd:session): session closed for user core Sep 13 01:03:57.997032 systemd[1]: sshd@12-10.230.35.122:22-139.178.68.195:59286.service: Deactivated successfully. Sep 13 01:03:58.000726 systemd-logind[1597]: Session 14 logged out. Waiting for processes to exit. Sep 13 01:03:58.001413 systemd[1]: session-14.scope: Deactivated successfully. Sep 13 01:03:58.005188 systemd-logind[1597]: Removed session 14. Sep 13 01:04:03.135953 systemd[1]: Started sshd@13-10.230.35.122:22-139.178.68.195:60724.service - OpenSSH per-connection server daemon (139.178.68.195:60724). Sep 13 01:04:04.031218 sshd[4307]: Accepted publickey for core from 139.178.68.195 port 60724 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:04.033907 sshd[4307]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:04.041460 systemd-logind[1597]: New session 15 of user core. Sep 13 01:04:04.051219 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 13 01:04:04.788700 sshd[4307]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:04.794045 systemd[1]: sshd@13-10.230.35.122:22-139.178.68.195:60724.service: Deactivated successfully. Sep 13 01:04:04.800455 systemd-logind[1597]: Session 15 logged out. Waiting for processes to exit. Sep 13 01:04:04.800920 systemd[1]: session-15.scope: Deactivated successfully. Sep 13 01:04:04.803353 systemd-logind[1597]: Removed session 15. Sep 13 01:04:04.944916 systemd[1]: Started sshd@14-10.230.35.122:22-139.178.68.195:60734.service - OpenSSH per-connection server daemon (139.178.68.195:60734). Sep 13 01:04:05.856443 sshd[4322]: Accepted publickey for core from 139.178.68.195 port 60734 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:05.859011 sshd[4322]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:05.865651 systemd-logind[1597]: New session 16 of user core. Sep 13 01:04:05.873115 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 13 01:04:06.631129 sshd[4322]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:06.638834 systemd[1]: sshd@14-10.230.35.122:22-139.178.68.195:60734.service: Deactivated successfully. Sep 13 01:04:06.643665 systemd[1]: session-16.scope: Deactivated successfully. Sep 13 01:04:06.644997 systemd-logind[1597]: Session 16 logged out. Waiting for processes to exit. Sep 13 01:04:06.647084 systemd-logind[1597]: Removed session 16. Sep 13 01:04:06.784092 systemd[1]: Started sshd@15-10.230.35.122:22-139.178.68.195:60748.service - OpenSSH per-connection server daemon (139.178.68.195:60748). Sep 13 01:04:07.672775 sshd[4334]: Accepted publickey for core from 139.178.68.195 port 60748 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:07.675331 sshd[4334]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:07.682668 systemd-logind[1597]: New session 17 of user core. Sep 13 01:04:07.688114 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 13 01:04:08.386963 sshd[4334]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:08.393426 systemd[1]: sshd@15-10.230.35.122:22-139.178.68.195:60748.service: Deactivated successfully. Sep 13 01:04:08.394763 systemd-logind[1597]: Session 17 logged out. Waiting for processes to exit. Sep 13 01:04:08.397772 systemd[1]: session-17.scope: Deactivated successfully. Sep 13 01:04:08.399221 systemd-logind[1597]: Removed session 17. Sep 13 01:04:13.543110 systemd[1]: Started sshd@16-10.230.35.122:22-139.178.68.195:42290.service - OpenSSH per-connection server daemon (139.178.68.195:42290). Sep 13 01:04:14.432787 sshd[4348]: Accepted publickey for core from 139.178.68.195 port 42290 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:14.435099 sshd[4348]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:14.442442 systemd-logind[1597]: New session 18 of user core. Sep 13 01:04:14.447017 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 13 01:04:15.137952 sshd[4348]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:15.143167 systemd[1]: sshd@16-10.230.35.122:22-139.178.68.195:42290.service: Deactivated successfully. Sep 13 01:04:15.148847 systemd-logind[1597]: Session 18 logged out. Waiting for processes to exit. Sep 13 01:04:15.149657 systemd[1]: session-18.scope: Deactivated successfully. Sep 13 01:04:15.151314 systemd-logind[1597]: Removed session 18. Sep 13 01:04:20.295083 systemd[1]: Started sshd@17-10.230.35.122:22-139.178.68.195:36260.service - OpenSSH per-connection server daemon (139.178.68.195:36260). Sep 13 01:04:21.187569 sshd[4362]: Accepted publickey for core from 139.178.68.195 port 36260 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:21.189965 sshd[4362]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:21.197975 systemd-logind[1597]: New session 19 of user core. Sep 13 01:04:21.203090 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 13 01:04:21.898356 sshd[4362]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:21.903103 systemd-logind[1597]: Session 19 logged out. Waiting for processes to exit. Sep 13 01:04:21.904296 systemd[1]: sshd@17-10.230.35.122:22-139.178.68.195:36260.service: Deactivated successfully. Sep 13 01:04:21.909952 systemd[1]: session-19.scope: Deactivated successfully. Sep 13 01:04:21.912252 systemd-logind[1597]: Removed session 19. Sep 13 01:04:22.047958 systemd[1]: Started sshd@18-10.230.35.122:22-139.178.68.195:36272.service - OpenSSH per-connection server daemon (139.178.68.195:36272). Sep 13 01:04:22.964172 sshd[4376]: Accepted publickey for core from 139.178.68.195 port 36272 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:22.966308 sshd[4376]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:22.973100 systemd-logind[1597]: New session 20 of user core. Sep 13 01:04:22.979500 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 13 01:04:24.077825 sshd[4376]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:24.088244 systemd[1]: sshd@18-10.230.35.122:22-139.178.68.195:36272.service: Deactivated successfully. Sep 13 01:04:24.093856 systemd[1]: session-20.scope: Deactivated successfully. Sep 13 01:04:24.095469 systemd-logind[1597]: Session 20 logged out. Waiting for processes to exit. Sep 13 01:04:24.097040 systemd-logind[1597]: Removed session 20. Sep 13 01:04:24.244987 systemd[1]: Started sshd@19-10.230.35.122:22-139.178.68.195:36282.service - OpenSSH per-connection server daemon (139.178.68.195:36282). Sep 13 01:04:25.144496 sshd[4388]: Accepted publickey for core from 139.178.68.195 port 36282 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:25.147198 sshd[4388]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:25.157704 systemd-logind[1597]: New session 21 of user core. Sep 13 01:04:25.165215 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 13 01:04:28.027274 sshd[4388]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:28.037196 systemd[1]: sshd@19-10.230.35.122:22-139.178.68.195:36282.service: Deactivated successfully. Sep 13 01:04:28.044693 systemd[1]: session-21.scope: Deactivated successfully. Sep 13 01:04:28.044898 systemd-logind[1597]: Session 21 logged out. Waiting for processes to exit. Sep 13 01:04:28.049269 systemd-logind[1597]: Removed session 21. Sep 13 01:04:28.197961 systemd[1]: Started sshd@20-10.230.35.122:22-139.178.68.195:36288.service - OpenSSH per-connection server daemon (139.178.68.195:36288). Sep 13 01:04:29.208033 sshd[4410]: Accepted publickey for core from 139.178.68.195 port 36288 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:29.210222 sshd[4410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:29.217826 systemd-logind[1597]: New session 22 of user core. Sep 13 01:04:29.223995 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 13 01:04:30.420630 sshd[4410]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:30.427470 systemd[1]: sshd@20-10.230.35.122:22-139.178.68.195:36288.service: Deactivated successfully. Sep 13 01:04:30.427686 systemd-logind[1597]: Session 22 logged out. Waiting for processes to exit. Sep 13 01:04:30.435454 systemd[1]: session-22.scope: Deactivated successfully. Sep 13 01:04:30.438663 systemd-logind[1597]: Removed session 22. Sep 13 01:04:30.570957 systemd[1]: Started sshd@21-10.230.35.122:22-139.178.68.195:35672.service - OpenSSH per-connection server daemon (139.178.68.195:35672). Sep 13 01:04:31.451664 sshd[4421]: Accepted publickey for core from 139.178.68.195 port 35672 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:31.453701 sshd[4421]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:31.460211 systemd-logind[1597]: New session 23 of user core. Sep 13 01:04:31.469149 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 13 01:04:32.145357 sshd[4421]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:32.150212 systemd-logind[1597]: Session 23 logged out. Waiting for processes to exit. Sep 13 01:04:32.150579 systemd[1]: sshd@21-10.230.35.122:22-139.178.68.195:35672.service: Deactivated successfully. Sep 13 01:04:32.156276 systemd[1]: session-23.scope: Deactivated successfully. Sep 13 01:04:32.157975 systemd-logind[1597]: Removed session 23. Sep 13 01:04:37.300068 systemd[1]: Started sshd@22-10.230.35.122:22-139.178.68.195:35686.service - OpenSSH per-connection server daemon (139.178.68.195:35686). Sep 13 01:04:38.189645 sshd[4440]: Accepted publickey for core from 139.178.68.195 port 35686 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:38.194990 sshd[4440]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:38.225903 systemd-logind[1597]: New session 24 of user core. Sep 13 01:04:38.231089 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 13 01:04:38.898187 sshd[4440]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:38.903089 systemd[1]: sshd@22-10.230.35.122:22-139.178.68.195:35686.service: Deactivated successfully. Sep 13 01:04:38.908225 systemd-logind[1597]: Session 24 logged out. Waiting for processes to exit. Sep 13 01:04:38.908920 systemd[1]: session-24.scope: Deactivated successfully. Sep 13 01:04:38.910517 systemd-logind[1597]: Removed session 24. Sep 13 01:04:44.048967 systemd[1]: Started sshd@23-10.230.35.122:22-139.178.68.195:59330.service - OpenSSH per-connection server daemon (139.178.68.195:59330). Sep 13 01:04:44.948393 sshd[4454]: Accepted publickey for core from 139.178.68.195 port 59330 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:44.950822 sshd[4454]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:44.958514 systemd-logind[1597]: New session 25 of user core. Sep 13 01:04:44.964118 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 13 01:04:45.669292 sshd[4454]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:45.674527 systemd[1]: sshd@23-10.230.35.122:22-139.178.68.195:59330.service: Deactivated successfully. Sep 13 01:04:45.680098 systemd[1]: session-25.scope: Deactivated successfully. Sep 13 01:04:45.680654 systemd-logind[1597]: Session 25 logged out. Waiting for processes to exit. Sep 13 01:04:45.682662 systemd-logind[1597]: Removed session 25. Sep 13 01:04:50.821951 systemd[1]: Started sshd@24-10.230.35.122:22-139.178.68.195:40128.service - OpenSSH per-connection server daemon (139.178.68.195:40128). Sep 13 01:04:51.717631 sshd[4468]: Accepted publickey for core from 139.178.68.195 port 40128 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:51.720220 sshd[4468]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:51.727874 systemd-logind[1597]: New session 26 of user core. Sep 13 01:04:51.733048 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 13 01:04:52.431014 sshd[4468]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:52.436229 systemd[1]: sshd@24-10.230.35.122:22-139.178.68.195:40128.service: Deactivated successfully. Sep 13 01:04:52.441967 systemd[1]: session-26.scope: Deactivated successfully. Sep 13 01:04:52.443731 systemd-logind[1597]: Session 26 logged out. Waiting for processes to exit. Sep 13 01:04:52.445141 systemd-logind[1597]: Removed session 26. Sep 13 01:04:52.586312 systemd[1]: Started sshd@25-10.230.35.122:22-139.178.68.195:40136.service - OpenSSH per-connection server daemon (139.178.68.195:40136). Sep 13 01:04:53.472451 sshd[4482]: Accepted publickey for core from 139.178.68.195 port 40136 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:53.475088 sshd[4482]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:53.485628 systemd-logind[1597]: New session 27 of user core. Sep 13 01:04:53.489165 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 13 01:04:55.765150 systemd[1]: run-containerd-runc-k8s.io-2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f-runc.4QNmAA.mount: Deactivated successfully. Sep 13 01:04:55.811690 containerd[1618]: time="2025-09-13T01:04:55.811102751Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 01:04:55.814539 containerd[1618]: time="2025-09-13T01:04:55.814332944Z" level=info msg="StopContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" with timeout 2 (s)" Sep 13 01:04:55.814905 containerd[1618]: time="2025-09-13T01:04:55.814693140Z" level=info msg="Stop container \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" with signal terminated" Sep 13 01:04:55.815582 containerd[1618]: time="2025-09-13T01:04:55.815308609Z" level=info msg="StopContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" with timeout 30 (s)" Sep 13 01:04:55.817741 containerd[1618]: time="2025-09-13T01:04:55.817711085Z" level=info msg="Stop container \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" with signal terminated" Sep 13 01:04:55.833688 systemd-networkd[1259]: lxc_health: Link DOWN Sep 13 01:04:55.833702 systemd-networkd[1259]: lxc_health: Lost carrier Sep 13 01:04:55.904971 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231-rootfs.mount: Deactivated successfully. Sep 13 01:04:55.912095 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f-rootfs.mount: Deactivated successfully. Sep 13 01:04:55.932445 containerd[1618]: time="2025-09-13T01:04:55.932033444Z" level=info msg="shim disconnected" id=2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f namespace=k8s.io Sep 13 01:04:55.932445 containerd[1618]: time="2025-09-13T01:04:55.932173253Z" level=warning msg="cleaning up after shim disconnected" id=2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f namespace=k8s.io Sep 13 01:04:55.932445 containerd[1618]: time="2025-09-13T01:04:55.932202410Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:04:55.933764 containerd[1618]: time="2025-09-13T01:04:55.932931690Z" level=info msg="shim disconnected" id=1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231 namespace=k8s.io Sep 13 01:04:55.933764 containerd[1618]: time="2025-09-13T01:04:55.933001313Z" level=warning msg="cleaning up after shim disconnected" id=1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231 namespace=k8s.io Sep 13 01:04:55.933764 containerd[1618]: time="2025-09-13T01:04:55.933019897Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:04:55.961229 containerd[1618]: time="2025-09-13T01:04:55.961175731Z" level=info msg="StopContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" returns successfully" Sep 13 01:04:55.962925 containerd[1618]: time="2025-09-13T01:04:55.962556270Z" level=info msg="StopContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" returns successfully" Sep 13 01:04:55.969889 containerd[1618]: time="2025-09-13T01:04:55.969839556Z" level=info msg="StopPodSandbox for \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\"" Sep 13 01:04:55.969989 containerd[1618]: time="2025-09-13T01:04:55.969911366Z" level=info msg="Container to stop \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.970167 containerd[1618]: time="2025-09-13T01:04:55.970137580Z" level=info msg="StopPodSandbox for \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\"" Sep 13 01:04:55.970240 containerd[1618]: time="2025-09-13T01:04:55.970178508Z" level=info msg="Container to stop \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.970240 containerd[1618]: time="2025-09-13T01:04:55.970199835Z" level=info msg="Container to stop \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.970240 containerd[1618]: time="2025-09-13T01:04:55.970217474Z" level=info msg="Container to stop \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.970240 containerd[1618]: time="2025-09-13T01:04:55.970233705Z" level=info msg="Container to stop \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.970440 containerd[1618]: time="2025-09-13T01:04:55.970249297Z" level=info msg="Container to stop \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 01:04:55.974391 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2-shm.mount: Deactivated successfully. Sep 13 01:04:55.976660 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede-shm.mount: Deactivated successfully. Sep 13 01:04:56.024337 containerd[1618]: time="2025-09-13T01:04:56.023606169Z" level=info msg="shim disconnected" id=b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede namespace=k8s.io Sep 13 01:04:56.024337 containerd[1618]: time="2025-09-13T01:04:56.024248607Z" level=warning msg="cleaning up after shim disconnected" id=b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede namespace=k8s.io Sep 13 01:04:56.024746 containerd[1618]: time="2025-09-13T01:04:56.024270168Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:04:56.041846 containerd[1618]: time="2025-09-13T01:04:56.041770344Z" level=info msg="shim disconnected" id=39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2 namespace=k8s.io Sep 13 01:04:56.041846 containerd[1618]: time="2025-09-13T01:04:56.041840879Z" level=warning msg="cleaning up after shim disconnected" id=39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2 namespace=k8s.io Sep 13 01:04:56.041846 containerd[1618]: time="2025-09-13T01:04:56.041856302Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:04:56.053468 containerd[1618]: time="2025-09-13T01:04:56.053408650Z" level=info msg="TearDown network for sandbox \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" successfully" Sep 13 01:04:56.053726 containerd[1618]: time="2025-09-13T01:04:56.053698518Z" level=info msg="StopPodSandbox for \"b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede\" returns successfully" Sep 13 01:04:56.091793 containerd[1618]: time="2025-09-13T01:04:56.091701157Z" level=warning msg="cleanup warnings time=\"2025-09-13T01:04:56Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 13 01:04:56.095233 containerd[1618]: time="2025-09-13T01:04:56.095006456Z" level=info msg="TearDown network for sandbox \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\" successfully" Sep 13 01:04:56.095233 containerd[1618]: time="2025-09-13T01:04:56.095075886Z" level=info msg="StopPodSandbox for \"39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2\" returns successfully" Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213717 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/48712337-4ca7-4808-8468-3fcbea712daa-cilium-config-path\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213792 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/48712337-4ca7-4808-8468-3fcbea712daa-clustermesh-secrets\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213830 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-kernel\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213871 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-bpf-maps\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213898 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cni-path\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.214447 kubelet[2876]: I0913 01:04:56.213938 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-etc-cni-netd\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.213969 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w497g\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-kube-api-access-w497g\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.214005 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-net\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.214048 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-lib-modules\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.214076 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-cgroup\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.214111 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7d6b\" (UniqueName: \"kubernetes.io/projected/cd73d6ce-5f91-43b9-8990-ced1697df22a-kube-api-access-k7d6b\") pod \"cd73d6ce-5f91-43b9-8990-ced1697df22a\" (UID: \"cd73d6ce-5f91-43b9-8990-ced1697df22a\") " Sep 13 01:04:56.215445 kubelet[2876]: I0913 01:04:56.214142 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd73d6ce-5f91-43b9-8990-ced1697df22a-cilium-config-path\") pod \"cd73d6ce-5f91-43b9-8990-ced1697df22a\" (UID: \"cd73d6ce-5f91-43b9-8990-ced1697df22a\") " Sep 13 01:04:56.215827 kubelet[2876]: I0913 01:04:56.214173 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-run\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215827 kubelet[2876]: I0913 01:04:56.214202 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-hubble-tls\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215827 kubelet[2876]: I0913 01:04:56.214240 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-xtables-lock\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.215827 kubelet[2876]: I0913 01:04:56.214264 2876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-hostproc\") pod \"48712337-4ca7-4808-8468-3fcbea712daa\" (UID: \"48712337-4ca7-4808-8468-3fcbea712daa\") " Sep 13 01:04:56.216772 kubelet[2876]: I0913 01:04:56.214355 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-hostproc" (OuterVolumeSpecName: "hostproc") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.225419 kubelet[2876]: I0913 01:04:56.225355 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48712337-4ca7-4808-8468-3fcbea712daa-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 01:04:56.225541 kubelet[2876]: I0913 01:04:56.225429 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.225541 kubelet[2876]: I0913 01:04:56.225464 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.225541 kubelet[2876]: I0913 01:04:56.225520 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.225856 kubelet[2876]: I0913 01:04:56.225551 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cni-path" (OuterVolumeSpecName: "cni-path") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.225856 kubelet[2876]: I0913 01:04:56.225578 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.228165 kubelet[2876]: I0913 01:04:56.228132 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48712337-4ca7-4808-8468-3fcbea712daa-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 01:04:56.228577 kubelet[2876]: I0913 01:04:56.228335 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.233728 kubelet[2876]: I0913 01:04:56.233672 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-kube-api-access-w497g" (OuterVolumeSpecName: "kube-api-access-w497g") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "kube-api-access-w497g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 01:04:56.233833 kubelet[2876]: I0913 01:04:56.233741 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.233833 kubelet[2876]: I0913 01:04:56.233780 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.234646 kubelet[2876]: I0913 01:04:56.234290 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd73d6ce-5f91-43b9-8990-ced1697df22a-kube-api-access-k7d6b" (OuterVolumeSpecName: "kube-api-access-k7d6b") pod "cd73d6ce-5f91-43b9-8990-ced1697df22a" (UID: "cd73d6ce-5f91-43b9-8990-ced1697df22a"). InnerVolumeSpecName "kube-api-access-k7d6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 01:04:56.237495 kubelet[2876]: I0913 01:04:56.237465 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 01:04:56.237677 kubelet[2876]: I0913 01:04:56.237563 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd73d6ce-5f91-43b9-8990-ced1697df22a-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "cd73d6ce-5f91-43b9-8990-ced1697df22a" (UID: "cd73d6ce-5f91-43b9-8990-ced1697df22a"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 01:04:56.237813 kubelet[2876]: I0913 01:04:56.237640 2876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "48712337-4ca7-4808-8468-3fcbea712daa" (UID: "48712337-4ca7-4808-8468-3fcbea712daa"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315303 2876 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-run\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315359 2876 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd73d6ce-5f91-43b9-8990-ced1697df22a-cilium-config-path\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315379 2876 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-hostproc\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315396 2876 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-hubble-tls\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315426 2876 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-xtables-lock\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315442 2876 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/48712337-4ca7-4808-8468-3fcbea712daa-cilium-config-path\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315459 2876 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/48712337-4ca7-4808-8468-3fcbea712daa-clustermesh-secrets\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.315807 kubelet[2876]: I0913 01:04:56.315487 2876 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-kernel\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315512 2876 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-bpf-maps\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315528 2876 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cni-path\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315564 2876 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-etc-cni-netd\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315578 2876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w497g\" (UniqueName: \"kubernetes.io/projected/48712337-4ca7-4808-8468-3fcbea712daa-kube-api-access-w497g\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315678 2876 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-host-proc-sys-net\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315700 2876 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-lib-modules\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315716 2876 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/48712337-4ca7-4808-8468-3fcbea712daa-cilium-cgroup\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.316319 kubelet[2876]: I0913 01:04:56.315733 2876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7d6b\" (UniqueName: \"kubernetes.io/projected/cd73d6ce-5f91-43b9-8990-ced1697df22a-kube-api-access-k7d6b\") on node \"srv-ngi99.gb1.brightbox.com\" DevicePath \"\"" Sep 13 01:04:56.465080 kubelet[2876]: E0913 01:04:56.463920 2876 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 01:04:56.711238 kubelet[2876]: I0913 01:04:56.711179 2876 scope.go:117] "RemoveContainer" containerID="2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f" Sep 13 01:04:56.725858 containerd[1618]: time="2025-09-13T01:04:56.725783078Z" level=info msg="RemoveContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\"" Sep 13 01:04:56.743035 containerd[1618]: time="2025-09-13T01:04:56.742978298Z" level=info msg="RemoveContainer for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" returns successfully" Sep 13 01:04:56.743408 kubelet[2876]: I0913 01:04:56.743366 2876 scope.go:117] "RemoveContainer" containerID="10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf" Sep 13 01:04:56.745168 containerd[1618]: time="2025-09-13T01:04:56.745136473Z" level=info msg="RemoveContainer for \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\"" Sep 13 01:04:56.756735 containerd[1618]: time="2025-09-13T01:04:56.751888202Z" level=info msg="RemoveContainer for \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\" returns successfully" Sep 13 01:04:56.761172 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-39588d2630cbe113c822eedf45553b6e28061927dd7360769fc3b12deec62df2-rootfs.mount: Deactivated successfully. Sep 13 01:04:56.763938 kubelet[2876]: I0913 01:04:56.762540 2876 scope.go:117] "RemoveContainer" containerID="1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a" Sep 13 01:04:56.762849 systemd[1]: var-lib-kubelet-pods-cd73d6ce\x2d5f91\x2d43b9\x2d8990\x2dced1697df22a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dk7d6b.mount: Deactivated successfully. Sep 13 01:04:56.763053 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b9a942a57ba67414995cdbe797a69886fe9abb0f53be7b2c39b4e46a29de9ede-rootfs.mount: Deactivated successfully. Sep 13 01:04:56.763237 systemd[1]: var-lib-kubelet-pods-48712337\x2d4ca7\x2d4808\x2d8468\x2d3fcbea712daa-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dw497g.mount: Deactivated successfully. Sep 13 01:04:56.763422 systemd[1]: var-lib-kubelet-pods-48712337\x2d4ca7\x2d4808\x2d8468\x2d3fcbea712daa-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 01:04:56.763992 systemd[1]: var-lib-kubelet-pods-48712337\x2d4ca7\x2d4808\x2d8468\x2d3fcbea712daa-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 01:04:56.768730 containerd[1618]: time="2025-09-13T01:04:56.768640795Z" level=info msg="RemoveContainer for \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\"" Sep 13 01:04:56.773594 containerd[1618]: time="2025-09-13T01:04:56.773559511Z" level=info msg="RemoveContainer for \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\" returns successfully" Sep 13 01:04:56.774275 kubelet[2876]: I0913 01:04:56.774063 2876 scope.go:117] "RemoveContainer" containerID="6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c" Sep 13 01:04:56.778355 containerd[1618]: time="2025-09-13T01:04:56.778273043Z" level=info msg="RemoveContainer for \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\"" Sep 13 01:04:56.781081 containerd[1618]: time="2025-09-13T01:04:56.781045887Z" level=info msg="RemoveContainer for \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\" returns successfully" Sep 13 01:04:56.781636 kubelet[2876]: I0913 01:04:56.781498 2876 scope.go:117] "RemoveContainer" containerID="529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c" Sep 13 01:04:56.783451 containerd[1618]: time="2025-09-13T01:04:56.783419447Z" level=info msg="RemoveContainer for \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\"" Sep 13 01:04:56.786497 containerd[1618]: time="2025-09-13T01:04:56.786461071Z" level=info msg="RemoveContainer for \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\" returns successfully" Sep 13 01:04:56.786711 kubelet[2876]: I0913 01:04:56.786671 2876 scope.go:117] "RemoveContainer" containerID="2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f" Sep 13 01:04:56.797609 containerd[1618]: time="2025-09-13T01:04:56.789315289Z" level=error msg="ContainerStatus for \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\": not found" Sep 13 01:04:56.803551 kubelet[2876]: E0913 01:04:56.803397 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\": not found" containerID="2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f" Sep 13 01:04:56.815020 kubelet[2876]: I0913 01:04:56.803470 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f"} err="failed to get container status \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\": rpc error: code = NotFound desc = an error occurred when try to find container \"2107197e8918c79364e6221f65cb40bfad44a5c752fab5808e6072746ab9620f\": not found" Sep 13 01:04:56.815020 kubelet[2876]: I0913 01:04:56.814838 2876 scope.go:117] "RemoveContainer" containerID="10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf" Sep 13 01:04:56.815536 kubelet[2876]: E0913 01:04:56.815443 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\": not found" containerID="10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf" Sep 13 01:04:56.815536 kubelet[2876]: I0913 01:04:56.815477 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf"} err="failed to get container status \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\": rpc error: code = NotFound desc = an error occurred when try to find container \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\": not found" Sep 13 01:04:56.815536 kubelet[2876]: I0913 01:04:56.815499 2876 scope.go:117] "RemoveContainer" containerID="1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a" Sep 13 01:04:56.815763 containerd[1618]: time="2025-09-13T01:04:56.815257434Z" level=error msg="ContainerStatus for \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"10cc07e72fb593443f3cc63b2687b02a414af4e57019f3a251cc2196045532cf\": not found" Sep 13 01:04:56.817359 containerd[1618]: time="2025-09-13T01:04:56.816515955Z" level=error msg="ContainerStatus for \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\": not found" Sep 13 01:04:56.817359 containerd[1618]: time="2025-09-13T01:04:56.817161309Z" level=error msg="ContainerStatus for \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\": not found" Sep 13 01:04:56.818407 kubelet[2876]: E0913 01:04:56.816889 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\": not found" containerID="1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a" Sep 13 01:04:56.818407 kubelet[2876]: I0913 01:04:56.816921 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a"} err="failed to get container status \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\": rpc error: code = NotFound desc = an error occurred when try to find container \"1029e7b3e731962bd45a33047e579fa3576aca54f58506f9f26eb2903c8ff87a\": not found" Sep 13 01:04:56.818407 kubelet[2876]: I0913 01:04:56.816945 2876 scope.go:117] "RemoveContainer" containerID="6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c" Sep 13 01:04:56.818407 kubelet[2876]: E0913 01:04:56.817310 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\": not found" containerID="6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c" Sep 13 01:04:56.818407 kubelet[2876]: I0913 01:04:56.817340 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c"} err="failed to get container status \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\": rpc error: code = NotFound desc = an error occurred when try to find container \"6393d98e72c8c2a12cc928321596e7f4d9f9b70810d8d50dd79f61a6f87e2b2c\": not found" Sep 13 01:04:56.818407 kubelet[2876]: I0913 01:04:56.817363 2876 scope.go:117] "RemoveContainer" containerID="529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c" Sep 13 01:04:56.818778 containerd[1618]: time="2025-09-13T01:04:56.817701798Z" level=error msg="ContainerStatus for \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\": not found" Sep 13 01:04:56.818846 kubelet[2876]: E0913 01:04:56.817854 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\": not found" containerID="529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c" Sep 13 01:04:56.818846 kubelet[2876]: I0913 01:04:56.817885 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c"} err="failed to get container status \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\": rpc error: code = NotFound desc = an error occurred when try to find container \"529c18898a415bc874bc60806cf12e6f8fbf4b821ca241bc5fdb7c60fccbb51c\": not found" Sep 13 01:04:56.818846 kubelet[2876]: I0913 01:04:56.817908 2876 scope.go:117] "RemoveContainer" containerID="1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231" Sep 13 01:04:56.820285 containerd[1618]: time="2025-09-13T01:04:56.819913443Z" level=info msg="RemoveContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\"" Sep 13 01:04:56.824102 containerd[1618]: time="2025-09-13T01:04:56.823927163Z" level=info msg="RemoveContainer for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" returns successfully" Sep 13 01:04:56.824213 kubelet[2876]: I0913 01:04:56.824112 2876 scope.go:117] "RemoveContainer" containerID="1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231" Sep 13 01:04:56.824554 containerd[1618]: time="2025-09-13T01:04:56.824438614Z" level=error msg="ContainerStatus for \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\": not found" Sep 13 01:04:56.824872 kubelet[2876]: E0913 01:04:56.824783 2876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\": not found" containerID="1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231" Sep 13 01:04:56.824872 kubelet[2876]: I0913 01:04:56.824817 2876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231"} err="failed to get container status \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\": rpc error: code = NotFound desc = an error occurred when try to find container \"1ff8173833cabc2c3b3b59561f3ecc95d59bf7922c0525bec1a587485f283231\": not found" Sep 13 01:04:57.652385 sshd[4482]: pam_unix(sshd:session): session closed for user core Sep 13 01:04:57.661780 systemd-logind[1597]: Session 27 logged out. Waiting for processes to exit. Sep 13 01:04:57.662978 systemd[1]: sshd@25-10.230.35.122:22-139.178.68.195:40136.service: Deactivated successfully. Sep 13 01:04:57.669883 systemd[1]: session-27.scope: Deactivated successfully. Sep 13 01:04:57.671984 systemd-logind[1597]: Removed session 27. Sep 13 01:04:57.811000 systemd[1]: Started sshd@26-10.230.35.122:22-139.178.68.195:40142.service - OpenSSH per-connection server daemon (139.178.68.195:40142). Sep 13 01:04:58.147555 kubelet[2876]: I0913 01:04:58.147439 2876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48712337-4ca7-4808-8468-3fcbea712daa" path="/var/lib/kubelet/pods/48712337-4ca7-4808-8468-3fcbea712daa/volumes" Sep 13 01:04:58.148854 kubelet[2876]: I0913 01:04:58.148818 2876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd73d6ce-5f91-43b9-8990-ced1697df22a" path="/var/lib/kubelet/pods/cd73d6ce-5f91-43b9-8990-ced1697df22a/volumes" Sep 13 01:04:58.718867 sshd[4648]: Accepted publickey for core from 139.178.68.195 port 40142 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:04:58.721372 sshd[4648]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:04:58.729710 systemd-logind[1597]: New session 28 of user core. Sep 13 01:04:58.735085 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 13 01:04:59.702235 kubelet[2876]: I0913 01:04:59.702131 2876 setters.go:600] "Node became not ready" node="srv-ngi99.gb1.brightbox.com" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-13T01:04:59Z","lastTransitionTime":"2025-09-13T01:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158476 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="apply-sysctl-overwrites" Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158524 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="mount-bpf-fs" Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158541 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="clean-cilium-state" Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158553 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="mount-cgroup" Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158563 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="cd73d6ce-5f91-43b9-8990-ced1697df22a" containerName="cilium-operator" Sep 13 01:05:00.160800 kubelet[2876]: E0913 01:05:00.158572 2876 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="cilium-agent" Sep 13 01:05:00.160800 kubelet[2876]: I0913 01:05:00.158666 2876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd73d6ce-5f91-43b9-8990-ced1697df22a" containerName="cilium-operator" Sep 13 01:05:00.160800 kubelet[2876]: I0913 01:05:00.158698 2876 memory_manager.go:354] "RemoveStaleState removing state" podUID="48712337-4ca7-4808-8468-3fcbea712daa" containerName="cilium-agent" Sep 13 01:05:00.248352 kubelet[2876]: I0913 01:05:00.248287 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-cilium-cgroup\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.250580 kubelet[2876]: I0913 01:05:00.250459 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-lib-modules\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.250580 kubelet[2876]: I0913 01:05:00.250523 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-xtables-lock\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250560 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-bpf-maps\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250829 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-hostproc\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250871 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-host-proc-sys-kernel\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250900 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/06129fd6-597d-4bb0-bb82-929474dfa53b-hubble-tls\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250929 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-etc-cni-netd\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251243 kubelet[2876]: I0913 01:05:00.250959 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/06129fd6-597d-4bb0-bb82-929474dfa53b-cilium-ipsec-secrets\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251002 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/06129fd6-597d-4bb0-bb82-929474dfa53b-clustermesh-secrets\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251034 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-host-proc-sys-net\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251075 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-cni-path\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251120 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7dck\" (UniqueName: \"kubernetes.io/projected/06129fd6-597d-4bb0-bb82-929474dfa53b-kube-api-access-g7dck\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251165 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/06129fd6-597d-4bb0-bb82-929474dfa53b-cilium-run\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.251553 kubelet[2876]: I0913 01:05:00.251195 2876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/06129fd6-597d-4bb0-bb82-929474dfa53b-cilium-config-path\") pod \"cilium-s7wp5\" (UID: \"06129fd6-597d-4bb0-bb82-929474dfa53b\") " pod="kube-system/cilium-s7wp5" Sep 13 01:05:00.259988 sshd[4648]: pam_unix(sshd:session): session closed for user core Sep 13 01:05:00.272559 systemd-logind[1597]: Session 28 logged out. Waiting for processes to exit. Sep 13 01:05:00.273322 systemd[1]: sshd@26-10.230.35.122:22-139.178.68.195:40142.service: Deactivated successfully. Sep 13 01:05:00.277283 systemd[1]: session-28.scope: Deactivated successfully. Sep 13 01:05:00.281275 systemd-logind[1597]: Removed session 28. Sep 13 01:05:00.413986 systemd[1]: Started sshd@27-10.230.35.122:22-139.178.68.195:56560.service - OpenSSH per-connection server daemon (139.178.68.195:56560). Sep 13 01:05:00.504172 containerd[1618]: time="2025-09-13T01:05:00.504120184Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s7wp5,Uid:06129fd6-597d-4bb0-bb82-929474dfa53b,Namespace:kube-system,Attempt:0,}" Sep 13 01:05:00.536203 containerd[1618]: time="2025-09-13T01:05:00.536083065Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 01:05:00.536995 containerd[1618]: time="2025-09-13T01:05:00.536651580Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 01:05:00.536995 containerd[1618]: time="2025-09-13T01:05:00.536683698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:05:00.536995 containerd[1618]: time="2025-09-13T01:05:00.536871658Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 01:05:00.606992 containerd[1618]: time="2025-09-13T01:05:00.606936066Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s7wp5,Uid:06129fd6-597d-4bb0-bb82-929474dfa53b,Namespace:kube-system,Attempt:0,} returns sandbox id \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\"" Sep 13 01:05:00.612531 containerd[1618]: time="2025-09-13T01:05:00.612451539Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 01:05:00.623519 containerd[1618]: time="2025-09-13T01:05:00.623443449Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3929a41590cc2b6cadbfb3a12ebe0ab6a8c073d8e86e91ba01efa2dbe5f11a3e\"" Sep 13 01:05:00.624792 containerd[1618]: time="2025-09-13T01:05:00.624107240Z" level=info msg="StartContainer for \"3929a41590cc2b6cadbfb3a12ebe0ab6a8c073d8e86e91ba01efa2dbe5f11a3e\"" Sep 13 01:05:00.696847 containerd[1618]: time="2025-09-13T01:05:00.696378335Z" level=info msg="StartContainer for \"3929a41590cc2b6cadbfb3a12ebe0ab6a8c073d8e86e91ba01efa2dbe5f11a3e\" returns successfully" Sep 13 01:05:00.755324 containerd[1618]: time="2025-09-13T01:05:00.755045051Z" level=info msg="shim disconnected" id=3929a41590cc2b6cadbfb3a12ebe0ab6a8c073d8e86e91ba01efa2dbe5f11a3e namespace=k8s.io Sep 13 01:05:00.755324 containerd[1618]: time="2025-09-13T01:05:00.755168831Z" level=warning msg="cleaning up after shim disconnected" id=3929a41590cc2b6cadbfb3a12ebe0ab6a8c073d8e86e91ba01efa2dbe5f11a3e namespace=k8s.io Sep 13 01:05:00.755324 containerd[1618]: time="2025-09-13T01:05:00.755209590Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:05:01.300486 sshd[4666]: Accepted publickey for core from 139.178.68.195 port 56560 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:05:01.302669 sshd[4666]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:05:01.311674 systemd-logind[1597]: New session 29 of user core. Sep 13 01:05:01.318240 systemd[1]: Started session-29.scope - Session 29 of User core. Sep 13 01:05:01.466674 kubelet[2876]: E0913 01:05:01.466470 2876 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 01:05:01.762156 containerd[1618]: time="2025-09-13T01:05:01.762089300Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 01:05:01.797627 containerd[1618]: time="2025-09-13T01:05:01.797538818Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74\"" Sep 13 01:05:01.800211 containerd[1618]: time="2025-09-13T01:05:01.800161629Z" level=info msg="StartContainer for \"32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74\"" Sep 13 01:05:01.888477 containerd[1618]: time="2025-09-13T01:05:01.888401874Z" level=info msg="StartContainer for \"32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74\" returns successfully" Sep 13 01:05:01.921035 sshd[4666]: pam_unix(sshd:session): session closed for user core Sep 13 01:05:01.926421 containerd[1618]: time="2025-09-13T01:05:01.926304181Z" level=info msg="shim disconnected" id=32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74 namespace=k8s.io Sep 13 01:05:01.926421 containerd[1618]: time="2025-09-13T01:05:01.926385414Z" level=warning msg="cleaning up after shim disconnected" id=32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74 namespace=k8s.io Sep 13 01:05:01.926421 containerd[1618]: time="2025-09-13T01:05:01.926401867Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:05:01.932125 systemd[1]: sshd@27-10.230.35.122:22-139.178.68.195:56560.service: Deactivated successfully. Sep 13 01:05:01.938689 systemd[1]: session-29.scope: Deactivated successfully. Sep 13 01:05:01.942292 systemd-logind[1597]: Session 29 logged out. Waiting for processes to exit. Sep 13 01:05:01.948180 systemd-logind[1597]: Removed session 29. Sep 13 01:05:02.074961 systemd[1]: Started sshd@28-10.230.35.122:22-139.178.68.195:56562.service - OpenSSH per-connection server daemon (139.178.68.195:56562). Sep 13 01:05:02.359523 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-32efb000548e40877de0a59131d10c180253919ab3ba6461a508b26cd91b2b74-rootfs.mount: Deactivated successfully. Sep 13 01:05:02.765287 containerd[1618]: time="2025-09-13T01:05:02.765222615Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 01:05:02.802082 containerd[1618]: time="2025-09-13T01:05:02.802024618Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590\"" Sep 13 01:05:02.802263 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3220786624.mount: Deactivated successfully. Sep 13 01:05:02.805503 containerd[1618]: time="2025-09-13T01:05:02.805168555Z" level=info msg="StartContainer for \"0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590\"" Sep 13 01:05:02.905835 containerd[1618]: time="2025-09-13T01:05:02.905780132Z" level=info msg="StartContainer for \"0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590\" returns successfully" Sep 13 01:05:02.971664 sshd[4844]: Accepted publickey for core from 139.178.68.195 port 56562 ssh2: RSA SHA256:nCFR9BVD/sBsaMzu6piX/nSqoN/UcYzTi/UCsy9A7bQ Sep 13 01:05:02.973229 containerd[1618]: time="2025-09-13T01:05:02.973144060Z" level=info msg="shim disconnected" id=0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590 namespace=k8s.io Sep 13 01:05:02.973347 containerd[1618]: time="2025-09-13T01:05:02.973232337Z" level=warning msg="cleaning up after shim disconnected" id=0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590 namespace=k8s.io Sep 13 01:05:02.973347 containerd[1618]: time="2025-09-13T01:05:02.973250277Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:05:02.973984 sshd[4844]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 13 01:05:02.986175 systemd-logind[1597]: New session 30 of user core. Sep 13 01:05:02.991106 systemd[1]: Started session-30.scope - Session 30 of User core. Sep 13 01:05:03.361881 systemd[1]: run-containerd-runc-k8s.io-0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590-runc.jWwx5P.mount: Deactivated successfully. Sep 13 01:05:03.363707 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0f637c3be5496d30f9b31ee22052cb57538a1c46e411778140dd514808ea3590-rootfs.mount: Deactivated successfully. Sep 13 01:05:03.770866 containerd[1618]: time="2025-09-13T01:05:03.770663827Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 01:05:03.794337 containerd[1618]: time="2025-09-13T01:05:03.794267862Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e\"" Sep 13 01:05:03.795268 containerd[1618]: time="2025-09-13T01:05:03.795236720Z" level=info msg="StartContainer for \"ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e\"" Sep 13 01:05:03.884887 containerd[1618]: time="2025-09-13T01:05:03.884835081Z" level=info msg="StartContainer for \"ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e\" returns successfully" Sep 13 01:05:03.911207 containerd[1618]: time="2025-09-13T01:05:03.911132709Z" level=info msg="shim disconnected" id=ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e namespace=k8s.io Sep 13 01:05:03.911207 containerd[1618]: time="2025-09-13T01:05:03.911202665Z" level=warning msg="cleaning up after shim disconnected" id=ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e namespace=k8s.io Sep 13 01:05:03.911738 containerd[1618]: time="2025-09-13T01:05:03.911223174Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 13 01:05:03.929657 containerd[1618]: time="2025-09-13T01:05:03.929511037Z" level=warning msg="cleanup warnings time=\"2025-09-13T01:05:03Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 13 01:05:04.362854 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ff3229fac4c75bd0c7e5f26effc3e46597822650369159d5ad50b8930b54bd9e-rootfs.mount: Deactivated successfully. Sep 13 01:05:04.783860 containerd[1618]: time="2025-09-13T01:05:04.783730445Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 01:05:04.812189 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2571612432.mount: Deactivated successfully. Sep 13 01:05:04.814232 containerd[1618]: time="2025-09-13T01:05:04.814187823Z" level=info msg="CreateContainer within sandbox \"7eb38835b0b9f7d421054876681fe9f418e7f14f65ae04735dd1b93fe896d09d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"cd1dde3f6f5a4470aba0486beccd16db26fd71eed0092e2ec09dae5101cd1f7a\"" Sep 13 01:05:04.818762 containerd[1618]: time="2025-09-13T01:05:04.816305740Z" level=info msg="StartContainer for \"cd1dde3f6f5a4470aba0486beccd16db26fd71eed0092e2ec09dae5101cd1f7a\"" Sep 13 01:05:04.914581 containerd[1618]: time="2025-09-13T01:05:04.914531389Z" level=info msg="StartContainer for \"cd1dde3f6f5a4470aba0486beccd16db26fd71eed0092e2ec09dae5101cd1f7a\" returns successfully" Sep 13 01:05:05.668669 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 13 01:05:05.826942 kubelet[2876]: I0913 01:05:05.826759 2876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-s7wp5" podStartSLOduration=5.826656766 podStartE2EDuration="5.826656766s" podCreationTimestamp="2025-09-13 01:05:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 01:05:05.823145004 +0000 UTC m=+159.893400052" watchObservedRunningTime="2025-09-13 01:05:05.826656766 +0000 UTC m=+159.896911806" Sep 13 01:05:08.121358 systemd[1]: run-containerd-runc-k8s.io-cd1dde3f6f5a4470aba0486beccd16db26fd71eed0092e2ec09dae5101cd1f7a-runc.eKrUfu.mount: Deactivated successfully. Sep 13 01:05:09.578818 systemd-networkd[1259]: lxc_health: Link UP Sep 13 01:05:09.590810 systemd-networkd[1259]: lxc_health: Gained carrier Sep 13 01:05:11.154116 systemd-networkd[1259]: lxc_health: Gained IPv6LL Sep 13 01:05:15.141957 kubelet[2876]: E0913 01:05:15.141822 2876 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:48876->127.0.0.1:35063: write tcp 127.0.0.1:48876->127.0.0.1:35063: write: broken pipe Sep 13 01:05:15.307379 sshd[4844]: pam_unix(sshd:session): session closed for user core Sep 13 01:05:15.320410 systemd[1]: sshd@28-10.230.35.122:22-139.178.68.195:56562.service: Deactivated successfully. Sep 13 01:05:15.331744 systemd[1]: session-30.scope: Deactivated successfully. Sep 13 01:05:15.337772 systemd-logind[1597]: Session 30 logged out. Waiting for processes to exit. Sep 13 01:05:15.342916 systemd-logind[1597]: Removed session 30.