Sep 12 10:10:59.936060 kernel: Linux version 6.6.105-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 08:42:12 -00 2025 Sep 12 10:10:59.936087 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=87e444606a7368354f582e8f746f078f97e75cf74b35edd9ec39d0d73a54ead2 Sep 12 10:10:59.936100 kernel: BIOS-provided physical RAM map: Sep 12 10:10:59.936107 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 12 10:10:59.936114 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 12 10:10:59.936120 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 12 10:10:59.936128 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Sep 12 10:10:59.936135 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Sep 12 10:10:59.936142 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 12 10:10:59.936149 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 12 10:10:59.936159 kernel: NX (Execute Disable) protection: active Sep 12 10:10:59.936165 kernel: APIC: Static calls initialized Sep 12 10:10:59.936176 kernel: SMBIOS 2.8 present. Sep 12 10:10:59.936184 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Sep 12 10:10:59.936192 kernel: Hypervisor detected: KVM Sep 12 10:10:59.936200 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 12 10:10:59.936214 kernel: kvm-clock: using sched offset of 2889771744 cycles Sep 12 10:10:59.936223 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 12 10:10:59.936230 kernel: tsc: Detected 2494.138 MHz processor Sep 12 10:10:59.936239 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 12 10:10:59.936247 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 12 10:10:59.936255 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Sep 12 10:10:59.936263 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 12 10:10:59.936271 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 12 10:10:59.936282 kernel: ACPI: Early table checksum verification disabled Sep 12 10:10:59.936289 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Sep 12 10:10:59.936297 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936305 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936313 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936321 kernel: ACPI: FACS 0x000000007FFE0000 000040 Sep 12 10:10:59.936328 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936336 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936344 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936354 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 12 10:10:59.936362 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Sep 12 10:10:59.936370 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Sep 12 10:10:59.936377 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Sep 12 10:10:59.936385 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Sep 12 10:10:59.936393 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Sep 12 10:10:59.936401 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Sep 12 10:10:59.936413 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Sep 12 10:10:59.936424 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 12 10:10:59.936432 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 12 10:10:59.936440 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Sep 12 10:10:59.936448 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Sep 12 10:10:59.936459 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Sep 12 10:10:59.936467 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Sep 12 10:10:59.936479 kernel: Zone ranges: Sep 12 10:10:59.936487 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 12 10:10:59.936495 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Sep 12 10:10:59.936503 kernel: Normal empty Sep 12 10:10:59.936511 kernel: Movable zone start for each node Sep 12 10:10:59.936520 kernel: Early memory node ranges Sep 12 10:10:59.936548 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 12 10:10:59.936557 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Sep 12 10:10:59.936565 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Sep 12 10:10:59.936576 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 12 10:10:59.936584 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 12 10:10:59.936595 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Sep 12 10:10:59.936603 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 12 10:10:59.936612 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 12 10:10:59.936620 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 12 10:10:59.936628 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 12 10:10:59.936636 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 12 10:10:59.936645 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 12 10:10:59.936653 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 12 10:10:59.936664 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 12 10:10:59.936672 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 12 10:10:59.936680 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 12 10:10:59.936688 kernel: TSC deadline timer available Sep 12 10:10:59.936697 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 12 10:10:59.936705 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 12 10:10:59.936713 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Sep 12 10:10:59.936724 kernel: Booting paravirtualized kernel on KVM Sep 12 10:10:59.936733 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 12 10:10:59.936744 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 12 10:10:59.937130 kernel: percpu: Embedded 58 pages/cpu s197160 r8192 d32216 u1048576 Sep 12 10:10:59.937141 kernel: pcpu-alloc: s197160 r8192 d32216 u1048576 alloc=1*2097152 Sep 12 10:10:59.937150 kernel: pcpu-alloc: [0] 0 1 Sep 12 10:10:59.937158 kernel: kvm-guest: PV spinlocks disabled, no host support Sep 12 10:10:59.937168 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=87e444606a7368354f582e8f746f078f97e75cf74b35edd9ec39d0d73a54ead2 Sep 12 10:10:59.937177 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 12 10:10:59.937185 kernel: random: crng init done Sep 12 10:10:59.937197 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 12 10:10:59.937206 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 12 10:10:59.937214 kernel: Fallback order for Node 0: 0 Sep 12 10:10:59.937223 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Sep 12 10:10:59.937231 kernel: Policy zone: DMA32 Sep 12 10:10:59.937239 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 12 10:10:59.937248 kernel: Memory: 1969156K/2096612K available (14336K kernel code, 2293K rwdata, 22868K rodata, 43508K init, 1568K bss, 127196K reserved, 0K cma-reserved) Sep 12 10:10:59.937257 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 12 10:10:59.937265 kernel: Kernel/User page tables isolation: enabled Sep 12 10:10:59.937276 kernel: ftrace: allocating 37946 entries in 149 pages Sep 12 10:10:59.937284 kernel: ftrace: allocated 149 pages with 4 groups Sep 12 10:10:59.937293 kernel: Dynamic Preempt: voluntary Sep 12 10:10:59.937301 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 12 10:10:59.937310 kernel: rcu: RCU event tracing is enabled. Sep 12 10:10:59.937319 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 12 10:10:59.937327 kernel: Trampoline variant of Tasks RCU enabled. Sep 12 10:10:59.937335 kernel: Rude variant of Tasks RCU enabled. Sep 12 10:10:59.937344 kernel: Tracing variant of Tasks RCU enabled. Sep 12 10:10:59.937355 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 12 10:10:59.937363 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 12 10:10:59.937372 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 12 10:10:59.937380 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 12 10:10:59.937391 kernel: Console: colour VGA+ 80x25 Sep 12 10:10:59.937399 kernel: printk: console [tty0] enabled Sep 12 10:10:59.937408 kernel: printk: console [ttyS0] enabled Sep 12 10:10:59.937416 kernel: ACPI: Core revision 20230628 Sep 12 10:10:59.937425 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 12 10:10:59.937436 kernel: APIC: Switch to symmetric I/O mode setup Sep 12 10:10:59.937444 kernel: x2apic enabled Sep 12 10:10:59.937453 kernel: APIC: Switched APIC routing to: physical x2apic Sep 12 10:10:59.937461 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 12 10:10:59.937470 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39838d43, max_idle_ns: 440795267131 ns Sep 12 10:10:59.937478 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494138) Sep 12 10:10:59.937486 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Sep 12 10:10:59.937495 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Sep 12 10:10:59.937515 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 12 10:10:59.937524 kernel: Spectre V2 : Mitigation: Retpolines Sep 12 10:10:59.937547 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 12 10:10:59.937556 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Sep 12 10:10:59.937588 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 12 10:10:59.937597 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Sep 12 10:10:59.937607 kernel: MDS: Mitigation: Clear CPU buffers Sep 12 10:10:59.937619 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 12 10:10:59.937633 kernel: active return thunk: its_return_thunk Sep 12 10:10:59.937650 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 12 10:10:59.937659 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 12 10:10:59.937668 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 12 10:10:59.937681 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 12 10:10:59.937694 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 12 10:10:59.937706 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Sep 12 10:10:59.937719 kernel: Freeing SMP alternatives memory: 32K Sep 12 10:10:59.937732 kernel: pid_max: default: 32768 minimum: 301 Sep 12 10:10:59.937748 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 12 10:10:59.937762 kernel: landlock: Up and running. Sep 12 10:10:59.937775 kernel: SELinux: Initializing. Sep 12 10:10:59.937787 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 10:10:59.937801 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 10:10:59.937813 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Sep 12 10:10:59.937826 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 10:10:59.937841 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 10:10:59.937850 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 10:10:59.937862 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Sep 12 10:10:59.937871 kernel: signal: max sigframe size: 1776 Sep 12 10:10:59.937880 kernel: rcu: Hierarchical SRCU implementation. Sep 12 10:10:59.937889 kernel: rcu: Max phase no-delay instances is 400. Sep 12 10:10:59.937898 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 12 10:10:59.937907 kernel: smp: Bringing up secondary CPUs ... Sep 12 10:10:59.937916 kernel: smpboot: x86: Booting SMP configuration: Sep 12 10:10:59.937925 kernel: .... node #0, CPUs: #1 Sep 12 10:10:59.937938 kernel: smp: Brought up 1 node, 2 CPUs Sep 12 10:10:59.937950 kernel: smpboot: Max logical packages: 1 Sep 12 10:10:59.937959 kernel: smpboot: Total of 2 processors activated (9976.55 BogoMIPS) Sep 12 10:10:59.937968 kernel: devtmpfs: initialized Sep 12 10:10:59.937976 kernel: x86/mm: Memory block size: 128MB Sep 12 10:10:59.937986 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 12 10:10:59.937994 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 12 10:10:59.938003 kernel: pinctrl core: initialized pinctrl subsystem Sep 12 10:10:59.938012 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 12 10:10:59.938021 kernel: audit: initializing netlink subsys (disabled) Sep 12 10:10:59.938033 kernel: audit: type=2000 audit(1757671858.995:1): state=initialized audit_enabled=0 res=1 Sep 12 10:10:59.938042 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 12 10:10:59.938050 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 12 10:10:59.938059 kernel: cpuidle: using governor menu Sep 12 10:10:59.938068 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 12 10:10:59.938077 kernel: dca service started, version 1.12.1 Sep 12 10:10:59.938086 kernel: PCI: Using configuration type 1 for base access Sep 12 10:10:59.938095 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 12 10:10:59.938103 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 12 10:10:59.938115 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 12 10:10:59.938124 kernel: ACPI: Added _OSI(Module Device) Sep 12 10:10:59.938139 kernel: ACPI: Added _OSI(Processor Device) Sep 12 10:10:59.938152 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 12 10:10:59.938164 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 12 10:10:59.938176 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 12 10:10:59.938188 kernel: ACPI: Interpreter enabled Sep 12 10:10:59.938199 kernel: ACPI: PM: (supports S0 S5) Sep 12 10:10:59.938208 kernel: ACPI: Using IOAPIC for interrupt routing Sep 12 10:10:59.938220 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 12 10:10:59.938229 kernel: PCI: Using E820 reservations for host bridge windows Sep 12 10:10:59.938238 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Sep 12 10:10:59.938247 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 12 10:10:59.938465 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 12 10:10:59.939124 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Sep 12 10:10:59.939242 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Sep 12 10:10:59.939260 kernel: acpiphp: Slot [3] registered Sep 12 10:10:59.939270 kernel: acpiphp: Slot [4] registered Sep 12 10:10:59.939279 kernel: acpiphp: Slot [5] registered Sep 12 10:10:59.939288 kernel: acpiphp: Slot [6] registered Sep 12 10:10:59.939297 kernel: acpiphp: Slot [7] registered Sep 12 10:10:59.939305 kernel: acpiphp: Slot [8] registered Sep 12 10:10:59.939314 kernel: acpiphp: Slot [9] registered Sep 12 10:10:59.939323 kernel: acpiphp: Slot [10] registered Sep 12 10:10:59.939332 kernel: acpiphp: Slot [11] registered Sep 12 10:10:59.939341 kernel: acpiphp: Slot [12] registered Sep 12 10:10:59.939353 kernel: acpiphp: Slot [13] registered Sep 12 10:10:59.939362 kernel: acpiphp: Slot [14] registered Sep 12 10:10:59.939371 kernel: acpiphp: Slot [15] registered Sep 12 10:10:59.939379 kernel: acpiphp: Slot [16] registered Sep 12 10:10:59.939388 kernel: acpiphp: Slot [17] registered Sep 12 10:10:59.939397 kernel: acpiphp: Slot [18] registered Sep 12 10:10:59.939406 kernel: acpiphp: Slot [19] registered Sep 12 10:10:59.939414 kernel: acpiphp: Slot [20] registered Sep 12 10:10:59.939423 kernel: acpiphp: Slot [21] registered Sep 12 10:10:59.939435 kernel: acpiphp: Slot [22] registered Sep 12 10:10:59.939444 kernel: acpiphp: Slot [23] registered Sep 12 10:10:59.939453 kernel: acpiphp: Slot [24] registered Sep 12 10:10:59.939462 kernel: acpiphp: Slot [25] registered Sep 12 10:10:59.939470 kernel: acpiphp: Slot [26] registered Sep 12 10:10:59.939479 kernel: acpiphp: Slot [27] registered Sep 12 10:10:59.939488 kernel: acpiphp: Slot [28] registered Sep 12 10:10:59.939497 kernel: acpiphp: Slot [29] registered Sep 12 10:10:59.939505 kernel: acpiphp: Slot [30] registered Sep 12 10:10:59.939514 kernel: acpiphp: Slot [31] registered Sep 12 10:10:59.939525 kernel: PCI host bridge to bus 0000:00 Sep 12 10:10:59.939688 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 12 10:10:59.939790 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 12 10:10:59.939901 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 12 10:10:59.939989 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 12 10:10:59.940075 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Sep 12 10:10:59.940165 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 12 10:10:59.940306 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Sep 12 10:10:59.940417 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Sep 12 10:10:59.940618 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Sep 12 10:10:59.940725 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Sep 12 10:10:59.940827 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Sep 12 10:10:59.940926 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Sep 12 10:10:59.941030 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Sep 12 10:10:59.941133 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Sep 12 10:10:59.941248 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Sep 12 10:10:59.941351 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Sep 12 10:10:59.941470 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Sep 12 10:10:59.941604 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Sep 12 10:10:59.941711 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Sep 12 10:10:59.941828 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Sep 12 10:10:59.942622 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Sep 12 10:10:59.942763 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Sep 12 10:10:59.942877 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Sep 12 10:10:59.942978 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Sep 12 10:10:59.943115 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 12 10:10:59.943247 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Sep 12 10:10:59.943411 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Sep 12 10:10:59.945677 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Sep 12 10:10:59.945812 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Sep 12 10:10:59.945927 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 12 10:10:59.946052 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Sep 12 10:10:59.946153 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Sep 12 10:10:59.946259 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Sep 12 10:10:59.946373 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Sep 12 10:10:59.946473 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Sep 12 10:10:59.946584 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Sep 12 10:10:59.946682 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Sep 12 10:10:59.946785 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Sep 12 10:10:59.946883 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Sep 12 10:10:59.947006 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Sep 12 10:10:59.947114 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Sep 12 10:10:59.947241 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Sep 12 10:10:59.947341 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Sep 12 10:10:59.947439 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Sep 12 10:10:59.950723 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Sep 12 10:10:59.950979 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Sep 12 10:10:59.951120 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Sep 12 10:10:59.951222 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Sep 12 10:10:59.951235 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 12 10:10:59.951244 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 12 10:10:59.951254 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 12 10:10:59.951263 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 12 10:10:59.951272 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 12 10:10:59.951285 kernel: iommu: Default domain type: Translated Sep 12 10:10:59.951294 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 12 10:10:59.951302 kernel: PCI: Using ACPI for IRQ routing Sep 12 10:10:59.951311 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 12 10:10:59.951320 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 12 10:10:59.951329 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Sep 12 10:10:59.951429 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Sep 12 10:10:59.951526 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Sep 12 10:10:59.951723 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 12 10:10:59.951745 kernel: vgaarb: loaded Sep 12 10:10:59.951754 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 12 10:10:59.951764 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 12 10:10:59.951773 kernel: clocksource: Switched to clocksource kvm-clock Sep 12 10:10:59.951782 kernel: VFS: Disk quotas dquot_6.6.0 Sep 12 10:10:59.951792 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 12 10:10:59.951800 kernel: pnp: PnP ACPI init Sep 12 10:10:59.951809 kernel: pnp: PnP ACPI: found 4 devices Sep 12 10:10:59.951822 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 12 10:10:59.951831 kernel: NET: Registered PF_INET protocol family Sep 12 10:10:59.951840 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 12 10:10:59.951849 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 12 10:10:59.951858 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 12 10:10:59.951867 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 12 10:10:59.951876 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 12 10:10:59.951885 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 12 10:10:59.951894 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 10:10:59.951905 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 10:10:59.951914 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 12 10:10:59.951923 kernel: NET: Registered PF_XDP protocol family Sep 12 10:10:59.952033 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 12 10:10:59.952122 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 12 10:10:59.952210 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 12 10:10:59.952296 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 12 10:10:59.952383 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Sep 12 10:10:59.952486 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Sep 12 10:10:59.956738 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 12 10:10:59.956777 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 12 10:10:59.956889 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7a0 took 30559 usecs Sep 12 10:10:59.956904 kernel: PCI: CLS 0 bytes, default 64 Sep 12 10:10:59.956914 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 12 10:10:59.956923 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39838d43, max_idle_ns: 440795267131 ns Sep 12 10:10:59.956933 kernel: Initialise system trusted keyrings Sep 12 10:10:59.956942 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 12 10:10:59.956958 kernel: Key type asymmetric registered Sep 12 10:10:59.956967 kernel: Asymmetric key parser 'x509' registered Sep 12 10:10:59.956977 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 12 10:10:59.956986 kernel: io scheduler mq-deadline registered Sep 12 10:10:59.956994 kernel: io scheduler kyber registered Sep 12 10:10:59.957003 kernel: io scheduler bfq registered Sep 12 10:10:59.957012 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 12 10:10:59.957028 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Sep 12 10:10:59.957041 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Sep 12 10:10:59.957059 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Sep 12 10:10:59.957068 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 12 10:10:59.957077 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 12 10:10:59.957087 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 12 10:10:59.957096 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 12 10:10:59.957105 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 12 10:10:59.957115 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 12 10:10:59.957249 kernel: rtc_cmos 00:03: RTC can wake from S4 Sep 12 10:10:59.957354 kernel: rtc_cmos 00:03: registered as rtc0 Sep 12 10:10:59.957446 kernel: rtc_cmos 00:03: setting system clock to 2025-09-12T10:10:59 UTC (1757671859) Sep 12 10:10:59.957616 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Sep 12 10:10:59.957628 kernel: intel_pstate: CPU model not supported Sep 12 10:10:59.957637 kernel: NET: Registered PF_INET6 protocol family Sep 12 10:10:59.957646 kernel: Segment Routing with IPv6 Sep 12 10:10:59.957655 kernel: In-situ OAM (IOAM) with IPv6 Sep 12 10:10:59.957668 kernel: NET: Registered PF_PACKET protocol family Sep 12 10:10:59.957683 kernel: Key type dns_resolver registered Sep 12 10:10:59.957697 kernel: IPI shorthand broadcast: enabled Sep 12 10:10:59.957706 kernel: sched_clock: Marking stable (808002800, 82575812)->(978998496, -88419884) Sep 12 10:10:59.957716 kernel: registered taskstats version 1 Sep 12 10:10:59.957724 kernel: Loading compiled-in X.509 certificates Sep 12 10:10:59.957734 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.105-flatcar: 0972efc09ee0bcd53f8cdb5573e11871ce7b16a9' Sep 12 10:10:59.957742 kernel: Key type .fscrypt registered Sep 12 10:10:59.957751 kernel: Key type fscrypt-provisioning registered Sep 12 10:10:59.957760 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 12 10:10:59.957772 kernel: ima: Allocated hash algorithm: sha1 Sep 12 10:10:59.957780 kernel: ima: No architecture policies found Sep 12 10:10:59.957789 kernel: clk: Disabling unused clocks Sep 12 10:10:59.957798 kernel: Freeing unused kernel image (initmem) memory: 43508K Sep 12 10:10:59.957808 kernel: Write protecting the kernel read-only data: 38912k Sep 12 10:10:59.957834 kernel: Freeing unused kernel image (rodata/data gap) memory: 1708K Sep 12 10:10:59.957846 kernel: Run /init as init process Sep 12 10:10:59.957856 kernel: with arguments: Sep 12 10:10:59.957865 kernel: /init Sep 12 10:10:59.957877 kernel: with environment: Sep 12 10:10:59.957886 kernel: HOME=/ Sep 12 10:10:59.957896 kernel: TERM=linux Sep 12 10:10:59.957905 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 12 10:10:59.957916 systemd[1]: Successfully made /usr/ read-only. Sep 12 10:10:59.957929 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 10:10:59.957940 systemd[1]: Detected virtualization kvm. Sep 12 10:10:59.957950 systemd[1]: Detected architecture x86-64. Sep 12 10:10:59.957962 systemd[1]: Running in initrd. Sep 12 10:10:59.957979 systemd[1]: No hostname configured, using default hostname. Sep 12 10:10:59.957993 systemd[1]: Hostname set to . Sep 12 10:10:59.958008 systemd[1]: Initializing machine ID from VM UUID. Sep 12 10:10:59.958018 systemd[1]: Queued start job for default target initrd.target. Sep 12 10:10:59.958028 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 10:10:59.958038 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 10:10:59.958049 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 12 10:10:59.958066 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 10:10:59.958076 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 12 10:10:59.958087 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 12 10:10:59.958098 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 12 10:10:59.958108 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 12 10:10:59.958119 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 10:10:59.958129 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 10:10:59.958143 systemd[1]: Reached target paths.target - Path Units. Sep 12 10:10:59.958154 systemd[1]: Reached target slices.target - Slice Units. Sep 12 10:10:59.958167 systemd[1]: Reached target swap.target - Swaps. Sep 12 10:10:59.958177 systemd[1]: Reached target timers.target - Timer Units. Sep 12 10:10:59.958187 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 10:10:59.958201 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 10:10:59.958211 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 12 10:10:59.958227 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 12 10:10:59.958242 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 10:10:59.958256 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 10:10:59.958270 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 10:10:59.958285 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 10:10:59.958295 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 12 10:10:59.958305 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 10:10:59.958320 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 12 10:10:59.958330 systemd[1]: Starting systemd-fsck-usr.service... Sep 12 10:10:59.958340 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 10:10:59.958350 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 10:10:59.958360 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:10:59.958370 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 12 10:10:59.958428 systemd-journald[184]: Collecting audit messages is disabled. Sep 12 10:10:59.958456 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 10:10:59.958468 systemd[1]: Finished systemd-fsck-usr.service. Sep 12 10:10:59.958481 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 10:10:59.958492 systemd-journald[184]: Journal started Sep 12 10:10:59.958517 systemd-journald[184]: Runtime Journal (/run/log/journal/659b0fce28794b2a94f53d716f568289) is 4.9M, max 39.3M, 34.4M free. Sep 12 10:10:59.958050 systemd-modules-load[185]: Inserted module 'overlay' Sep 12 10:10:59.962557 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 10:10:59.988119 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 10:11:00.001983 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 12 10:11:00.002015 kernel: Bridge firewalling registered Sep 12 10:10:59.998436 systemd-modules-load[185]: Inserted module 'br_netfilter' Sep 12 10:11:00.011468 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 10:11:00.013275 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:00.014641 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 10:11:00.024919 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 10:11:00.027653 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 10:11:00.034728 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 10:11:00.037134 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 10:11:00.051013 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 10:11:00.052374 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 10:11:00.059991 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 12 10:11:00.062720 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 10:11:00.063927 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 10:11:00.082193 dracut-cmdline[218]: dracut-dracut-053 Sep 12 10:11:00.093052 dracut-cmdline[218]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=87e444606a7368354f582e8f746f078f97e75cf74b35edd9ec39d0d73a54ead2 Sep 12 10:11:00.115739 systemd-resolved[219]: Positive Trust Anchors: Sep 12 10:11:00.115756 systemd-resolved[219]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 10:11:00.115794 systemd-resolved[219]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 10:11:00.119607 systemd-resolved[219]: Defaulting to hostname 'linux'. Sep 12 10:11:00.121135 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 10:11:00.121725 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 10:11:00.206599 kernel: SCSI subsystem initialized Sep 12 10:11:00.218568 kernel: Loading iSCSI transport class v2.0-870. Sep 12 10:11:00.230586 kernel: iscsi: registered transport (tcp) Sep 12 10:11:00.253588 kernel: iscsi: registered transport (qla4xxx) Sep 12 10:11:00.253724 kernel: QLogic iSCSI HBA Driver Sep 12 10:11:00.306745 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 12 10:11:00.312887 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 12 10:11:00.341839 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 12 10:11:00.341956 kernel: device-mapper: uevent: version 1.0.3 Sep 12 10:11:00.343311 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 12 10:11:00.387614 kernel: raid6: avx2x4 gen() 21726 MB/s Sep 12 10:11:00.404598 kernel: raid6: avx2x2 gen() 25441 MB/s Sep 12 10:11:00.421693 kernel: raid6: avx2x1 gen() 22647 MB/s Sep 12 10:11:00.421807 kernel: raid6: using algorithm avx2x2 gen() 25441 MB/s Sep 12 10:11:00.439943 kernel: raid6: .... xor() 18977 MB/s, rmw enabled Sep 12 10:11:00.440071 kernel: raid6: using avx2x2 recovery algorithm Sep 12 10:11:00.462572 kernel: xor: automatically using best checksumming function avx Sep 12 10:11:00.622589 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 12 10:11:00.638454 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 12 10:11:00.649866 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 10:11:00.667253 systemd-udevd[403]: Using default interface naming scheme 'v255'. Sep 12 10:11:00.673702 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 10:11:00.682776 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 12 10:11:00.699784 dracut-pre-trigger[408]: rd.md=0: removing MD RAID activation Sep 12 10:11:00.742930 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 10:11:00.747859 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 10:11:00.832590 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 10:11:00.842120 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 12 10:11:00.865887 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 12 10:11:00.867506 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 10:11:00.868123 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 10:11:00.868433 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 10:11:00.875834 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 12 10:11:00.891215 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 12 10:11:00.957116 kernel: scsi host0: Virtio SCSI HBA Sep 12 10:11:00.957611 kernel: cryptd: max_cpu_qlen set to 1000 Sep 12 10:11:00.959556 kernel: libata version 3.00 loaded. Sep 12 10:11:00.963603 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Sep 12 10:11:00.966558 kernel: ata_piix 0000:00:01.1: version 2.13 Sep 12 10:11:00.973730 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Sep 12 10:11:00.981720 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 12 10:11:00.981828 kernel: GPT:9289727 != 125829119 Sep 12 10:11:00.982763 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 12 10:11:00.982822 kernel: GPT:9289727 != 125829119 Sep 12 10:11:00.983967 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 12 10:11:00.984022 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 10:11:01.008565 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Sep 12 10:11:01.010996 kernel: AVX2 version of gcm_enc/dec engaged. Sep 12 10:11:01.011081 kernel: AES CTR mode by8 optimization enabled Sep 12 10:11:01.011554 kernel: scsi host1: ata_piix Sep 12 10:11:01.011913 kernel: virtio_blk virtio5: [vdb] 976 512-byte logical blocks (500 kB/488 KiB) Sep 12 10:11:01.016227 kernel: scsi host2: ata_piix Sep 12 10:11:01.016618 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Sep 12 10:11:01.016647 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Sep 12 10:11:01.019573 kernel: ACPI: bus type USB registered Sep 12 10:11:01.023564 kernel: usbcore: registered new interface driver usbfs Sep 12 10:11:01.027542 kernel: usbcore: registered new interface driver hub Sep 12 10:11:01.027695 kernel: usbcore: registered new device driver usb Sep 12 10:11:01.042100 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 12 10:11:01.042208 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 10:11:01.043411 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 10:11:01.044468 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 10:11:01.044587 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:01.045226 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:11:01.050804 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:11:01.052983 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 10:11:01.099218 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:01.104876 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 10:11:01.128254 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 10:11:01.213897 kernel: BTRFS: device fsid 2566299d-dd4a-4826-ba43-7397a17991fb devid 1 transid 35 /dev/vda3 scanned by (udev-worker) (448) Sep 12 10:11:01.217563 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by (udev-worker) (455) Sep 12 10:11:01.230700 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 12 10:11:01.252601 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Sep 12 10:11:01.256118 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Sep 12 10:11:01.256373 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Sep 12 10:11:01.256080 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 12 10:11:01.259554 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Sep 12 10:11:01.261816 kernel: hub 1-0:1.0: USB hub found Sep 12 10:11:01.262075 kernel: hub 1-0:1.0: 2 ports detected Sep 12 10:11:01.265348 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 12 10:11:01.266523 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 12 10:11:01.276065 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 10:11:01.284798 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 12 10:11:01.291037 disk-uuid[552]: Primary Header is updated. Sep 12 10:11:01.291037 disk-uuid[552]: Secondary Entries is updated. Sep 12 10:11:01.291037 disk-uuid[552]: Secondary Header is updated. Sep 12 10:11:01.307696 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 10:11:02.319614 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 12 10:11:02.321829 disk-uuid[553]: The operation has completed successfully. Sep 12 10:11:02.387560 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 12 10:11:02.387767 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 12 10:11:02.419838 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 12 10:11:02.431984 sh[564]: Success Sep 12 10:11:02.452576 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 12 10:11:02.519261 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 12 10:11:02.521272 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 12 10:11:02.522010 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 12 10:11:02.552066 kernel: BTRFS info (device dm-0): first mount of filesystem 2566299d-dd4a-4826-ba43-7397a17991fb Sep 12 10:11:02.552145 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 12 10:11:02.552160 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 12 10:11:02.553868 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 12 10:11:02.555551 kernel: BTRFS info (device dm-0): using free space tree Sep 12 10:11:02.564838 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 12 10:11:02.566031 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 12 10:11:02.572826 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 12 10:11:02.576813 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 12 10:11:02.594592 kernel: BTRFS info (device vda6): first mount of filesystem 36a15e30-b48e-4687-be9c-f68c3ae1825b Sep 12 10:11:02.594666 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 10:11:02.595774 kernel: BTRFS info (device vda6): using free space tree Sep 12 10:11:02.600558 kernel: BTRFS info (device vda6): auto enabling async discard Sep 12 10:11:02.606640 kernel: BTRFS info (device vda6): last unmount of filesystem 36a15e30-b48e-4687-be9c-f68c3ae1825b Sep 12 10:11:02.608132 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 12 10:11:02.615852 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 12 10:11:02.729505 ignition[638]: Ignition 2.20.0 Sep 12 10:11:02.730090 ignition[638]: Stage: fetch-offline Sep 12 10:11:02.730137 ignition[638]: no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:02.730146 ignition[638]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:02.731457 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 10:11:02.730265 ignition[638]: parsed url from cmdline: "" Sep 12 10:11:02.730268 ignition[638]: no config URL provided Sep 12 10:11:02.730274 ignition[638]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 10:11:02.730282 ignition[638]: no config at "/usr/lib/ignition/user.ign" Sep 12 10:11:02.730289 ignition[638]: failed to fetch config: resource requires networking Sep 12 10:11:02.730469 ignition[638]: Ignition finished successfully Sep 12 10:11:02.751581 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 10:11:02.758818 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 10:11:02.803050 systemd-networkd[750]: lo: Link UP Sep 12 10:11:02.803062 systemd-networkd[750]: lo: Gained carrier Sep 12 10:11:02.805854 systemd-networkd[750]: Enumeration completed Sep 12 10:11:02.806239 systemd-networkd[750]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Sep 12 10:11:02.806243 systemd-networkd[750]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Sep 12 10:11:02.806966 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 10:11:02.807034 systemd-networkd[750]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 10:11:02.807039 systemd-networkd[750]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 10:11:02.807704 systemd[1]: Reached target network.target - Network. Sep 12 10:11:02.807947 systemd-networkd[750]: eth0: Link UP Sep 12 10:11:02.807954 systemd-networkd[750]: eth0: Gained carrier Sep 12 10:11:02.807965 systemd-networkd[750]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Sep 12 10:11:02.813885 systemd-networkd[750]: eth1: Link UP Sep 12 10:11:02.813893 systemd-networkd[750]: eth1: Gained carrier Sep 12 10:11:02.813909 systemd-networkd[750]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 10:11:02.815894 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 12 10:11:02.824614 systemd-networkd[750]: eth0: DHCPv4 address 164.92.125.86/19, gateway 164.92.96.1 acquired from 169.254.169.253 Sep 12 10:11:02.829671 systemd-networkd[750]: eth1: DHCPv4 address 10.124.0.31/20 acquired from 169.254.169.253 Sep 12 10:11:02.838327 ignition[753]: Ignition 2.20.0 Sep 12 10:11:02.838968 ignition[753]: Stage: fetch Sep 12 10:11:02.839156 ignition[753]: no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:02.839167 ignition[753]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:02.839267 ignition[753]: parsed url from cmdline: "" Sep 12 10:11:02.839271 ignition[753]: no config URL provided Sep 12 10:11:02.839276 ignition[753]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 10:11:02.839286 ignition[753]: no config at "/usr/lib/ignition/user.ign" Sep 12 10:11:02.839309 ignition[753]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Sep 12 10:11:02.861487 ignition[753]: GET result: OK Sep 12 10:11:02.861647 ignition[753]: parsing config with SHA512: c4ba1848fdc49f52c3a8be3c742faec8505176c778a65856f203b7c9a740f609c6bfe52047039c0cfc350785e3c28a5def0b068cc124adba380aebd777165158 Sep 12 10:11:02.867898 unknown[753]: fetched base config from "system" Sep 12 10:11:02.867915 unknown[753]: fetched base config from "system" Sep 12 10:11:02.867925 unknown[753]: fetched user config from "digitalocean" Sep 12 10:11:02.868597 ignition[753]: fetch: fetch complete Sep 12 10:11:02.868604 ignition[753]: fetch: fetch passed Sep 12 10:11:02.868664 ignition[753]: Ignition finished successfully Sep 12 10:11:02.870439 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 12 10:11:02.875948 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 12 10:11:02.898233 ignition[761]: Ignition 2.20.0 Sep 12 10:11:02.898244 ignition[761]: Stage: kargs Sep 12 10:11:02.898455 ignition[761]: no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:02.898465 ignition[761]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:02.899351 ignition[761]: kargs: kargs passed Sep 12 10:11:02.899409 ignition[761]: Ignition finished successfully Sep 12 10:11:02.900583 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 12 10:11:02.907838 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 12 10:11:02.923029 ignition[767]: Ignition 2.20.0 Sep 12 10:11:02.923044 ignition[767]: Stage: disks Sep 12 10:11:02.924031 ignition[767]: no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:02.924045 ignition[767]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:02.924947 ignition[767]: disks: disks passed Sep 12 10:11:02.928802 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 12 10:11:02.925000 ignition[767]: Ignition finished successfully Sep 12 10:11:02.930121 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 12 10:11:02.931031 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 12 10:11:02.931813 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 10:11:02.932677 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 10:11:02.933246 systemd[1]: Reached target basic.target - Basic System. Sep 12 10:11:02.938806 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 12 10:11:02.960564 systemd-fsck[775]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 12 10:11:02.963574 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 12 10:11:02.971819 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 12 10:11:03.122563 kernel: EXT4-fs (vda9): mounted filesystem 4caafea7-bbab-4a47-b77b-37af606fc08b r/w with ordered data mode. Quota mode: none. Sep 12 10:11:03.123751 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 12 10:11:03.125353 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 12 10:11:03.141727 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 10:11:03.144964 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 12 10:11:03.147292 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... Sep 12 10:11:03.154837 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 12 10:11:03.159276 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by mount (783) Sep 12 10:11:03.158246 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 12 10:11:03.167149 kernel: BTRFS info (device vda6): first mount of filesystem 36a15e30-b48e-4687-be9c-f68c3ae1825b Sep 12 10:11:03.167179 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 10:11:03.167192 kernel: BTRFS info (device vda6): using free space tree Sep 12 10:11:03.158299 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 10:11:03.168994 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 12 10:11:03.185688 kernel: BTRFS info (device vda6): auto enabling async discard Sep 12 10:11:03.188068 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 12 10:11:03.194689 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 10:11:03.265952 coreos-metadata[785]: Sep 12 10:11:03.265 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 10:11:03.270135 initrd-setup-root[814]: cut: /sysroot/etc/passwd: No such file or directory Sep 12 10:11:03.271400 coreos-metadata[786]: Sep 12 10:11:03.271 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 10:11:03.277197 initrd-setup-root[821]: cut: /sysroot/etc/group: No such file or directory Sep 12 10:11:03.280107 coreos-metadata[785]: Sep 12 10:11:03.277 INFO Fetch successful Sep 12 10:11:03.282345 coreos-metadata[786]: Sep 12 10:11:03.280 INFO Fetch successful Sep 12 10:11:03.286259 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. Sep 12 10:11:03.286413 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. Sep 12 10:11:03.291144 coreos-metadata[786]: Sep 12 10:11:03.288 INFO wrote hostname ci-4230.2.2-n-dc4800d201 to /sysroot/etc/hostname Sep 12 10:11:03.289989 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 12 10:11:03.293379 initrd-setup-root[828]: cut: /sysroot/etc/shadow: No such file or directory Sep 12 10:11:03.298480 initrd-setup-root[837]: cut: /sysroot/etc/gshadow: No such file or directory Sep 12 10:11:03.424969 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 12 10:11:03.428738 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 12 10:11:03.431765 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 12 10:11:03.447561 kernel: BTRFS info (device vda6): last unmount of filesystem 36a15e30-b48e-4687-be9c-f68c3ae1825b Sep 12 10:11:03.471266 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 12 10:11:03.478468 ignition[906]: INFO : Ignition 2.20.0 Sep 12 10:11:03.480515 ignition[906]: INFO : Stage: mount Sep 12 10:11:03.480515 ignition[906]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:03.480515 ignition[906]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:03.480515 ignition[906]: INFO : mount: mount passed Sep 12 10:11:03.480515 ignition[906]: INFO : Ignition finished successfully Sep 12 10:11:03.483959 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 12 10:11:03.488806 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 12 10:11:03.550795 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 12 10:11:03.556895 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 10:11:03.570215 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (916) Sep 12 10:11:03.570313 kernel: BTRFS info (device vda6): first mount of filesystem 36a15e30-b48e-4687-be9c-f68c3ae1825b Sep 12 10:11:03.570338 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 12 10:11:03.571845 kernel: BTRFS info (device vda6): using free space tree Sep 12 10:11:03.580580 kernel: BTRFS info (device vda6): auto enabling async discard Sep 12 10:11:03.583093 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 10:11:03.610832 ignition[932]: INFO : Ignition 2.20.0 Sep 12 10:11:03.610832 ignition[932]: INFO : Stage: files Sep 12 10:11:03.611959 ignition[932]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:03.611959 ignition[932]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:03.612859 ignition[932]: DEBUG : files: compiled without relabeling support, skipping Sep 12 10:11:03.613618 ignition[932]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 12 10:11:03.613618 ignition[932]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 12 10:11:03.617101 ignition[932]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 12 10:11:03.617660 ignition[932]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 12 10:11:03.618277 ignition[932]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 12 10:11:03.618266 unknown[932]: wrote ssh authorized keys file for user: core Sep 12 10:11:03.619496 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 12 10:11:03.620245 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Sep 12 10:11:03.656298 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 12 10:11:03.761756 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Sep 12 10:11:03.761756 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 12 10:11:03.763339 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 12 10:11:03.963856 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 12 10:11:04.023603 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 12 10:11:04.023603 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 10:11:04.024930 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 12 10:11:04.029456 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Sep 12 10:11:04.358401 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 12 10:11:04.469750 systemd-networkd[750]: eth1: Gained IPv6LL Sep 12 10:11:04.646395 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Sep 12 10:11:04.646395 ignition[932]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Sep 12 10:11:04.648256 ignition[932]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 10:11:04.648256 ignition[932]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 10:11:04.648256 ignition[932]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Sep 12 10:11:04.648256 ignition[932]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Sep 12 10:11:04.650455 ignition[932]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Sep 12 10:11:04.650455 ignition[932]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 12 10:11:04.650455 ignition[932]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 12 10:11:04.650455 ignition[932]: INFO : files: files passed Sep 12 10:11:04.650455 ignition[932]: INFO : Ignition finished successfully Sep 12 10:11:04.651619 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 12 10:11:04.660953 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 12 10:11:04.664907 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 12 10:11:04.667692 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 12 10:11:04.668421 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 12 10:11:04.680869 initrd-setup-root-after-ignition[961]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 10:11:04.680869 initrd-setup-root-after-ignition[961]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 12 10:11:04.683899 initrd-setup-root-after-ignition[965]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 10:11:04.685914 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 10:11:04.687092 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 12 10:11:04.691887 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 12 10:11:04.725763 systemd-networkd[750]: eth0: Gained IPv6LL Sep 12 10:11:04.741973 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 12 10:11:04.742158 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 12 10:11:04.743375 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 12 10:11:04.744002 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 12 10:11:04.744848 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 12 10:11:04.749820 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 12 10:11:04.775292 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 10:11:04.781850 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 12 10:11:04.794391 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 12 10:11:04.796062 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 10:11:04.796737 systemd[1]: Stopped target timers.target - Timer Units. Sep 12 10:11:04.797651 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 12 10:11:04.797825 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 10:11:04.798888 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 12 10:11:04.799382 systemd[1]: Stopped target basic.target - Basic System. Sep 12 10:11:04.800117 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 12 10:11:04.800853 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 10:11:04.801565 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 12 10:11:04.802223 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 12 10:11:04.802997 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 10:11:04.803890 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 12 10:11:04.804634 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 12 10:11:04.805256 systemd[1]: Stopped target swap.target - Swaps. Sep 12 10:11:04.805819 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 12 10:11:04.805976 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 12 10:11:04.807015 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 12 10:11:04.807844 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 10:11:04.808279 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 12 10:11:04.809045 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 10:11:04.810187 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 12 10:11:04.810409 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 12 10:11:04.811958 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 12 10:11:04.812158 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 10:11:04.813437 systemd[1]: ignition-files.service: Deactivated successfully. Sep 12 10:11:04.813672 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 12 10:11:04.814436 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 12 10:11:04.814577 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 12 10:11:04.823000 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 12 10:11:04.827032 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 12 10:11:04.827569 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 12 10:11:04.827830 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 10:11:04.828739 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 12 10:11:04.828856 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 10:11:04.838431 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 12 10:11:04.839134 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 12 10:11:04.853875 ignition[987]: INFO : Ignition 2.20.0 Sep 12 10:11:04.856452 ignition[987]: INFO : Stage: umount Sep 12 10:11:04.856452 ignition[987]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 10:11:04.856452 ignition[987]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 12 10:11:04.856452 ignition[987]: INFO : umount: umount passed Sep 12 10:11:04.856452 ignition[987]: INFO : Ignition finished successfully Sep 12 10:11:04.864480 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 12 10:11:04.865089 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 12 10:11:04.868666 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 12 10:11:04.869687 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 12 10:11:04.869763 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 12 10:11:04.870242 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 12 10:11:04.870291 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 12 10:11:04.871076 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 12 10:11:04.871136 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 12 10:11:04.872211 systemd[1]: Stopped target network.target - Network. Sep 12 10:11:04.872940 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 12 10:11:04.873001 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 10:11:04.873832 systemd[1]: Stopped target paths.target - Path Units. Sep 12 10:11:04.874588 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 12 10:11:04.878629 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 10:11:04.879754 systemd[1]: Stopped target slices.target - Slice Units. Sep 12 10:11:04.880122 systemd[1]: Stopped target sockets.target - Socket Units. Sep 12 10:11:04.881103 systemd[1]: iscsid.socket: Deactivated successfully. Sep 12 10:11:04.881178 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 10:11:04.881856 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 12 10:11:04.881915 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 10:11:04.882599 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 12 10:11:04.882676 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 12 10:11:04.883419 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 12 10:11:04.883497 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 12 10:11:04.884685 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 12 10:11:04.885444 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 12 10:11:04.887197 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 12 10:11:04.887392 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 12 10:11:04.890445 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 12 10:11:04.890623 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 12 10:11:04.892583 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 12 10:11:04.892810 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 12 10:11:04.900028 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 12 10:11:04.900472 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 12 10:11:04.900690 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 12 10:11:04.903001 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 12 10:11:04.905213 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 12 10:11:04.905289 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 12 10:11:04.910778 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 12 10:11:04.911323 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 12 10:11:04.911442 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 10:11:04.912215 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 12 10:11:04.912302 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 12 10:11:04.915128 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 12 10:11:04.915218 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 12 10:11:04.916008 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 12 10:11:04.916079 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 10:11:04.917218 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 10:11:04.922546 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 12 10:11:04.922938 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 12 10:11:04.933987 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 12 10:11:04.934235 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 10:11:04.936318 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 12 10:11:04.936412 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 12 10:11:04.937124 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 12 10:11:04.937180 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 10:11:04.937982 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 12 10:11:04.938057 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 12 10:11:04.939332 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 12 10:11:04.939406 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 12 10:11:04.940715 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 12 10:11:04.940792 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 10:11:04.949832 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 12 10:11:04.950374 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 12 10:11:04.950562 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 10:11:04.952747 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 12 10:11:04.952825 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 10:11:04.953395 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 12 10:11:04.953454 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 10:11:04.953939 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 10:11:04.953995 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:04.958672 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 12 10:11:04.958771 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 10:11:04.959312 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 12 10:11:04.960746 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 12 10:11:04.961882 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 12 10:11:04.962035 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 12 10:11:04.963992 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 12 10:11:04.975966 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 12 10:11:04.986489 systemd[1]: Switching root. Sep 12 10:11:05.020821 systemd-journald[184]: Journal stopped Sep 12 10:11:06.410811 systemd-journald[184]: Received SIGTERM from PID 1 (systemd). Sep 12 10:11:06.410915 kernel: SELinux: policy capability network_peer_controls=1 Sep 12 10:11:06.410941 kernel: SELinux: policy capability open_perms=1 Sep 12 10:11:06.410968 kernel: SELinux: policy capability extended_socket_class=1 Sep 12 10:11:06.410987 kernel: SELinux: policy capability always_check_network=0 Sep 12 10:11:06.411012 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 12 10:11:06.411040 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 12 10:11:06.411059 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 12 10:11:06.411079 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 12 10:11:06.411099 kernel: audit: type=1403 audit(1757671865.192:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 12 10:11:06.411121 systemd[1]: Successfully loaded SELinux policy in 44.826ms. Sep 12 10:11:06.411159 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 16.174ms. Sep 12 10:11:06.411182 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 10:11:06.411203 systemd[1]: Detected virtualization kvm. Sep 12 10:11:06.411228 systemd[1]: Detected architecture x86-64. Sep 12 10:11:06.411250 systemd[1]: Detected first boot. Sep 12 10:11:06.411271 systemd[1]: Hostname set to . Sep 12 10:11:06.411313 systemd[1]: Initializing machine ID from VM UUID. Sep 12 10:11:06.411334 zram_generator::config[1039]: No configuration found. Sep 12 10:11:06.411355 kernel: Guest personality initialized and is inactive Sep 12 10:11:06.411373 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 12 10:11:06.411395 kernel: Initialized host personality Sep 12 10:11:06.411411 kernel: NET: Registered PF_VSOCK protocol family Sep 12 10:11:06.411427 systemd[1]: Populated /etc with preset unit settings. Sep 12 10:11:06.411447 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 12 10:11:06.411474 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 12 10:11:06.411492 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 12 10:11:06.411512 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 12 10:11:06.419383 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 12 10:11:06.419438 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 12 10:11:06.419474 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 12 10:11:06.419494 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 12 10:11:06.419516 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 12 10:11:06.421961 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 12 10:11:06.422005 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 12 10:11:06.422029 systemd[1]: Created slice user.slice - User and Session Slice. Sep 12 10:11:06.422050 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 10:11:06.422071 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 10:11:06.422092 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 12 10:11:06.422126 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 12 10:11:06.422149 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 12 10:11:06.422171 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 10:11:06.422192 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 12 10:11:06.422215 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 10:11:06.422235 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 12 10:11:06.422259 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 12 10:11:06.422310 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 12 10:11:06.422329 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 12 10:11:06.422348 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 10:11:06.422376 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 10:11:06.422396 systemd[1]: Reached target slices.target - Slice Units. Sep 12 10:11:06.422416 systemd[1]: Reached target swap.target - Swaps. Sep 12 10:11:06.422438 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 12 10:11:06.422458 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 12 10:11:06.422483 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 12 10:11:06.422503 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 10:11:06.422524 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 10:11:06.422561 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 10:11:06.422582 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 12 10:11:06.422602 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 12 10:11:06.422622 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 12 10:11:06.422642 systemd[1]: Mounting media.mount - External Media Directory... Sep 12 10:11:06.422663 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:06.422687 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 12 10:11:06.422708 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 12 10:11:06.422728 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 12 10:11:06.422750 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 12 10:11:06.422782 systemd[1]: Reached target machines.target - Containers. Sep 12 10:11:06.422804 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 12 10:11:06.422825 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 10:11:06.422846 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 10:11:06.422863 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 12 10:11:06.422887 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 10:11:06.422905 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 10:11:06.422923 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 10:11:06.422945 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 12 10:11:06.422965 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 10:11:06.422986 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 12 10:11:06.423007 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 12 10:11:06.423029 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 12 10:11:06.423056 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 12 10:11:06.423076 systemd[1]: Stopped systemd-fsck-usr.service. Sep 12 10:11:06.423098 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 10:11:06.423120 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 10:11:06.423141 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 10:11:06.423163 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 10:11:06.423185 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 12 10:11:06.423207 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 12 10:11:06.423245 kernel: fuse: init (API version 7.39) Sep 12 10:11:06.423271 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 10:11:06.423290 systemd[1]: verity-setup.service: Deactivated successfully. Sep 12 10:11:06.423312 systemd[1]: Stopped verity-setup.service. Sep 12 10:11:06.423340 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:06.423362 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 12 10:11:06.423384 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 12 10:11:06.423404 systemd[1]: Mounted media.mount - External Media Directory. Sep 12 10:11:06.423423 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 12 10:11:06.423441 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 12 10:11:06.423465 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 12 10:11:06.423486 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 10:11:06.423508 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 12 10:11:06.424942 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 12 10:11:06.424994 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 10:11:06.425017 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 10:11:06.425040 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 10:11:06.425060 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 10:11:06.425081 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 12 10:11:06.425109 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 12 10:11:06.425129 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 10:11:06.425152 kernel: loop: module loaded Sep 12 10:11:06.425193 kernel: ACPI: bus type drm_connector registered Sep 12 10:11:06.425212 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 12 10:11:06.425232 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 10:11:06.425252 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 10:11:06.425271 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 10:11:06.425291 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 10:11:06.425317 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 12 10:11:06.425338 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 12 10:11:06.425362 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 12 10:11:06.425388 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 10:11:06.425408 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 12 10:11:06.425476 systemd-journald[1106]: Collecting audit messages is disabled. Sep 12 10:11:06.425519 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 12 10:11:06.425561 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 12 10:11:06.425587 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 10:11:06.425608 systemd-journald[1106]: Journal started Sep 12 10:11:06.425647 systemd-journald[1106]: Runtime Journal (/run/log/journal/659b0fce28794b2a94f53d716f568289) is 4.9M, max 39.3M, 34.4M free. Sep 12 10:11:06.431121 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 12 10:11:06.431216 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 10:11:05.999858 systemd[1]: Queued start job for default target multi-user.target. Sep 12 10:11:06.014298 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 12 10:11:06.015066 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 12 10:11:06.441669 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 12 10:11:06.441791 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 10:11:06.456380 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 10:11:06.471615 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 12 10:11:06.476447 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 10:11:06.479582 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 10:11:06.483642 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 10:11:06.484841 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 12 10:11:06.485884 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 12 10:11:06.488648 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 12 10:11:06.509188 kernel: loop0: detected capacity change from 0 to 224512 Sep 12 10:11:06.514585 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 12 10:11:06.532080 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 12 10:11:06.549183 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 12 10:11:06.552079 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 10:11:06.555876 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 12 10:11:06.563480 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 12 10:11:06.570935 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 12 10:11:06.577464 systemd-tmpfiles[1137]: ACLs are not supported, ignoring. Sep 12 10:11:06.577484 systemd-tmpfiles[1137]: ACLs are not supported, ignoring. Sep 12 10:11:06.587260 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 12 10:11:06.597343 systemd-journald[1106]: Time spent on flushing to /var/log/journal/659b0fce28794b2a94f53d716f568289 is 98.314ms for 1012 entries. Sep 12 10:11:06.597343 systemd-journald[1106]: System Journal (/var/log/journal/659b0fce28794b2a94f53d716f568289) is 8M, max 195.6M, 187.6M free. Sep 12 10:11:06.710777 systemd-journald[1106]: Received client request to flush runtime journal. Sep 12 10:11:06.710870 kernel: loop1: detected capacity change from 0 to 8 Sep 12 10:11:06.710890 kernel: loop2: detected capacity change from 0 to 147912 Sep 12 10:11:06.603787 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 10:11:06.622809 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 12 10:11:06.623555 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 10:11:06.649759 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 12 10:11:06.685582 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 10:11:06.692829 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 12 10:11:06.715732 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 12 10:11:06.734908 udevadm[1176]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 12 10:11:06.737428 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 12 10:11:06.740789 kernel: loop3: detected capacity change from 0 to 138176 Sep 12 10:11:06.760815 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 10:11:06.806488 kernel: loop4: detected capacity change from 0 to 224512 Sep 12 10:11:06.839981 kernel: loop5: detected capacity change from 0 to 8 Sep 12 10:11:06.850996 kernel: loop6: detected capacity change from 0 to 147912 Sep 12 10:11:06.878211 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Sep 12 10:11:06.878248 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Sep 12 10:11:06.886915 kernel: loop7: detected capacity change from 0 to 138176 Sep 12 10:11:06.903217 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 10:11:06.915047 (sd-merge)[1185]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Sep 12 10:11:06.915749 (sd-merge)[1185]: Merged extensions into '/usr'. Sep 12 10:11:06.929204 systemd[1]: Reload requested from client PID 1136 ('systemd-sysext') (unit systemd-sysext.service)... Sep 12 10:11:06.929239 systemd[1]: Reloading... Sep 12 10:11:07.118575 zram_generator::config[1214]: No configuration found. Sep 12 10:11:07.300274 ldconfig[1128]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 12 10:11:07.395449 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 12 10:11:07.489886 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 12 10:11:07.490409 systemd[1]: Reloading finished in 559 ms. Sep 12 10:11:07.508640 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 12 10:11:07.509478 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 12 10:11:07.526847 systemd[1]: Starting ensure-sysext.service... Sep 12 10:11:07.530051 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 10:11:07.546317 systemd[1]: Reload requested from client PID 1257 ('systemctl') (unit ensure-sysext.service)... Sep 12 10:11:07.546335 systemd[1]: Reloading... Sep 12 10:11:07.591226 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 12 10:11:07.591677 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 12 10:11:07.597038 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 12 10:11:07.601808 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Sep 12 10:11:07.601889 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Sep 12 10:11:07.613910 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 10:11:07.614594 systemd-tmpfiles[1258]: Skipping /boot Sep 12 10:11:07.671026 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 10:11:07.671049 systemd-tmpfiles[1258]: Skipping /boot Sep 12 10:11:07.728563 zram_generator::config[1286]: No configuration found. Sep 12 10:11:07.959031 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 12 10:11:08.085310 systemd[1]: Reloading finished in 538 ms. Sep 12 10:11:08.102048 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 12 10:11:08.115880 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 10:11:08.134013 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 10:11:08.141187 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 12 10:11:08.146018 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 12 10:11:08.150393 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 10:11:08.161059 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 10:11:08.165915 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 12 10:11:08.171586 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.171805 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 10:11:08.185964 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 10:11:08.193163 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 10:11:08.198890 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 10:11:08.199452 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 10:11:08.200064 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 10:11:08.200181 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.207998 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.208272 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 10:11:08.208469 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 10:11:08.208577 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 10:11:08.217911 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 12 10:11:08.219204 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.220223 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 10:11:08.221781 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 10:11:08.235303 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.236901 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 10:11:08.246904 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 10:11:08.249718 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 10:11:08.250758 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 10:11:08.251486 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 10:11:08.251678 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.253780 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 12 10:11:08.255047 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 12 10:11:08.255413 systemd-udevd[1341]: Using default interface naming scheme 'v255'. Sep 12 10:11:08.273800 systemd[1]: Finished ensure-sysext.service. Sep 12 10:11:08.281930 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 10:11:08.282428 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 10:11:08.284027 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 10:11:08.284904 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 10:11:08.286934 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 10:11:08.294364 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 10:11:08.294724 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 10:11:08.296020 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 10:11:08.296789 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 10:11:08.306911 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 12 10:11:08.321970 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 10:11:08.322835 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 10:11:08.322935 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 10:11:08.333776 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 12 10:11:08.345831 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 12 10:11:08.346423 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 10:11:08.363389 augenrules[1385]: No rules Sep 12 10:11:08.362547 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 10:11:08.362809 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 10:11:08.393003 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 12 10:11:08.396578 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 12 10:11:08.516807 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1366) Sep 12 10:11:08.518387 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. Sep 12 10:11:08.535762 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Sep 12 10:11:08.536719 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.536889 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 10:11:08.543844 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 10:11:08.554830 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 10:11:08.567838 kernel: ISO 9660 Extensions: RRIP_1991A Sep 12 10:11:08.563925 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 10:11:08.564420 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 10:11:08.564462 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 10:11:08.564497 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 10:11:08.564516 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 10:11:08.564821 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 12 10:11:08.576662 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Sep 12 10:11:08.582919 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 10:11:08.584138 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 10:11:08.585901 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 10:11:08.587591 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 10:11:08.589863 systemd[1]: Reached target time-set.target - System Time Set. Sep 12 10:11:08.592776 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 10:11:08.614712 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 10:11:08.614986 systemd-networkd[1378]: lo: Link UP Sep 12 10:11:08.616907 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 10:11:08.617300 systemd-networkd[1378]: lo: Gained carrier Sep 12 10:11:08.621085 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 10:11:08.624717 systemd-networkd[1378]: Enumeration completed Sep 12 10:11:08.624869 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 10:11:08.625546 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 12 10:11:08.633569 systemd-networkd[1378]: eth1: Configuring with /run/systemd/network/10-5a:3a:a8:a8:3c:0a.network. Sep 12 10:11:08.637462 systemd-networkd[1378]: eth1: Link UP Sep 12 10:11:08.637677 systemd-networkd[1378]: eth1: Gained carrier Sep 12 10:11:08.638235 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 12 10:11:08.649810 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 12 10:11:08.650300 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:08.657155 systemd-resolved[1340]: Positive Trust Anchors: Sep 12 10:11:08.657173 systemd-resolved[1340]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 10:11:08.657211 systemd-resolved[1340]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 10:11:08.666579 systemd-resolved[1340]: Using system hostname 'ci-4230.2.2-n-dc4800d201'. Sep 12 10:11:08.669016 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 10:11:08.670849 systemd[1]: Reached target network.target - Network. Sep 12 10:11:08.671227 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 10:11:08.697560 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 12 10:11:08.698822 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 12 10:11:08.703658 systemd-networkd[1378]: eth0: Configuring with /run/systemd/network/10-2a:65:1e:2a:5f:42.network. Sep 12 10:11:08.704832 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:08.705118 systemd-networkd[1378]: eth0: Link UP Sep 12 10:11:08.705326 systemd-networkd[1378]: eth0: Gained carrier Sep 12 10:11:08.707573 kernel: ACPI: button: Power Button [PWRF] Sep 12 10:11:08.711417 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:08.713496 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:08.748665 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Sep 12 10:11:08.754571 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Sep 12 10:11:08.756628 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 12 10:11:08.763985 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 12 10:11:08.791462 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 12 10:11:08.827670 kernel: mousedev: PS/2 mouse device common for all mice Sep 12 10:11:08.844005 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:11:08.871506 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Sep 12 10:11:08.874064 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Sep 12 10:11:08.881751 kernel: Console: switching to colour dummy device 80x25 Sep 12 10:11:08.881849 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Sep 12 10:11:08.881868 kernel: [drm] features: -context_init Sep 12 10:11:08.887553 kernel: [drm] number of scanouts: 1 Sep 12 10:11:08.887680 kernel: [drm] number of cap sets: 0 Sep 12 10:11:08.889644 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Sep 12 10:11:08.908795 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 10:11:08.910067 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:08.912552 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Sep 12 10:11:08.912482 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 10:11:08.914325 kernel: Console: switching to colour frame buffer device 128x48 Sep 12 10:11:08.927560 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Sep 12 10:11:08.963676 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:11:08.984448 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 10:11:08.985610 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:09.005965 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 10:11:09.020569 kernel: EDAC MC: Ver: 3.0.0 Sep 12 10:11:09.056060 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 12 10:11:09.066942 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 12 10:11:09.078958 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 10:11:09.083071 lvm[1448]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 12 10:11:09.115214 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 12 10:11:09.117063 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 10:11:09.117205 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 10:11:09.117391 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 12 10:11:09.117501 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 12 10:11:09.117834 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 12 10:11:09.117997 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 12 10:11:09.118069 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 12 10:11:09.118137 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 12 10:11:09.118169 systemd[1]: Reached target paths.target - Path Units. Sep 12 10:11:09.118227 systemd[1]: Reached target timers.target - Timer Units. Sep 12 10:11:09.120150 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 12 10:11:09.123626 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 12 10:11:09.129148 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 12 10:11:09.130895 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 12 10:11:09.131930 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 12 10:11:09.146904 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 12 10:11:09.149952 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 12 10:11:09.158886 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 12 10:11:09.162297 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 12 10:11:09.164340 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 10:11:09.164975 systemd[1]: Reached target basic.target - Basic System. Sep 12 10:11:09.165465 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 12 10:11:09.165496 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 12 10:11:09.167264 lvm[1454]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 12 10:11:09.172825 systemd[1]: Starting containerd.service - containerd container runtime... Sep 12 10:11:09.185791 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 12 10:11:09.196784 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 12 10:11:09.202749 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 12 10:11:09.213886 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 12 10:11:09.214626 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 12 10:11:09.218817 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 12 10:11:09.230080 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 12 10:11:09.238841 jq[1458]: false Sep 12 10:11:09.237748 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 12 10:11:09.237380 dbus-daemon[1457]: [system] SELinux support is enabled Sep 12 10:11:09.241117 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 12 10:11:09.256810 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 12 10:11:09.260878 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 12 10:11:09.261630 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 12 10:11:09.265805 systemd[1]: Starting update-engine.service - Update Engine... Sep 12 10:11:09.277711 coreos-metadata[1456]: Sep 12 10:11:09.277 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 10:11:09.279960 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 12 10:11:09.281464 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 12 10:11:09.293624 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 12 10:11:09.302694 coreos-metadata[1456]: Sep 12 10:11:09.302 INFO Fetch successful Sep 12 10:11:09.306215 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 12 10:11:09.307652 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 12 10:11:09.310375 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 12 10:11:09.310858 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 12 10:11:09.325552 jq[1468]: true Sep 12 10:11:09.337138 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 12 10:11:09.337231 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 12 10:11:09.341353 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 12 10:11:09.341509 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Sep 12 10:11:09.343772 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 12 10:11:09.368040 extend-filesystems[1461]: Found loop4 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found loop5 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found loop6 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found loop7 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda1 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda2 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda3 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found usr Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda4 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda6 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda7 Sep 12 10:11:09.377818 extend-filesystems[1461]: Found vda9 Sep 12 10:11:09.440189 extend-filesystems[1461]: Checking size of /dev/vda9 Sep 12 10:11:09.453924 update_engine[1467]: I20250912 10:11:09.399989 1467 main.cc:92] Flatcar Update Engine starting Sep 12 10:11:09.453924 update_engine[1467]: I20250912 10:11:09.419102 1467 update_check_scheduler.cc:74] Next update check in 8m32s Sep 12 10:11:09.383026 (ntainerd)[1490]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 12 10:11:09.457611 extend-filesystems[1461]: Resized partition /dev/vda9 Sep 12 10:11:09.410448 systemd[1]: motdgen.service: Deactivated successfully. Sep 12 10:11:09.461432 jq[1481]: true Sep 12 10:11:09.411653 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 12 10:11:09.461762 tar[1472]: linux-amd64/LICENSE Sep 12 10:11:09.461762 tar[1472]: linux-amd64/helm Sep 12 10:11:09.432403 systemd[1]: Started update-engine.service - Update Engine. Sep 12 10:11:09.442935 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 12 10:11:09.466790 systemd-logind[1466]: New seat seat0. Sep 12 10:11:09.487235 extend-filesystems[1501]: resize2fs 1.47.1 (20-May-2024) Sep 12 10:11:09.501438 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Sep 12 10:11:09.479923 systemd-logind[1466]: Watching system buttons on /dev/input/event1 (Power Button) Sep 12 10:11:09.479945 systemd-logind[1466]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 12 10:11:09.480272 systemd[1]: Started systemd-logind.service - User Login Management. Sep 12 10:11:09.511629 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 12 10:11:09.512575 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 12 10:11:09.591882 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Sep 12 10:11:09.620281 extend-filesystems[1501]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 12 10:11:09.620281 extend-filesystems[1501]: old_desc_blocks = 1, new_desc_blocks = 8 Sep 12 10:11:09.620281 extend-filesystems[1501]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Sep 12 10:11:09.634514 extend-filesystems[1461]: Resized filesystem in /dev/vda9 Sep 12 10:11:09.634514 extend-filesystems[1461]: Found vdb Sep 12 10:11:09.621394 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 12 10:11:09.641117 bash[1518]: Updated "/home/core/.ssh/authorized_keys" Sep 12 10:11:09.622890 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 12 10:11:09.636617 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 12 10:11:09.646669 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1363) Sep 12 10:11:09.654119 systemd[1]: Starting sshkeys.service... Sep 12 10:11:09.767708 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 12 10:11:09.779844 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 12 10:11:09.854894 coreos-metadata[1529]: Sep 12 10:11:09.854 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 12 10:11:09.857578 sshd_keygen[1492]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 12 10:11:09.871341 coreos-metadata[1529]: Sep 12 10:11:09.871 INFO Fetch successful Sep 12 10:11:09.884368 unknown[1529]: wrote ssh authorized keys file for user: core Sep 12 10:11:09.889615 locksmithd[1497]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 12 10:11:09.920300 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 12 10:11:09.925402 update-ssh-keys[1543]: Updated "/home/core/.ssh/authorized_keys" Sep 12 10:11:09.929266 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 12 10:11:09.944868 systemd[1]: Finished sshkeys.service. Sep 12 10:11:09.964594 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 12 10:11:09.974306 systemd-networkd[1378]: eth0: Gained IPv6LL Sep 12 10:11:09.975196 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:09.981724 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 12 10:11:09.983057 systemd[1]: Reached target network-online.target - Network is Online. Sep 12 10:11:09.998990 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:10.003040 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 12 10:11:10.008207 systemd[1]: issuegen.service: Deactivated successfully. Sep 12 10:11:10.008525 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 12 10:11:10.025563 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 12 10:11:10.077600 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 12 10:11:10.091973 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 12 10:11:10.104159 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 12 10:11:10.106947 systemd[1]: Reached target getty.target - Login Prompts. Sep 12 10:11:10.114825 containerd[1490]: time="2025-09-12T10:11:10.113510694Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Sep 12 10:11:10.125348 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 12 10:11:10.201352 containerd[1490]: time="2025-09-12T10:11:10.200249905Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.205778 containerd[1490]: time="2025-09-12T10:11:10.205724121Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.105-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 12 10:11:10.206594 containerd[1490]: time="2025-09-12T10:11:10.206561287Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 12 10:11:10.206709 containerd[1490]: time="2025-09-12T10:11:10.206695567Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.206983114Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207007466Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207078603Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207090755Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207395354Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207421395Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207436355Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207449289Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207588130Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.207866459Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208422 containerd[1490]: time="2025-09-12T10:11:10.208031453Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 12 10:11:10.208728 containerd[1490]: time="2025-09-12T10:11:10.208045684Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 12 10:11:10.208728 containerd[1490]: time="2025-09-12T10:11:10.208147659Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 12 10:11:10.208728 containerd[1490]: time="2025-09-12T10:11:10.208198435Z" level=info msg="metadata content store policy set" policy=shared Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215192617Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215264721Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215284222Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215301356Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215318427Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215557351Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.215837430Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216020021Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216044122Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216065462Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216085726Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216100441Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216113270Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.218617 containerd[1490]: time="2025-09-12T10:11:10.216127859Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216143802Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216159914Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216172437Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216184673Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216205854Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216415932Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216435651Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216482318Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216504182Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216522319Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216550424Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.216567740Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.217647132Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219033 containerd[1490]: time="2025-09-12T10:11:10.217716304Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217736485Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217755077Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217769422Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217788287Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217823880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217842852Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217857031Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217911071Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217932760Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217948198Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217963771Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217976192Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.217991826Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 12 10:11:10.219324 containerd[1490]: time="2025-09-12T10:11:10.218003055Z" level=info msg="NRI interface is disabled by configuration." Sep 12 10:11:10.219620 containerd[1490]: time="2025-09-12T10:11:10.218016224Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 12 10:11:10.219645 containerd[1490]: time="2025-09-12T10:11:10.218360492Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 12 10:11:10.219645 containerd[1490]: time="2025-09-12T10:11:10.218434930Z" level=info msg="Connect containerd service" Sep 12 10:11:10.219645 containerd[1490]: time="2025-09-12T10:11:10.218500194Z" level=info msg="using legacy CRI server" Sep 12 10:11:10.219645 containerd[1490]: time="2025-09-12T10:11:10.218514309Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 12 10:11:10.219645 containerd[1490]: time="2025-09-12T10:11:10.219239431Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 12 10:11:10.225832 containerd[1490]: time="2025-09-12T10:11:10.225764153Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226676286Z" level=info msg="Start subscribing containerd event" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226758298Z" level=info msg="Start recovering state" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226870112Z" level=info msg="Start event monitor" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226899176Z" level=info msg="Start snapshots syncer" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226914358Z" level=info msg="Start cni network conf syncer for default" Sep 12 10:11:10.227293 containerd[1490]: time="2025-09-12T10:11:10.226924189Z" level=info msg="Start streaming server" Sep 12 10:11:10.227513 containerd[1490]: time="2025-09-12T10:11:10.227415661Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 12 10:11:10.227513 containerd[1490]: time="2025-09-12T10:11:10.227485613Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 12 10:11:10.231970 containerd[1490]: time="2025-09-12T10:11:10.227719446Z" level=info msg="containerd successfully booted in 0.125867s" Sep 12 10:11:10.228024 systemd[1]: Started containerd.service - containerd container runtime. Sep 12 10:11:10.294848 systemd-networkd[1378]: eth1: Gained IPv6LL Sep 12 10:11:10.295682 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:10.551158 tar[1472]: linux-amd64/README.md Sep 12 10:11:10.566458 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 12 10:11:11.106682 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:11.110051 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 12 10:11:11.112841 systemd[1]: Startup finished in 939ms (kernel) + 5.495s (initrd) + 5.964s (userspace) = 12.399s. Sep 12 10:11:11.118066 (kubelet)[1582]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 10:11:11.701706 kubelet[1582]: E0912 10:11:11.701597 1582 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 10:11:11.704988 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 10:11:11.705164 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 10:11:11.705909 systemd[1]: kubelet.service: Consumed 1.198s CPU time, 264.9M memory peak. Sep 12 10:11:13.860045 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 12 10:11:13.864863 systemd[1]: Started sshd@0-164.92.125.86:22-139.178.68.195:45534.service - OpenSSH per-connection server daemon (139.178.68.195:45534). Sep 12 10:11:13.940916 sshd[1594]: Accepted publickey for core from 139.178.68.195 port 45534 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:13.943353 sshd-session[1594]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:13.951090 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 12 10:11:13.955939 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 12 10:11:13.965556 systemd-logind[1466]: New session 1 of user core. Sep 12 10:11:13.977417 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 12 10:11:13.986083 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 12 10:11:13.996496 (systemd)[1598]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 12 10:11:13.999799 systemd-logind[1466]: New session c1 of user core. Sep 12 10:11:14.215417 systemd[1598]: Queued start job for default target default.target. Sep 12 10:11:14.226937 systemd[1598]: Created slice app.slice - User Application Slice. Sep 12 10:11:14.226973 systemd[1598]: Reached target paths.target - Paths. Sep 12 10:11:14.227024 systemd[1598]: Reached target timers.target - Timers. Sep 12 10:11:14.228760 systemd[1598]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 12 10:11:14.243666 systemd[1598]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 12 10:11:14.243878 systemd[1598]: Reached target sockets.target - Sockets. Sep 12 10:11:14.243931 systemd[1598]: Reached target basic.target - Basic System. Sep 12 10:11:14.243971 systemd[1598]: Reached target default.target - Main User Target. Sep 12 10:11:14.244009 systemd[1598]: Startup finished in 234ms. Sep 12 10:11:14.244187 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 12 10:11:14.253211 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 12 10:11:14.326117 systemd[1]: Started sshd@1-164.92.125.86:22-139.178.68.195:45540.service - OpenSSH per-connection server daemon (139.178.68.195:45540). Sep 12 10:11:14.372520 sshd[1609]: Accepted publickey for core from 139.178.68.195 port 45540 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:14.374611 sshd-session[1609]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:14.380465 systemd-logind[1466]: New session 2 of user core. Sep 12 10:11:14.390894 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 12 10:11:14.454154 sshd[1611]: Connection closed by 139.178.68.195 port 45540 Sep 12 10:11:14.453948 sshd-session[1609]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:14.465213 systemd[1]: sshd@1-164.92.125.86:22-139.178.68.195:45540.service: Deactivated successfully. Sep 12 10:11:14.467410 systemd[1]: session-2.scope: Deactivated successfully. Sep 12 10:11:14.469420 systemd-logind[1466]: Session 2 logged out. Waiting for processes to exit. Sep 12 10:11:14.474073 systemd[1]: Started sshd@2-164.92.125.86:22-139.178.68.195:45552.service - OpenSSH per-connection server daemon (139.178.68.195:45552). Sep 12 10:11:14.476148 systemd-logind[1466]: Removed session 2. Sep 12 10:11:14.529390 sshd[1616]: Accepted publickey for core from 139.178.68.195 port 45552 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:14.531164 sshd-session[1616]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:14.537816 systemd-logind[1466]: New session 3 of user core. Sep 12 10:11:14.543881 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 12 10:11:14.607573 sshd[1619]: Connection closed by 139.178.68.195 port 45552 Sep 12 10:11:14.607911 sshd-session[1616]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:14.622095 systemd[1]: sshd@2-164.92.125.86:22-139.178.68.195:45552.service: Deactivated successfully. Sep 12 10:11:14.624824 systemd[1]: session-3.scope: Deactivated successfully. Sep 12 10:11:14.627081 systemd-logind[1466]: Session 3 logged out. Waiting for processes to exit. Sep 12 10:11:14.632992 systemd[1]: Started sshd@3-164.92.125.86:22-139.178.68.195:45556.service - OpenSSH per-connection server daemon (139.178.68.195:45556). Sep 12 10:11:14.634911 systemd-logind[1466]: Removed session 3. Sep 12 10:11:14.682434 sshd[1624]: Accepted publickey for core from 139.178.68.195 port 45556 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:14.684341 sshd-session[1624]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:14.689969 systemd-logind[1466]: New session 4 of user core. Sep 12 10:11:14.700960 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 12 10:11:14.765371 sshd[1627]: Connection closed by 139.178.68.195 port 45556 Sep 12 10:11:14.764679 sshd-session[1624]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:14.774703 systemd[1]: sshd@3-164.92.125.86:22-139.178.68.195:45556.service: Deactivated successfully. Sep 12 10:11:14.777064 systemd[1]: session-4.scope: Deactivated successfully. Sep 12 10:11:14.779794 systemd-logind[1466]: Session 4 logged out. Waiting for processes to exit. Sep 12 10:11:14.785987 systemd[1]: Started sshd@4-164.92.125.86:22-139.178.68.195:45560.service - OpenSSH per-connection server daemon (139.178.68.195:45560). Sep 12 10:11:14.788572 systemd-logind[1466]: Removed session 4. Sep 12 10:11:14.835213 sshd[1632]: Accepted publickey for core from 139.178.68.195 port 45560 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:14.837185 sshd-session[1632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:14.842956 systemd-logind[1466]: New session 5 of user core. Sep 12 10:11:14.849820 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 12 10:11:14.924649 sudo[1636]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 12 10:11:14.925094 sudo[1636]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 10:11:14.943832 sudo[1636]: pam_unix(sudo:session): session closed for user root Sep 12 10:11:14.946988 sshd[1635]: Connection closed by 139.178.68.195 port 45560 Sep 12 10:11:14.947944 sshd-session[1632]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:14.961633 systemd[1]: sshd@4-164.92.125.86:22-139.178.68.195:45560.service: Deactivated successfully. Sep 12 10:11:14.964247 systemd[1]: session-5.scope: Deactivated successfully. Sep 12 10:11:14.966723 systemd-logind[1466]: Session 5 logged out. Waiting for processes to exit. Sep 12 10:11:14.973028 systemd[1]: Started sshd@5-164.92.125.86:22-139.178.68.195:45574.service - OpenSSH per-connection server daemon (139.178.68.195:45574). Sep 12 10:11:14.974670 systemd-logind[1466]: Removed session 5. Sep 12 10:11:15.036743 sshd[1641]: Accepted publickey for core from 139.178.68.195 port 45574 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:15.038052 sshd-session[1641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:15.045411 systemd-logind[1466]: New session 6 of user core. Sep 12 10:11:15.055925 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 12 10:11:15.118941 sudo[1646]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 12 10:11:15.119272 sudo[1646]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 10:11:15.125513 sudo[1646]: pam_unix(sudo:session): session closed for user root Sep 12 10:11:15.133168 sudo[1645]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 12 10:11:15.133526 sudo[1645]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 10:11:15.153106 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 10:11:15.194451 augenrules[1668]: No rules Sep 12 10:11:15.196716 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 10:11:15.197053 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 10:11:15.198859 sudo[1645]: pam_unix(sudo:session): session closed for user root Sep 12 10:11:15.203003 sshd[1644]: Connection closed by 139.178.68.195 port 45574 Sep 12 10:11:15.203958 sshd-session[1641]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:15.219097 systemd[1]: sshd@5-164.92.125.86:22-139.178.68.195:45574.service: Deactivated successfully. Sep 12 10:11:15.221648 systemd[1]: session-6.scope: Deactivated successfully. Sep 12 10:11:15.222676 systemd-logind[1466]: Session 6 logged out. Waiting for processes to exit. Sep 12 10:11:15.229927 systemd[1]: Started sshd@6-164.92.125.86:22-139.178.68.195:45584.service - OpenSSH per-connection server daemon (139.178.68.195:45584). Sep 12 10:11:15.232182 systemd-logind[1466]: Removed session 6. Sep 12 10:11:15.291443 sshd[1676]: Accepted publickey for core from 139.178.68.195 port 45584 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:11:15.293642 sshd-session[1676]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:11:15.300809 systemd-logind[1466]: New session 7 of user core. Sep 12 10:11:15.308910 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 12 10:11:15.370117 sudo[1680]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 12 10:11:15.370912 sudo[1680]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 10:11:15.813957 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 12 10:11:15.816667 (dockerd)[1698]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 12 10:11:16.438592 dockerd[1698]: time="2025-09-12T10:11:16.437823600Z" level=info msg="Starting up" Sep 12 10:11:16.560381 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport3910523536-merged.mount: Deactivated successfully. Sep 12 10:11:16.593856 dockerd[1698]: time="2025-09-12T10:11:16.593587780Z" level=info msg="Loading containers: start." Sep 12 10:11:16.789602 kernel: Initializing XFRM netlink socket Sep 12 10:11:16.828403 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:16.828503 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:16.845337 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:16.892630 systemd-networkd[1378]: docker0: Link UP Sep 12 10:11:16.892960 systemd-timesyncd[1380]: Network configuration changed, trying to establish connection. Sep 12 10:11:16.935888 dockerd[1698]: time="2025-09-12T10:11:16.935822385Z" level=info msg="Loading containers: done." Sep 12 10:11:16.952640 dockerd[1698]: time="2025-09-12T10:11:16.952305733Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 12 10:11:16.952640 dockerd[1698]: time="2025-09-12T10:11:16.952427348Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Sep 12 10:11:16.952908 dockerd[1698]: time="2025-09-12T10:11:16.952889401Z" level=info msg="Daemon has completed initialization" Sep 12 10:11:16.987884 dockerd[1698]: time="2025-09-12T10:11:16.987651859Z" level=info msg="API listen on /run/docker.sock" Sep 12 10:11:16.990867 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 12 10:11:17.854748 containerd[1490]: time="2025-09-12T10:11:17.854703493Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Sep 12 10:11:18.507679 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2812500081.mount: Deactivated successfully. Sep 12 10:11:19.617189 containerd[1490]: time="2025-09-12T10:11:19.615730278Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:19.617189 containerd[1490]: time="2025-09-12T10:11:19.616541575Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Sep 12 10:11:19.617189 containerd[1490]: time="2025-09-12T10:11:19.617124184Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:19.620613 containerd[1490]: time="2025-09-12T10:11:19.620566447Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:19.622083 containerd[1490]: time="2025-09-12T10:11:19.622042384Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.767293415s" Sep 12 10:11:19.622264 containerd[1490]: time="2025-09-12T10:11:19.622240383Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Sep 12 10:11:19.623487 containerd[1490]: time="2025-09-12T10:11:19.623457704Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Sep 12 10:11:20.972559 containerd[1490]: time="2025-09-12T10:11:20.971459977Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:20.972559 containerd[1490]: time="2025-09-12T10:11:20.972083728Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Sep 12 10:11:20.973797 containerd[1490]: time="2025-09-12T10:11:20.973295156Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:20.976727 containerd[1490]: time="2025-09-12T10:11:20.976687180Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:20.977903 containerd[1490]: time="2025-09-12T10:11:20.977867532Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.354234731s" Sep 12 10:11:20.978033 containerd[1490]: time="2025-09-12T10:11:20.978017513Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Sep 12 10:11:20.978929 containerd[1490]: time="2025-09-12T10:11:20.978905713Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Sep 12 10:11:21.759847 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 12 10:11:21.767849 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:21.932748 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:21.942472 (kubelet)[1968]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 10:11:22.001597 kubelet[1968]: E0912 10:11:22.001428 1968 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 10:11:22.007336 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 10:11:22.007563 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 10:11:22.008287 systemd[1]: kubelet.service: Consumed 183ms CPU time, 112.5M memory peak. Sep 12 10:11:22.376304 containerd[1490]: time="2025-09-12T10:11:22.376254164Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:22.377416 containerd[1490]: time="2025-09-12T10:11:22.376935791Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Sep 12 10:11:22.379514 containerd[1490]: time="2025-09-12T10:11:22.377637170Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:22.380822 containerd[1490]: time="2025-09-12T10:11:22.380790452Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:22.382989 containerd[1490]: time="2025-09-12T10:11:22.382935644Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 1.403997118s" Sep 12 10:11:22.382989 containerd[1490]: time="2025-09-12T10:11:22.382988895Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Sep 12 10:11:22.383611 containerd[1490]: time="2025-09-12T10:11:22.383579913Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Sep 12 10:11:23.374924 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1285796011.mount: Deactivated successfully. Sep 12 10:11:23.837313 containerd[1490]: time="2025-09-12T10:11:23.836578535Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:23.837857 containerd[1490]: time="2025-09-12T10:11:23.837818627Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Sep 12 10:11:23.838048 containerd[1490]: time="2025-09-12T10:11:23.838028166Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:23.839969 containerd[1490]: time="2025-09-12T10:11:23.839934131Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:23.840755 containerd[1490]: time="2025-09-12T10:11:23.840729214Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.456902104s" Sep 12 10:11:23.840893 containerd[1490]: time="2025-09-12T10:11:23.840876906Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Sep 12 10:11:23.841465 containerd[1490]: time="2025-09-12T10:11:23.841432894Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 12 10:11:24.016854 systemd-resolved[1340]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Sep 12 10:11:24.345646 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3635067557.mount: Deactivated successfully. Sep 12 10:11:25.193028 containerd[1490]: time="2025-09-12T10:11:25.192956716Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.195296 containerd[1490]: time="2025-09-12T10:11:25.194804617Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Sep 12 10:11:25.195296 containerd[1490]: time="2025-09-12T10:11:25.195240349Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.200115 containerd[1490]: time="2025-09-12T10:11:25.198566556Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.200115 containerd[1490]: time="2025-09-12T10:11:25.199960186Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.358358608s" Sep 12 10:11:25.200115 containerd[1490]: time="2025-09-12T10:11:25.200004239Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 12 10:11:25.201100 containerd[1490]: time="2025-09-12T10:11:25.201056643Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 12 10:11:25.689285 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1320587819.mount: Deactivated successfully. Sep 12 10:11:25.693364 containerd[1490]: time="2025-09-12T10:11:25.693307098Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.694710 containerd[1490]: time="2025-09-12T10:11:25.694652017Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 12 10:11:25.695140 containerd[1490]: time="2025-09-12T10:11:25.695086053Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.698708 containerd[1490]: time="2025-09-12T10:11:25.698644636Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:25.699562 containerd[1490]: time="2025-09-12T10:11:25.699246516Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 498.150044ms" Sep 12 10:11:25.699562 containerd[1490]: time="2025-09-12T10:11:25.699292229Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 12 10:11:25.700460 containerd[1490]: time="2025-09-12T10:11:25.700431774Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Sep 12 10:11:26.230167 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3362620832.mount: Deactivated successfully. Sep 12 10:11:27.125796 systemd-resolved[1340]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Sep 12 10:11:27.857582 containerd[1490]: time="2025-09-12T10:11:27.856590738Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:27.858759 containerd[1490]: time="2025-09-12T10:11:27.858471670Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Sep 12 10:11:27.860557 containerd[1490]: time="2025-09-12T10:11:27.859286328Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:27.862258 containerd[1490]: time="2025-09-12T10:11:27.862210413Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:27.864559 containerd[1490]: time="2025-09-12T10:11:27.863806591Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.163341936s" Sep 12 10:11:27.864559 containerd[1490]: time="2025-09-12T10:11:27.863846229Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Sep 12 10:11:30.296983 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:30.297305 systemd[1]: kubelet.service: Consumed 183ms CPU time, 112.5M memory peak. Sep 12 10:11:30.306993 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:30.355036 systemd[1]: Reload requested from client PID 2120 ('systemctl') (unit session-7.scope)... Sep 12 10:11:30.355080 systemd[1]: Reloading... Sep 12 10:11:30.518959 zram_generator::config[2164]: No configuration found. Sep 12 10:11:30.707180 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 12 10:11:30.855899 systemd[1]: Reloading finished in 499 ms. Sep 12 10:11:30.916650 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:30.922497 (kubelet)[2209]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 10:11:30.926039 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:30.927372 systemd[1]: kubelet.service: Deactivated successfully. Sep 12 10:11:30.927792 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:30.927963 systemd[1]: kubelet.service: Consumed 121ms CPU time, 99.1M memory peak. Sep 12 10:11:30.933881 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:31.071660 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:31.077405 (kubelet)[2221]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 10:11:31.129279 kubelet[2221]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 10:11:31.129723 kubelet[2221]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 12 10:11:31.129773 kubelet[2221]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 10:11:31.129925 kubelet[2221]: I0912 10:11:31.129898 2221 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 10:11:32.059959 kubelet[2221]: I0912 10:11:32.059904 2221 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 12 10:11:32.060198 kubelet[2221]: I0912 10:11:32.060186 2221 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 10:11:32.060593 kubelet[2221]: I0912 10:11:32.060576 2221 server.go:954] "Client rotation is on, will bootstrap in background" Sep 12 10:11:32.089672 kubelet[2221]: E0912 10:11:32.089615 2221 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://164.92.125.86:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:32.090716 kubelet[2221]: I0912 10:11:32.090688 2221 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 10:11:32.103245 kubelet[2221]: E0912 10:11:32.103201 2221 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 12 10:11:32.103245 kubelet[2221]: I0912 10:11:32.103231 2221 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 12 10:11:32.107924 kubelet[2221]: I0912 10:11:32.107875 2221 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 10:11:32.110583 kubelet[2221]: I0912 10:11:32.109971 2221 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 10:11:32.110583 kubelet[2221]: I0912 10:11:32.110041 2221 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.2.2-n-dc4800d201","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 10:11:32.110583 kubelet[2221]: I0912 10:11:32.110304 2221 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 10:11:32.110583 kubelet[2221]: I0912 10:11:32.110315 2221 container_manager_linux.go:304] "Creating device plugin manager" Sep 12 10:11:32.110923 kubelet[2221]: I0912 10:11:32.110465 2221 state_mem.go:36] "Initialized new in-memory state store" Sep 12 10:11:32.114773 kubelet[2221]: I0912 10:11:32.114735 2221 kubelet.go:446] "Attempting to sync node with API server" Sep 12 10:11:32.114950 kubelet[2221]: I0912 10:11:32.114939 2221 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 10:11:32.115021 kubelet[2221]: I0912 10:11:32.115014 2221 kubelet.go:352] "Adding apiserver pod source" Sep 12 10:11:32.115066 kubelet[2221]: I0912 10:11:32.115059 2221 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 10:11:32.119030 kubelet[2221]: W0912 10:11:32.118678 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://164.92.125.86:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-n-dc4800d201&limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:32.119030 kubelet[2221]: E0912 10:11:32.118740 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://164.92.125.86:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-n-dc4800d201&limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:32.119606 kubelet[2221]: W0912 10:11:32.119297 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://164.92.125.86:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:32.119606 kubelet[2221]: E0912 10:11:32.119344 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://164.92.125.86:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:32.121007 kubelet[2221]: I0912 10:11:32.120963 2221 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Sep 12 10:11:32.125051 kubelet[2221]: I0912 10:11:32.125008 2221 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 10:11:32.126563 kubelet[2221]: W0912 10:11:32.125733 2221 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 12 10:11:32.127130 kubelet[2221]: I0912 10:11:32.127095 2221 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 12 10:11:32.127223 kubelet[2221]: I0912 10:11:32.127148 2221 server.go:1287] "Started kubelet" Sep 12 10:11:32.128573 kubelet[2221]: I0912 10:11:32.128316 2221 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 10:11:32.129343 kubelet[2221]: I0912 10:11:32.129325 2221 server.go:479] "Adding debug handlers to kubelet server" Sep 12 10:11:32.132480 kubelet[2221]: I0912 10:11:32.132298 2221 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 10:11:32.133426 kubelet[2221]: I0912 10:11:32.133359 2221 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 10:11:32.133646 kubelet[2221]: I0912 10:11:32.133627 2221 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 10:11:32.137341 kubelet[2221]: E0912 10:11:32.134844 2221 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://164.92.125.86:6443/api/v1/namespaces/default/events\": dial tcp 164.92.125.86:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4230.2.2-n-dc4800d201.18648145f9ecf588 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230.2.2-n-dc4800d201,UID:ci-4230.2.2-n-dc4800d201,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230.2.2-n-dc4800d201,},FirstTimestamp:2025-09-12 10:11:32.127118728 +0000 UTC m=+1.045273254,LastTimestamp:2025-09-12 10:11:32.127118728 +0000 UTC m=+1.045273254,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230.2.2-n-dc4800d201,}" Sep 12 10:11:32.137526 kubelet[2221]: I0912 10:11:32.137473 2221 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 10:11:32.141403 kubelet[2221]: I0912 10:11:32.141375 2221 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 12 10:11:32.142557 kubelet[2221]: E0912 10:11:32.141869 2221 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4230.2.2-n-dc4800d201\" not found" Sep 12 10:11:32.142557 kubelet[2221]: I0912 10:11:32.142197 2221 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 12 10:11:32.142557 kubelet[2221]: I0912 10:11:32.142253 2221 reconciler.go:26] "Reconciler: start to sync state" Sep 12 10:11:32.145307 kubelet[2221]: W0912 10:11:32.145239 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://164.92.125.86:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:32.145438 kubelet[2221]: E0912 10:11:32.145344 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://164.92.125.86:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:32.145477 kubelet[2221]: E0912 10:11:32.145447 2221 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://164.92.125.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-n-dc4800d201?timeout=10s\": dial tcp 164.92.125.86:6443: connect: connection refused" interval="200ms" Sep 12 10:11:32.146668 kubelet[2221]: I0912 10:11:32.146165 2221 factory.go:221] Registration of the systemd container factory successfully Sep 12 10:11:32.146668 kubelet[2221]: I0912 10:11:32.146264 2221 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 10:11:32.148615 kubelet[2221]: I0912 10:11:32.148217 2221 factory.go:221] Registration of the containerd container factory successfully Sep 12 10:11:32.158232 kubelet[2221]: I0912 10:11:32.158175 2221 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 10:11:32.159799 kubelet[2221]: I0912 10:11:32.159765 2221 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 10:11:32.159943 kubelet[2221]: I0912 10:11:32.159933 2221 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 12 10:11:32.160015 kubelet[2221]: I0912 10:11:32.160006 2221 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 12 10:11:32.160075 kubelet[2221]: I0912 10:11:32.160067 2221 kubelet.go:2382] "Starting kubelet main sync loop" Sep 12 10:11:32.160185 kubelet[2221]: E0912 10:11:32.160167 2221 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 10:11:32.168858 kubelet[2221]: W0912 10:11:32.168792 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://164.92.125.86:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:32.168858 kubelet[2221]: E0912 10:11:32.168849 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://164.92.125.86:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:32.175567 kubelet[2221]: E0912 10:11:32.175494 2221 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 10:11:32.178661 kubelet[2221]: I0912 10:11:32.178633 2221 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 12 10:11:32.178837 kubelet[2221]: I0912 10:11:32.178826 2221 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 12 10:11:32.178903 kubelet[2221]: I0912 10:11:32.178895 2221 state_mem.go:36] "Initialized new in-memory state store" Sep 12 10:11:32.180471 kubelet[2221]: I0912 10:11:32.180449 2221 policy_none.go:49] "None policy: Start" Sep 12 10:11:32.180638 kubelet[2221]: I0912 10:11:32.180627 2221 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 12 10:11:32.180696 kubelet[2221]: I0912 10:11:32.180689 2221 state_mem.go:35] "Initializing new in-memory state store" Sep 12 10:11:32.186517 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 12 10:11:32.195600 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 12 10:11:32.199872 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 12 10:11:32.212781 kubelet[2221]: I0912 10:11:32.212742 2221 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 10:11:32.213441 kubelet[2221]: I0912 10:11:32.213422 2221 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 10:11:32.213825 kubelet[2221]: I0912 10:11:32.213778 2221 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 10:11:32.214104 kubelet[2221]: I0912 10:11:32.214092 2221 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 10:11:32.215872 kubelet[2221]: E0912 10:11:32.215713 2221 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 12 10:11:32.215872 kubelet[2221]: E0912 10:11:32.215817 2221 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4230.2.2-n-dc4800d201\" not found" Sep 12 10:11:32.276770 systemd[1]: Created slice kubepods-burstable-pod724ea47755c3eb2d655fa4e278e349e1.slice - libcontainer container kubepods-burstable-pod724ea47755c3eb2d655fa4e278e349e1.slice. Sep 12 10:11:32.294889 kubelet[2221]: E0912 10:11:32.294824 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.299918 systemd[1]: Created slice kubepods-burstable-pod87e38bff5b8b029a6f3c49d3080af986.slice - libcontainer container kubepods-burstable-pod87e38bff5b8b029a6f3c49d3080af986.slice. Sep 12 10:11:32.302108 kubelet[2221]: E0912 10:11:32.302076 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.309964 systemd[1]: Created slice kubepods-burstable-pod219a36a897fd955b6971f95b188cf014.slice - libcontainer container kubepods-burstable-pod219a36a897fd955b6971f95b188cf014.slice. Sep 12 10:11:32.315305 kubelet[2221]: E0912 10:11:32.315011 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.316170 kubelet[2221]: I0912 10:11:32.316139 2221 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.316633 kubelet[2221]: E0912 10:11:32.316568 2221 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://164.92.125.86:6443/api/v1/nodes\": dial tcp 164.92.125.86:6443: connect: connection refused" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.346757 kubelet[2221]: E0912 10:11:32.346707 2221 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://164.92.125.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-n-dc4800d201?timeout=10s\": dial tcp 164.92.125.86:6443: connect: connection refused" interval="400ms" Sep 12 10:11:32.444345 kubelet[2221]: I0912 10:11:32.444219 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-ca-certs\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444345 kubelet[2221]: I0912 10:11:32.444269 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-k8s-certs\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444345 kubelet[2221]: I0912 10:11:32.444291 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444345 kubelet[2221]: I0912 10:11:32.444331 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444345 kubelet[2221]: I0912 10:11:32.444353 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-k8s-certs\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444665 kubelet[2221]: I0912 10:11:32.444368 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-ca-certs\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444665 kubelet[2221]: I0912 10:11:32.444384 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-kubeconfig\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444665 kubelet[2221]: I0912 10:11:32.444402 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.444665 kubelet[2221]: I0912 10:11:32.444418 2221 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/219a36a897fd955b6971f95b188cf014-kubeconfig\") pod \"kube-scheduler-ci-4230.2.2-n-dc4800d201\" (UID: \"219a36a897fd955b6971f95b188cf014\") " pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.518595 kubelet[2221]: I0912 10:11:32.518546 2221 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.518945 kubelet[2221]: E0912 10:11:32.518916 2221 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://164.92.125.86:6443/api/v1/nodes\": dial tcp 164.92.125.86:6443: connect: connection refused" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.596349 kubelet[2221]: E0912 10:11:32.596213 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:32.597971 containerd[1490]: time="2025-09-12T10:11:32.597925528Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.2.2-n-dc4800d201,Uid:724ea47755c3eb2d655fa4e278e349e1,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:32.603426 kubelet[2221]: E0912 10:11:32.603353 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:32.604412 containerd[1490]: time="2025-09-12T10:11:32.604131196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.2.2-n-dc4800d201,Uid:87e38bff5b8b029a6f3c49d3080af986,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:32.604741 systemd-resolved[1340]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.3. Sep 12 10:11:32.615989 kubelet[2221]: E0912 10:11:32.615940 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:32.617185 containerd[1490]: time="2025-09-12T10:11:32.616813893Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.2.2-n-dc4800d201,Uid:219a36a897fd955b6971f95b188cf014,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:32.747604 kubelet[2221]: E0912 10:11:32.747559 2221 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://164.92.125.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-n-dc4800d201?timeout=10s\": dial tcp 164.92.125.86:6443: connect: connection refused" interval="800ms" Sep 12 10:11:32.921018 kubelet[2221]: I0912 10:11:32.920890 2221 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.922022 kubelet[2221]: E0912 10:11:32.921613 2221 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://164.92.125.86:6443/api/v1/nodes\": dial tcp 164.92.125.86:6443: connect: connection refused" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:32.988767 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3762300471.mount: Deactivated successfully. Sep 12 10:11:32.992121 containerd[1490]: time="2025-09-12T10:11:32.992066557Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 10:11:32.993136 containerd[1490]: time="2025-09-12T10:11:32.993095697Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 10:11:32.993977 containerd[1490]: time="2025-09-12T10:11:32.993942676Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 12 10:11:32.994301 containerd[1490]: time="2025-09-12T10:11:32.994271221Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 12 10:11:32.996554 containerd[1490]: time="2025-09-12T10:11:32.994947658Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 10:11:32.996554 containerd[1490]: time="2025-09-12T10:11:32.995654783Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 12 10:11:32.998807 containerd[1490]: time="2025-09-12T10:11:32.998776551Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 10:11:33.000735 containerd[1490]: time="2025-09-12T10:11:33.000696222Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 396.474379ms" Sep 12 10:11:33.002359 containerd[1490]: time="2025-09-12T10:11:33.001942106Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 399.41505ms" Sep 12 10:11:33.004553 containerd[1490]: time="2025-09-12T10:11:33.003770885Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 10:11:33.006275 containerd[1490]: time="2025-09-12T10:11:33.006236096Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 389.312781ms" Sep 12 10:11:33.079656 kubelet[2221]: W0912 10:11:33.079516 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://164.92.125.86:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:33.079656 kubelet[2221]: E0912 10:11:33.079614 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://164.92.125.86:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:33.092073 kubelet[2221]: W0912 10:11:33.091954 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://164.92.125.86:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-n-dc4800d201&limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:33.092073 kubelet[2221]: E0912 10:11:33.092031 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://164.92.125.86:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-n-dc4800d201&limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:33.164282 kubelet[2221]: W0912 10:11:33.164204 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://164.92.125.86:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:33.164282 kubelet[2221]: E0912 10:11:33.164289 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://164.92.125.86:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:33.168713 containerd[1490]: time="2025-09-12T10:11:33.168122486Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:33.169822 containerd[1490]: time="2025-09-12T10:11:33.169588653Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:33.169822 containerd[1490]: time="2025-09-12T10:11:33.169626904Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.169822 containerd[1490]: time="2025-09-12T10:11:33.169726474Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.174269 containerd[1490]: time="2025-09-12T10:11:33.173675622Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:33.174269 containerd[1490]: time="2025-09-12T10:11:33.173735852Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:33.174269 containerd[1490]: time="2025-09-12T10:11:33.173751542Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.174269 containerd[1490]: time="2025-09-12T10:11:33.174027698Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:33.174269 containerd[1490]: time="2025-09-12T10:11:33.174209651Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:33.175039 containerd[1490]: time="2025-09-12T10:11:33.174964888Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.182873 containerd[1490]: time="2025-09-12T10:11:33.179580751Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.182873 containerd[1490]: time="2025-09-12T10:11:33.180606919Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:33.213863 systemd[1]: Started cri-containerd-b51fa4472850950ddae44de5286a42453f01b6f0fafc85c6a97d8f5792f8373a.scope - libcontainer container b51fa4472850950ddae44de5286a42453f01b6f0fafc85c6a97d8f5792f8373a. Sep 12 10:11:33.220774 systemd[1]: Started cri-containerd-0d8f06bf60ffbed149fa185f07f03c4623ced71d1183b5630f368a04a49350a7.scope - libcontainer container 0d8f06bf60ffbed149fa185f07f03c4623ced71d1183b5630f368a04a49350a7. Sep 12 10:11:33.222655 systemd[1]: Started cri-containerd-25a88108ecffdf603f40f59a2f9823be89e98f74a98b61395dadc9e71b93987c.scope - libcontainer container 25a88108ecffdf603f40f59a2f9823be89e98f74a98b61395dadc9e71b93987c. Sep 12 10:11:33.291719 containerd[1490]: time="2025-09-12T10:11:33.291673600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.2.2-n-dc4800d201,Uid:724ea47755c3eb2d655fa4e278e349e1,Namespace:kube-system,Attempt:0,} returns sandbox id \"0d8f06bf60ffbed149fa185f07f03c4623ced71d1183b5630f368a04a49350a7\"" Sep 12 10:11:33.295680 kubelet[2221]: E0912 10:11:33.295557 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:33.303847 containerd[1490]: time="2025-09-12T10:11:33.303351723Z" level=info msg="CreateContainer within sandbox \"0d8f06bf60ffbed149fa185f07f03c4623ced71d1183b5630f368a04a49350a7\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 12 10:11:33.319550 containerd[1490]: time="2025-09-12T10:11:33.319492902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.2.2-n-dc4800d201,Uid:87e38bff5b8b029a6f3c49d3080af986,Namespace:kube-system,Attempt:0,} returns sandbox id \"b51fa4472850950ddae44de5286a42453f01b6f0fafc85c6a97d8f5792f8373a\"" Sep 12 10:11:33.320510 containerd[1490]: time="2025-09-12T10:11:33.320480421Z" level=info msg="CreateContainer within sandbox \"0d8f06bf60ffbed149fa185f07f03c4623ced71d1183b5630f368a04a49350a7\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"14e1ed7d2616ed1b707976377e47c6ce22a18374c94773498fbcaf13ec749291\"" Sep 12 10:11:33.320620 kubelet[2221]: E0912 10:11:33.320518 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:33.321568 containerd[1490]: time="2025-09-12T10:11:33.321544945Z" level=info msg="StartContainer for \"14e1ed7d2616ed1b707976377e47c6ce22a18374c94773498fbcaf13ec749291\"" Sep 12 10:11:33.325404 containerd[1490]: time="2025-09-12T10:11:33.325278407Z" level=info msg="CreateContainer within sandbox \"b51fa4472850950ddae44de5286a42453f01b6f0fafc85c6a97d8f5792f8373a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 12 10:11:33.329670 containerd[1490]: time="2025-09-12T10:11:33.329631032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.2.2-n-dc4800d201,Uid:219a36a897fd955b6971f95b188cf014,Namespace:kube-system,Attempt:0,} returns sandbox id \"25a88108ecffdf603f40f59a2f9823be89e98f74a98b61395dadc9e71b93987c\"" Sep 12 10:11:33.331094 kubelet[2221]: E0912 10:11:33.331064 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:33.333959 containerd[1490]: time="2025-09-12T10:11:33.333580242Z" level=info msg="CreateContainer within sandbox \"25a88108ecffdf603f40f59a2f9823be89e98f74a98b61395dadc9e71b93987c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 12 10:11:33.345285 containerd[1490]: time="2025-09-12T10:11:33.345234484Z" level=info msg="CreateContainer within sandbox \"b51fa4472850950ddae44de5286a42453f01b6f0fafc85c6a97d8f5792f8373a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2292b15b699d46abdcee8b0bbc10954f45effd8018ddd43700ee6a102174ba4a\"" Sep 12 10:11:33.346014 containerd[1490]: time="2025-09-12T10:11:33.345978423Z" level=info msg="StartContainer for \"2292b15b699d46abdcee8b0bbc10954f45effd8018ddd43700ee6a102174ba4a\"" Sep 12 10:11:33.352468 containerd[1490]: time="2025-09-12T10:11:33.352419568Z" level=info msg="CreateContainer within sandbox \"25a88108ecffdf603f40f59a2f9823be89e98f74a98b61395dadc9e71b93987c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"490a1a81afc400d1b8f9eea7167c8b326add191ba6e2a9872aca138287413cc6\"" Sep 12 10:11:33.353236 containerd[1490]: time="2025-09-12T10:11:33.353212300Z" level=info msg="StartContainer for \"490a1a81afc400d1b8f9eea7167c8b326add191ba6e2a9872aca138287413cc6\"" Sep 12 10:11:33.368093 systemd[1]: Started cri-containerd-14e1ed7d2616ed1b707976377e47c6ce22a18374c94773498fbcaf13ec749291.scope - libcontainer container 14e1ed7d2616ed1b707976377e47c6ce22a18374c94773498fbcaf13ec749291. Sep 12 10:11:33.392751 systemd[1]: Started cri-containerd-2292b15b699d46abdcee8b0bbc10954f45effd8018ddd43700ee6a102174ba4a.scope - libcontainer container 2292b15b699d46abdcee8b0bbc10954f45effd8018ddd43700ee6a102174ba4a. Sep 12 10:11:33.413707 systemd[1]: Started cri-containerd-490a1a81afc400d1b8f9eea7167c8b326add191ba6e2a9872aca138287413cc6.scope - libcontainer container 490a1a81afc400d1b8f9eea7167c8b326add191ba6e2a9872aca138287413cc6. Sep 12 10:11:33.457302 containerd[1490]: time="2025-09-12T10:11:33.456509818Z" level=info msg="StartContainer for \"14e1ed7d2616ed1b707976377e47c6ce22a18374c94773498fbcaf13ec749291\" returns successfully" Sep 12 10:11:33.491875 containerd[1490]: time="2025-09-12T10:11:33.491525843Z" level=info msg="StartContainer for \"490a1a81afc400d1b8f9eea7167c8b326add191ba6e2a9872aca138287413cc6\" returns successfully" Sep 12 10:11:33.497073 containerd[1490]: time="2025-09-12T10:11:33.496967957Z" level=info msg="StartContainer for \"2292b15b699d46abdcee8b0bbc10954f45effd8018ddd43700ee6a102174ba4a\" returns successfully" Sep 12 10:11:33.548797 kubelet[2221]: E0912 10:11:33.548724 2221 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://164.92.125.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-n-dc4800d201?timeout=10s\": dial tcp 164.92.125.86:6443: connect: connection refused" interval="1.6s" Sep 12 10:11:33.565974 kubelet[2221]: W0912 10:11:33.565883 2221 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://164.92.125.86:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 164.92.125.86:6443: connect: connection refused Sep 12 10:11:33.565974 kubelet[2221]: E0912 10:11:33.565978 2221 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://164.92.125.86:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 164.92.125.86:6443: connect: connection refused" logger="UnhandledError" Sep 12 10:11:33.724089 kubelet[2221]: I0912 10:11:33.723281 2221 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:33.724089 kubelet[2221]: E0912 10:11:33.723705 2221 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://164.92.125.86:6443/api/v1/nodes\": dial tcp 164.92.125.86:6443: connect: connection refused" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:34.188060 kubelet[2221]: E0912 10:11:34.186788 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:34.188060 kubelet[2221]: E0912 10:11:34.186949 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:34.191659 kubelet[2221]: E0912 10:11:34.189020 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:34.191659 kubelet[2221]: E0912 10:11:34.189146 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:34.194648 kubelet[2221]: E0912 10:11:34.194349 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:34.194648 kubelet[2221]: E0912 10:11:34.194494 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:35.200929 kubelet[2221]: E0912 10:11:35.200129 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.200929 kubelet[2221]: E0912 10:11:35.200156 2221 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.200929 kubelet[2221]: E0912 10:11:35.200296 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:35.200929 kubelet[2221]: E0912 10:11:35.200295 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:35.326830 kubelet[2221]: I0912 10:11:35.326792 2221 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.686329 kubelet[2221]: E0912 10:11:35.686272 2221 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4230.2.2-n-dc4800d201\" not found" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.779308 kubelet[2221]: E0912 10:11:35.778445 2221 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230.2.2-n-dc4800d201.18648145f9ecf588 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230.2.2-n-dc4800d201,UID:ci-4230.2.2-n-dc4800d201,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230.2.2-n-dc4800d201,},FirstTimestamp:2025-09-12 10:11:32.127118728 +0000 UTC m=+1.045273254,LastTimestamp:2025-09-12 10:11:32.127118728 +0000 UTC m=+1.045273254,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230.2.2-n-dc4800d201,}" Sep 12 10:11:35.869037 kubelet[2221]: I0912 10:11:35.868976 2221 kubelet_node_status.go:78] "Successfully registered node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.869037 kubelet[2221]: E0912 10:11:35.869037 2221 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4230.2.2-n-dc4800d201\": node \"ci-4230.2.2-n-dc4800d201\" not found" Sep 12 10:11:35.944988 kubelet[2221]: I0912 10:11:35.944285 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.956323 kubelet[2221]: E0912 10:11:35.956274 2221 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4230.2.2-n-dc4800d201\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.956323 kubelet[2221]: I0912 10:11:35.956319 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.964741 kubelet[2221]: E0912 10:11:35.964454 2221 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.964741 kubelet[2221]: I0912 10:11:35.964490 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:35.966747 kubelet[2221]: E0912 10:11:35.966709 2221 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.124173 kubelet[2221]: I0912 10:11:36.123858 2221 apiserver.go:52] "Watching apiserver" Sep 12 10:11:36.142439 kubelet[2221]: I0912 10:11:36.142369 2221 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 12 10:11:36.200019 kubelet[2221]: I0912 10:11:36.199628 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.200019 kubelet[2221]: I0912 10:11:36.199697 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.203573 kubelet[2221]: E0912 10:11:36.203080 2221 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4230.2.2-n-dc4800d201\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.203573 kubelet[2221]: E0912 10:11:36.203329 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:36.204230 kubelet[2221]: E0912 10:11:36.203085 2221 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.204450 kubelet[2221]: E0912 10:11:36.204381 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:36.653176 kubelet[2221]: I0912 10:11:36.652927 2221 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:36.666175 kubelet[2221]: W0912 10:11:36.665474 2221 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 10:11:36.666175 kubelet[2221]: E0912 10:11:36.665821 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:37.201031 kubelet[2221]: E0912 10:11:37.200986 2221 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:38.029301 systemd[1]: Reload requested from client PID 2495 ('systemctl') (unit session-7.scope)... Sep 12 10:11:38.029321 systemd[1]: Reloading... Sep 12 10:11:38.143563 zram_generator::config[2551]: No configuration found. Sep 12 10:11:38.260539 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 12 10:11:38.389708 systemd[1]: Reloading finished in 359 ms. Sep 12 10:11:38.418915 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:38.434400 systemd[1]: kubelet.service: Deactivated successfully. Sep 12 10:11:38.434865 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:38.435037 systemd[1]: kubelet.service: Consumed 1.496s CPU time, 126.7M memory peak. Sep 12 10:11:38.444978 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 10:11:38.587357 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 10:11:38.600152 (kubelet)[2590]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 10:11:38.669723 kubelet[2590]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 10:11:38.669723 kubelet[2590]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 12 10:11:38.669723 kubelet[2590]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 10:11:38.670158 kubelet[2590]: I0912 10:11:38.669750 2590 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 10:11:38.680583 kubelet[2590]: I0912 10:11:38.679256 2590 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Sep 12 10:11:38.680583 kubelet[2590]: I0912 10:11:38.679292 2590 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 10:11:38.680583 kubelet[2590]: I0912 10:11:38.679856 2590 server.go:954] "Client rotation is on, will bootstrap in background" Sep 12 10:11:38.681843 kubelet[2590]: I0912 10:11:38.681803 2590 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 12 10:11:38.687565 kubelet[2590]: I0912 10:11:38.686991 2590 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 10:11:38.689617 kubelet[2590]: E0912 10:11:38.689585 2590 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 12 10:11:38.689750 kubelet[2590]: I0912 10:11:38.689739 2590 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 12 10:11:38.697917 kubelet[2590]: I0912 10:11:38.697876 2590 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 10:11:38.698191 kubelet[2590]: I0912 10:11:38.698152 2590 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 10:11:38.698375 kubelet[2590]: I0912 10:11:38.698192 2590 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.2.2-n-dc4800d201","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 10:11:38.698460 kubelet[2590]: I0912 10:11:38.698387 2590 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 10:11:38.698460 kubelet[2590]: I0912 10:11:38.698398 2590 container_manager_linux.go:304] "Creating device plugin manager" Sep 12 10:11:38.698460 kubelet[2590]: I0912 10:11:38.698448 2590 state_mem.go:36] "Initialized new in-memory state store" Sep 12 10:11:38.698639 kubelet[2590]: I0912 10:11:38.698627 2590 kubelet.go:446] "Attempting to sync node with API server" Sep 12 10:11:38.698675 kubelet[2590]: I0912 10:11:38.698651 2590 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 10:11:38.698675 kubelet[2590]: I0912 10:11:38.698673 2590 kubelet.go:352] "Adding apiserver pod source" Sep 12 10:11:38.698732 kubelet[2590]: I0912 10:11:38.698685 2590 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 10:11:38.702510 kubelet[2590]: I0912 10:11:38.700867 2590 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Sep 12 10:11:38.702510 kubelet[2590]: I0912 10:11:38.701260 2590 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 12 10:11:38.702510 kubelet[2590]: I0912 10:11:38.701683 2590 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 12 10:11:38.702510 kubelet[2590]: I0912 10:11:38.701708 2590 server.go:1287] "Started kubelet" Sep 12 10:11:38.706151 kubelet[2590]: I0912 10:11:38.706045 2590 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 10:11:38.708279 kubelet[2590]: I0912 10:11:38.708223 2590 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 10:11:38.713475 kubelet[2590]: I0912 10:11:38.712508 2590 server.go:479] "Adding debug handlers to kubelet server" Sep 12 10:11:38.714654 kubelet[2590]: I0912 10:11:38.714579 2590 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 10:11:38.714874 kubelet[2590]: I0912 10:11:38.714859 2590 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 10:11:38.715122 kubelet[2590]: I0912 10:11:38.715105 2590 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 10:11:38.718129 kubelet[2590]: I0912 10:11:38.718103 2590 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 12 10:11:38.718407 kubelet[2590]: E0912 10:11:38.718383 2590 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4230.2.2-n-dc4800d201\" not found" Sep 12 10:11:38.719830 kubelet[2590]: I0912 10:11:38.719806 2590 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 12 10:11:38.720005 kubelet[2590]: I0912 10:11:38.719992 2590 reconciler.go:26] "Reconciler: start to sync state" Sep 12 10:11:38.735566 kubelet[2590]: I0912 10:11:38.734632 2590 factory.go:221] Registration of the systemd container factory successfully Sep 12 10:11:38.736691 kubelet[2590]: I0912 10:11:38.735890 2590 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 10:11:38.743565 kubelet[2590]: I0912 10:11:38.743505 2590 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 12 10:11:38.746788 kubelet[2590]: E0912 10:11:38.746748 2590 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 10:11:38.748439 kubelet[2590]: I0912 10:11:38.747732 2590 factory.go:221] Registration of the containerd container factory successfully Sep 12 10:11:38.748439 kubelet[2590]: I0912 10:11:38.747865 2590 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 12 10:11:38.748439 kubelet[2590]: I0912 10:11:38.747927 2590 status_manager.go:227] "Starting to sync pod status with apiserver" Sep 12 10:11:38.748439 kubelet[2590]: I0912 10:11:38.747951 2590 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 12 10:11:38.748439 kubelet[2590]: I0912 10:11:38.747959 2590 kubelet.go:2382] "Starting kubelet main sync loop" Sep 12 10:11:38.748439 kubelet[2590]: E0912 10:11:38.748016 2590 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 10:11:38.811245 kubelet[2590]: I0912 10:11:38.811218 2590 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811418 2590 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811442 2590 state_mem.go:36] "Initialized new in-memory state store" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811703 2590 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811714 2590 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811732 2590 policy_none.go:49] "None policy: Start" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811760 2590 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811774 2590 state_mem.go:35] "Initializing new in-memory state store" Sep 12 10:11:38.812728 kubelet[2590]: I0912 10:11:38.811894 2590 state_mem.go:75] "Updated machine memory state" Sep 12 10:11:38.816988 kubelet[2590]: I0912 10:11:38.816343 2590 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 12 10:11:38.816988 kubelet[2590]: I0912 10:11:38.816578 2590 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 10:11:38.816988 kubelet[2590]: I0912 10:11:38.816598 2590 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 10:11:38.818184 kubelet[2590]: I0912 10:11:38.817938 2590 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 10:11:38.818969 kubelet[2590]: E0912 10:11:38.818837 2590 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 12 10:11:38.850875 kubelet[2590]: I0912 10:11:38.850467 2590 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.850875 kubelet[2590]: I0912 10:11:38.850657 2590 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.850875 kubelet[2590]: I0912 10:11:38.850470 2590 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.859068 kubelet[2590]: W0912 10:11:38.858875 2590 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 10:11:38.859676 kubelet[2590]: W0912 10:11:38.859562 2590 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 10:11:38.859676 kubelet[2590]: E0912 10:11:38.859612 2590 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" already exists" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.859945 kubelet[2590]: W0912 10:11:38.859847 2590 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 10:11:38.918573 kubelet[2590]: I0912 10:11:38.917934 2590 kubelet_node_status.go:75] "Attempting to register node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924174 kubelet[2590]: I0912 10:11:38.921421 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924174 kubelet[2590]: I0912 10:11:38.921457 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-k8s-certs\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924174 kubelet[2590]: I0912 10:11:38.921475 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-k8s-certs\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924174 kubelet[2590]: I0912 10:11:38.921958 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-ca-certs\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924174 kubelet[2590]: I0912 10:11:38.921979 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/724ea47755c3eb2d655fa4e278e349e1-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.2.2-n-dc4800d201\" (UID: \"724ea47755c3eb2d655fa4e278e349e1\") " pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924556 kubelet[2590]: I0912 10:11:38.921998 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-ca-certs\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924556 kubelet[2590]: I0912 10:11:38.922014 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-kubeconfig\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924556 kubelet[2590]: I0912 10:11:38.922030 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/87e38bff5b8b029a6f3c49d3080af986-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.2.2-n-dc4800d201\" (UID: \"87e38bff5b8b029a6f3c49d3080af986\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.924556 kubelet[2590]: I0912 10:11:38.922050 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/219a36a897fd955b6971f95b188cf014-kubeconfig\") pod \"kube-scheduler-ci-4230.2.2-n-dc4800d201\" (UID: \"219a36a897fd955b6971f95b188cf014\") " pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.927989 kubelet[2590]: I0912 10:11:38.927928 2590 kubelet_node_status.go:124] "Node was previously registered" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:38.928168 kubelet[2590]: I0912 10:11:38.928015 2590 kubelet_node_status.go:78] "Successfully registered node" node="ci-4230.2.2-n-dc4800d201" Sep 12 10:11:39.042278 sudo[2621]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 12 10:11:39.043516 sudo[2621]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 12 10:11:39.160071 kubelet[2590]: E0912 10:11:39.160018 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.160233 kubelet[2590]: E0912 10:11:39.160208 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.160843 kubelet[2590]: E0912 10:11:39.160801 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.655279 sudo[2621]: pam_unix(sudo:session): session closed for user root Sep 12 10:11:39.699952 kubelet[2590]: I0912 10:11:39.699426 2590 apiserver.go:52] "Watching apiserver" Sep 12 10:11:39.721074 kubelet[2590]: I0912 10:11:39.720780 2590 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 12 10:11:39.782440 kubelet[2590]: E0912 10:11:39.782381 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.783056 kubelet[2590]: I0912 10:11:39.782952 2590 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:39.783491 kubelet[2590]: E0912 10:11:39.783466 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.805454 kubelet[2590]: W0912 10:11:39.805406 2590 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 12 10:11:39.805659 kubelet[2590]: E0912 10:11:39.805485 2590 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4230.2.2-n-dc4800d201\" already exists" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" Sep 12 10:11:39.805701 kubelet[2590]: E0912 10:11:39.805679 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:39.835475 kubelet[2590]: I0912 10:11:39.835412 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4230.2.2-n-dc4800d201" podStartSLOduration=1.835393953 podStartE2EDuration="1.835393953s" podCreationTimestamp="2025-09-12 10:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:11:39.825838193 +0000 UTC m=+1.218579313" watchObservedRunningTime="2025-09-12 10:11:39.835393953 +0000 UTC m=+1.228135070" Sep 12 10:11:39.845929 kubelet[2590]: I0912 10:11:39.845873 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4230.2.2-n-dc4800d201" podStartSLOduration=3.8458538410000003 podStartE2EDuration="3.845853841s" podCreationTimestamp="2025-09-12 10:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:11:39.836041819 +0000 UTC m=+1.228782945" watchObservedRunningTime="2025-09-12 10:11:39.845853841 +0000 UTC m=+1.238594968" Sep 12 10:11:39.857044 kubelet[2590]: I0912 10:11:39.856824 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4230.2.2-n-dc4800d201" podStartSLOduration=1.856802419 podStartE2EDuration="1.856802419s" podCreationTimestamp="2025-09-12 10:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:11:39.846013252 +0000 UTC m=+1.238754348" watchObservedRunningTime="2025-09-12 10:11:39.856802419 +0000 UTC m=+1.249543528" Sep 12 10:11:40.783359 kubelet[2590]: E0912 10:11:40.782934 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:40.785743 kubelet[2590]: E0912 10:11:40.785129 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:41.199153 sudo[1680]: pam_unix(sudo:session): session closed for user root Sep 12 10:11:41.203622 sshd[1679]: Connection closed by 139.178.68.195 port 45584 Sep 12 10:11:41.204672 sshd-session[1676]: pam_unix(sshd:session): session closed for user core Sep 12 10:11:41.210490 systemd[1]: sshd@6-164.92.125.86:22-139.178.68.195:45584.service: Deactivated successfully. Sep 12 10:11:41.216480 systemd[1]: session-7.scope: Deactivated successfully. Sep 12 10:11:41.217015 systemd[1]: session-7.scope: Consumed 4.598s CPU time, 219.1M memory peak. Sep 12 10:11:41.220415 systemd-logind[1466]: Session 7 logged out. Waiting for processes to exit. Sep 12 10:11:41.222001 systemd-logind[1466]: Removed session 7. Sep 12 10:11:44.285267 kubelet[2590]: I0912 10:11:44.285235 2590 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 12 10:11:44.286211 containerd[1490]: time="2025-09-12T10:11:44.286168352Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 12 10:11:44.286606 kubelet[2590]: I0912 10:11:44.286470 2590 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 12 10:11:44.453457 kubelet[2590]: E0912 10:11:44.452610 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:44.790717 kubelet[2590]: E0912 10:11:44.790681 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.093329 kubelet[2590]: I0912 10:11:45.093171 2590 status_manager.go:890] "Failed to get status for pod" podUID="4998d8e1-cf9d-450a-8a92-766e938ce703" pod="kube-system/kube-proxy-p577b" err="pods \"kube-proxy-p577b\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" Sep 12 10:11:45.106153 systemd[1]: Created slice kubepods-besteffort-pod4998d8e1_cf9d_450a_8a92_766e938ce703.slice - libcontainer container kubepods-besteffort-pod4998d8e1_cf9d_450a_8a92_766e938ce703.slice. Sep 12 10:11:45.134844 systemd[1]: Created slice kubepods-burstable-podb636ea9b_fac2_43d4_abde_838ed8abdb63.slice - libcontainer container kubepods-burstable-podb636ea9b_fac2_43d4_abde_838ed8abdb63.slice. Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161132 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-run\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161178 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nzq\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-kube-api-access-98nzq\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161200 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4998d8e1-cf9d-450a-8a92-766e938ce703-lib-modules\") pod \"kube-proxy-p577b\" (UID: \"4998d8e1-cf9d-450a-8a92-766e938ce703\") " pod="kube-system/kube-proxy-p577b" Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161218 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4998d8e1-cf9d-450a-8a92-766e938ce703-xtables-lock\") pod \"kube-proxy-p577b\" (UID: \"4998d8e1-cf9d-450a-8a92-766e938ce703\") " pod="kube-system/kube-proxy-p577b" Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161232 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cni-path\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161651 kubelet[2590]: I0912 10:11:45.161250 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-kernel\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161266 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-cgroup\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161282 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b636ea9b-fac2-43d4-abde-838ed8abdb63-clustermesh-secrets\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161298 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-net\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161347 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-hubble-tls\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161376 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb6h5\" (UniqueName: \"kubernetes.io/projected/4998d8e1-cf9d-450a-8a92-766e938ce703-kube-api-access-qb6h5\") pod \"kube-proxy-p577b\" (UID: \"4998d8e1-cf9d-450a-8a92-766e938ce703\") " pod="kube-system/kube-proxy-p577b" Sep 12 10:11:45.161953 kubelet[2590]: I0912 10:11:45.161406 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-bpf-maps\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161423 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-hostproc\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161440 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-xtables-lock\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161455 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/4998d8e1-cf9d-450a-8a92-766e938ce703-kube-proxy\") pod \"kube-proxy-p577b\" (UID: \"4998d8e1-cf9d-450a-8a92-766e938ce703\") " pod="kube-system/kube-proxy-p577b" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161470 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-config-path\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161485 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-etc-cni-netd\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.162103 kubelet[2590]: I0912 10:11:45.161502 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-lib-modules\") pod \"cilium-s88lt\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " pod="kube-system/cilium-s88lt" Sep 12 10:11:45.424087 systemd[1]: Created slice kubepods-besteffort-pod1bc90cd3_fbf9_4494_8740_a14bf04b19ca.slice - libcontainer container kubepods-besteffort-pod1bc90cd3_fbf9_4494_8740_a14bf04b19ca.slice. Sep 12 10:11:45.428098 kubelet[2590]: E0912 10:11:45.426894 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.428828 containerd[1490]: time="2025-09-12T10:11:45.428744628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-p577b,Uid:4998d8e1-cf9d-450a-8a92-766e938ce703,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:45.445288 kubelet[2590]: E0912 10:11:45.444058 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.445442 containerd[1490]: time="2025-09-12T10:11:45.444838999Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s88lt,Uid:b636ea9b-fac2-43d4-abde-838ed8abdb63,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:45.463287 kubelet[2590]: I0912 10:11:45.463242 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c4cm\" (UniqueName: \"kubernetes.io/projected/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-kube-api-access-8c4cm\") pod \"cilium-operator-6c4d7847fc-r8j2g\" (UID: \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\") " pod="kube-system/cilium-operator-6c4d7847fc-r8j2g" Sep 12 10:11:45.463626 kubelet[2590]: I0912 10:11:45.463600 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-r8j2g\" (UID: \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\") " pod="kube-system/cilium-operator-6c4d7847fc-r8j2g" Sep 12 10:11:45.472517 containerd[1490]: time="2025-09-12T10:11:45.472165201Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:45.472517 containerd[1490]: time="2025-09-12T10:11:45.472280052Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:45.472517 containerd[1490]: time="2025-09-12T10:11:45.472293527Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.474935 containerd[1490]: time="2025-09-12T10:11:45.473689852Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.475349 containerd[1490]: time="2025-09-12T10:11:45.475008934Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:45.475349 containerd[1490]: time="2025-09-12T10:11:45.475089104Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:45.475349 containerd[1490]: time="2025-09-12T10:11:45.475102096Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.475349 containerd[1490]: time="2025-09-12T10:11:45.475212064Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.506841 systemd[1]: Started cri-containerd-0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9.scope - libcontainer container 0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9. Sep 12 10:11:45.509874 systemd[1]: Started cri-containerd-53993b2d3c20411e69bcdf9f87748e718e1c5c0b87ce8ce8f04b72c4f7db1fd0.scope - libcontainer container 53993b2d3c20411e69bcdf9f87748e718e1c5c0b87ce8ce8f04b72c4f7db1fd0. Sep 12 10:11:45.553190 containerd[1490]: time="2025-09-12T10:11:45.553134450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s88lt,Uid:b636ea9b-fac2-43d4-abde-838ed8abdb63,Namespace:kube-system,Attempt:0,} returns sandbox id \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\"" Sep 12 10:11:45.554891 kubelet[2590]: E0912 10:11:45.554861 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.559777 containerd[1490]: time="2025-09-12T10:11:45.558148179Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 12 10:11:45.568802 containerd[1490]: time="2025-09-12T10:11:45.568739471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-p577b,Uid:4998d8e1-cf9d-450a-8a92-766e938ce703,Namespace:kube-system,Attempt:0,} returns sandbox id \"53993b2d3c20411e69bcdf9f87748e718e1c5c0b87ce8ce8f04b72c4f7db1fd0\"" Sep 12 10:11:45.571912 kubelet[2590]: E0912 10:11:45.571777 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.579173 containerd[1490]: time="2025-09-12T10:11:45.579125579Z" level=info msg="CreateContainer within sandbox \"53993b2d3c20411e69bcdf9f87748e718e1c5c0b87ce8ce8f04b72c4f7db1fd0\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 12 10:11:45.601872 containerd[1490]: time="2025-09-12T10:11:45.601809879Z" level=info msg="CreateContainer within sandbox \"53993b2d3c20411e69bcdf9f87748e718e1c5c0b87ce8ce8f04b72c4f7db1fd0\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d394489ee48ed778f68353b5df1414173f8b032f7c0f2668bae576d6854179ef\"" Sep 12 10:11:45.604661 containerd[1490]: time="2025-09-12T10:11:45.604615190Z" level=info msg="StartContainer for \"d394489ee48ed778f68353b5df1414173f8b032f7c0f2668bae576d6854179ef\"" Sep 12 10:11:45.636432 systemd[1]: Started cri-containerd-d394489ee48ed778f68353b5df1414173f8b032f7c0f2668bae576d6854179ef.scope - libcontainer container d394489ee48ed778f68353b5df1414173f8b032f7c0f2668bae576d6854179ef. Sep 12 10:11:45.672246 containerd[1490]: time="2025-09-12T10:11:45.672163132Z" level=info msg="StartContainer for \"d394489ee48ed778f68353b5df1414173f8b032f7c0f2668bae576d6854179ef\" returns successfully" Sep 12 10:11:45.732993 kubelet[2590]: E0912 10:11:45.732860 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.734422 containerd[1490]: time="2025-09-12T10:11:45.734366989Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-r8j2g,Uid:1bc90cd3-fbf9-4494-8740-a14bf04b19ca,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:45.762321 containerd[1490]: time="2025-09-12T10:11:45.761715433Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:11:45.762321 containerd[1490]: time="2025-09-12T10:11:45.761801561Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:11:45.762321 containerd[1490]: time="2025-09-12T10:11:45.761820391Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.762991 containerd[1490]: time="2025-09-12T10:11:45.762864503Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:11:45.791603 systemd[1]: Started cri-containerd-4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353.scope - libcontainer container 4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353. Sep 12 10:11:45.798770 kubelet[2590]: E0912 10:11:45.798006 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:45.866779 containerd[1490]: time="2025-09-12T10:11:45.866626270Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-r8j2g,Uid:1bc90cd3-fbf9-4494-8740-a14bf04b19ca,Namespace:kube-system,Attempt:0,} returns sandbox id \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\"" Sep 12 10:11:45.869268 kubelet[2590]: E0912 10:11:45.869243 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:46.759639 kubelet[2590]: E0912 10:11:46.758276 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:46.781095 kubelet[2590]: I0912 10:11:46.780972 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-p577b" podStartSLOduration=1.78094412 podStartE2EDuration="1.78094412s" podCreationTimestamp="2025-09-12 10:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:11:45.815444659 +0000 UTC m=+7.208185778" watchObservedRunningTime="2025-09-12 10:11:46.78094412 +0000 UTC m=+8.173685254" Sep 12 10:11:46.802695 kubelet[2590]: E0912 10:11:46.801638 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:47.065472 systemd-timesyncd[1380]: Contacted time server 23.142.248.9:123 (2.flatcar.pool.ntp.org). Sep 12 10:11:47.065991 systemd-timesyncd[1380]: Initial clock synchronization to Fri 2025-09-12 10:11:47.430188 UTC. Sep 12 10:11:47.815738 kubelet[2590]: E0912 10:11:47.815311 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:48.275407 kubelet[2590]: E0912 10:11:48.275351 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:48.820213 kubelet[2590]: E0912 10:11:48.819157 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:49.821488 kubelet[2590]: E0912 10:11:49.821455 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:50.628991 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3552940215.mount: Deactivated successfully. Sep 12 10:11:52.905407 containerd[1490]: time="2025-09-12T10:11:52.905344975Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:52.906187 containerd[1490]: time="2025-09-12T10:11:52.905662382Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Sep 12 10:11:52.907569 containerd[1490]: time="2025-09-12T10:11:52.907286751Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:52.910154 containerd[1490]: time="2025-09-12T10:11:52.910018722Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 7.35181745s" Sep 12 10:11:52.910154 containerd[1490]: time="2025-09-12T10:11:52.910090723Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 12 10:11:52.912839 containerd[1490]: time="2025-09-12T10:11:52.912798496Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 12 10:11:52.917697 containerd[1490]: time="2025-09-12T10:11:52.917445929Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 12 10:11:53.026623 containerd[1490]: time="2025-09-12T10:11:53.026570493Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\"" Sep 12 10:11:53.028131 containerd[1490]: time="2025-09-12T10:11:53.027984507Z" level=info msg="StartContainer for \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\"" Sep 12 10:11:53.136233 systemd[1]: run-containerd-runc-k8s.io-4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c-runc.KU9IAW.mount: Deactivated successfully. Sep 12 10:11:53.145802 systemd[1]: Started cri-containerd-4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c.scope - libcontainer container 4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c. Sep 12 10:11:53.191403 containerd[1490]: time="2025-09-12T10:11:53.191238798Z" level=info msg="StartContainer for \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\" returns successfully" Sep 12 10:11:53.207258 systemd[1]: cri-containerd-4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c.scope: Deactivated successfully. Sep 12 10:11:53.318748 containerd[1490]: time="2025-09-12T10:11:53.304959708Z" level=info msg="shim disconnected" id=4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c namespace=k8s.io Sep 12 10:11:53.318748 containerd[1490]: time="2025-09-12T10:11:53.318727522Z" level=warning msg="cleaning up after shim disconnected" id=4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c namespace=k8s.io Sep 12 10:11:53.318748 containerd[1490]: time="2025-09-12T10:11:53.318745084Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:11:53.338315 containerd[1490]: time="2025-09-12T10:11:53.338057008Z" level=warning msg="cleanup warnings time=\"2025-09-12T10:11:53Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 12 10:11:53.849832 kubelet[2590]: E0912 10:11:53.849791 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:53.880375 containerd[1490]: time="2025-09-12T10:11:53.880304778Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 12 10:11:53.902000 containerd[1490]: time="2025-09-12T10:11:53.901940490Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\"" Sep 12 10:11:53.902844 containerd[1490]: time="2025-09-12T10:11:53.902783330Z" level=info msg="StartContainer for \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\"" Sep 12 10:11:53.938865 systemd[1]: Started cri-containerd-aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e.scope - libcontainer container aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e. Sep 12 10:11:53.981393 containerd[1490]: time="2025-09-12T10:11:53.981348059Z" level=info msg="StartContainer for \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\" returns successfully" Sep 12 10:11:53.990437 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c-rootfs.mount: Deactivated successfully. Sep 12 10:11:54.005569 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 12 10:11:54.006261 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 12 10:11:54.006773 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 12 10:11:54.016267 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 10:11:54.020830 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 12 10:11:54.021465 systemd[1]: cri-containerd-aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e.scope: Deactivated successfully. Sep 12 10:11:54.061447 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e-rootfs.mount: Deactivated successfully. Sep 12 10:11:54.064850 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 10:11:54.070245 containerd[1490]: time="2025-09-12T10:11:54.070006785Z" level=info msg="shim disconnected" id=aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e namespace=k8s.io Sep 12 10:11:54.070245 containerd[1490]: time="2025-09-12T10:11:54.070067249Z" level=warning msg="cleaning up after shim disconnected" id=aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e namespace=k8s.io Sep 12 10:11:54.070245 containerd[1490]: time="2025-09-12T10:11:54.070075871Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:11:54.714903 update_engine[1467]: I20250912 10:11:54.714795 1467 update_attempter.cc:509] Updating boot flags... Sep 12 10:11:54.773700 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (3127) Sep 12 10:11:54.848755 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (3127) Sep 12 10:11:54.869116 kubelet[2590]: E0912 10:11:54.868341 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:54.907014 containerd[1490]: time="2025-09-12T10:11:54.904688129Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 12 10:11:54.974851 containerd[1490]: time="2025-09-12T10:11:54.974299909Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\"" Sep 12 10:11:54.977304 containerd[1490]: time="2025-09-12T10:11:54.977168182Z" level=info msg="StartContainer for \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\"" Sep 12 10:11:54.989275 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount298214456.mount: Deactivated successfully. Sep 12 10:11:55.041322 systemd[1]: run-containerd-runc-k8s.io-afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf-runc.P3h7SO.mount: Deactivated successfully. Sep 12 10:11:55.050870 systemd[1]: Started cri-containerd-afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf.scope - libcontainer container afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf. Sep 12 10:11:55.103607 containerd[1490]: time="2025-09-12T10:11:55.103254277Z" level=info msg="StartContainer for \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\" returns successfully" Sep 12 10:11:55.105743 containerd[1490]: time="2025-09-12T10:11:55.103733214Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Sep 12 10:11:55.105989 containerd[1490]: time="2025-09-12T10:11:55.105081818Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:55.108110 containerd[1490]: time="2025-09-12T10:11:55.108068810Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 10:11:55.108994 containerd[1490]: time="2025-09-12T10:11:55.108824499Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.19597454s" Sep 12 10:11:55.108994 containerd[1490]: time="2025-09-12T10:11:55.108859072Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 12 10:11:55.116606 containerd[1490]: time="2025-09-12T10:11:55.116031650Z" level=info msg="CreateContainer within sandbox \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 12 10:11:55.117197 systemd[1]: cri-containerd-afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf.scope: Deactivated successfully. Sep 12 10:11:55.163166 containerd[1490]: time="2025-09-12T10:11:55.162668628Z" level=info msg="shim disconnected" id=afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf namespace=k8s.io Sep 12 10:11:55.163166 containerd[1490]: time="2025-09-12T10:11:55.162753145Z" level=warning msg="cleaning up after shim disconnected" id=afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf namespace=k8s.io Sep 12 10:11:55.163166 containerd[1490]: time="2025-09-12T10:11:55.162766633Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:11:55.164983 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf-rootfs.mount: Deactivated successfully. Sep 12 10:11:55.170066 containerd[1490]: time="2025-09-12T10:11:55.170018930Z" level=info msg="CreateContainer within sandbox \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\"" Sep 12 10:11:55.172450 containerd[1490]: time="2025-09-12T10:11:55.172415979Z" level=info msg="StartContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\"" Sep 12 10:11:55.249878 systemd[1]: Started cri-containerd-7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f.scope - libcontainer container 7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f. Sep 12 10:11:55.291139 containerd[1490]: time="2025-09-12T10:11:55.291047209Z" level=info msg="StartContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" returns successfully" Sep 12 10:11:55.910497 kubelet[2590]: E0912 10:11:55.910447 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:55.915401 containerd[1490]: time="2025-09-12T10:11:55.915359946Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 12 10:11:55.918202 kubelet[2590]: E0912 10:11:55.917867 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:55.929830 containerd[1490]: time="2025-09-12T10:11:55.929590508Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\"" Sep 12 10:11:55.930249 containerd[1490]: time="2025-09-12T10:11:55.930201125Z" level=info msg="StartContainer for \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\"" Sep 12 10:11:56.022472 systemd[1]: run-containerd-runc-k8s.io-97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b-runc.aImPrI.mount: Deactivated successfully. Sep 12 10:11:56.034786 systemd[1]: Started cri-containerd-97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b.scope - libcontainer container 97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b. Sep 12 10:11:56.110821 systemd[1]: cri-containerd-97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b.scope: Deactivated successfully. Sep 12 10:11:56.114348 containerd[1490]: time="2025-09-12T10:11:56.112701508Z" level=info msg="StartContainer for \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\" returns successfully" Sep 12 10:11:56.164707 containerd[1490]: time="2025-09-12T10:11:56.162731581Z" level=info msg="shim disconnected" id=97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b namespace=k8s.io Sep 12 10:11:56.164707 containerd[1490]: time="2025-09-12T10:11:56.162802938Z" level=warning msg="cleaning up after shim disconnected" id=97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b namespace=k8s.io Sep 12 10:11:56.164707 containerd[1490]: time="2025-09-12T10:11:56.162818953Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:11:56.163429 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b-rootfs.mount: Deactivated successfully. Sep 12 10:11:56.196843 containerd[1490]: time="2025-09-12T10:11:56.196780790Z" level=warning msg="cleanup warnings time=\"2025-09-12T10:11:56Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 12 10:11:56.923059 kubelet[2590]: E0912 10:11:56.922744 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:56.923059 kubelet[2590]: E0912 10:11:56.922799 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:56.927584 containerd[1490]: time="2025-09-12T10:11:56.925969975Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 12 10:11:56.940904 containerd[1490]: time="2025-09-12T10:11:56.940763074Z" level=info msg="CreateContainer within sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\"" Sep 12 10:11:56.946997 containerd[1490]: time="2025-09-12T10:11:56.946680737Z" level=info msg="StartContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\"" Sep 12 10:11:56.963415 kubelet[2590]: I0912 10:11:56.962366 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-r8j2g" podStartSLOduration=2.7211297610000003 podStartE2EDuration="11.96233345s" podCreationTimestamp="2025-09-12 10:11:45 +0000 UTC" firstStartedPulling="2025-09-12 10:11:45.870642095 +0000 UTC m=+7.263383195" lastFinishedPulling="2025-09-12 10:11:55.111845784 +0000 UTC m=+16.504586884" observedRunningTime="2025-09-12 10:11:56.244589643 +0000 UTC m=+17.637330755" watchObservedRunningTime="2025-09-12 10:11:56.96233345 +0000 UTC m=+18.355074571" Sep 12 10:11:56.988811 systemd[1]: Started cri-containerd-e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83.scope - libcontainer container e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83. Sep 12 10:11:57.030897 containerd[1490]: time="2025-09-12T10:11:57.030837713Z" level=info msg="StartContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" returns successfully" Sep 12 10:11:57.218730 kubelet[2590]: I0912 10:11:57.218111 2590 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 12 10:11:57.266485 systemd[1]: Created slice kubepods-burstable-pod8dffd854_3556_4ce2_94b0_eb318a35c5e8.slice - libcontainer container kubepods-burstable-pod8dffd854_3556_4ce2_94b0_eb318a35c5e8.slice. Sep 12 10:11:57.277779 systemd[1]: Created slice kubepods-burstable-pod25ae4374_754c_4a3c_a7c5_b18eefd707e4.slice - libcontainer container kubepods-burstable-pod25ae4374_754c_4a3c_a7c5_b18eefd707e4.slice. Sep 12 10:11:57.352511 kubelet[2590]: I0912 10:11:57.352324 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dffd854-3556-4ce2-94b0-eb318a35c5e8-config-volume\") pod \"coredns-668d6bf9bc-65q6r\" (UID: \"8dffd854-3556-4ce2-94b0-eb318a35c5e8\") " pod="kube-system/coredns-668d6bf9bc-65q6r" Sep 12 10:11:57.352511 kubelet[2590]: I0912 10:11:57.352391 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhzwx\" (UniqueName: \"kubernetes.io/projected/25ae4374-754c-4a3c-a7c5-b18eefd707e4-kube-api-access-mhzwx\") pod \"coredns-668d6bf9bc-5l2kb\" (UID: \"25ae4374-754c-4a3c-a7c5-b18eefd707e4\") " pod="kube-system/coredns-668d6bf9bc-5l2kb" Sep 12 10:11:57.352511 kubelet[2590]: I0912 10:11:57.352427 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae4374-754c-4a3c-a7c5-b18eefd707e4-config-volume\") pod \"coredns-668d6bf9bc-5l2kb\" (UID: \"25ae4374-754c-4a3c-a7c5-b18eefd707e4\") " pod="kube-system/coredns-668d6bf9bc-5l2kb" Sep 12 10:11:57.352511 kubelet[2590]: I0912 10:11:57.352452 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdtdl\" (UniqueName: \"kubernetes.io/projected/8dffd854-3556-4ce2-94b0-eb318a35c5e8-kube-api-access-pdtdl\") pod \"coredns-668d6bf9bc-65q6r\" (UID: \"8dffd854-3556-4ce2-94b0-eb318a35c5e8\") " pod="kube-system/coredns-668d6bf9bc-65q6r" Sep 12 10:11:57.574151 kubelet[2590]: E0912 10:11:57.573496 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:57.575164 containerd[1490]: time="2025-09-12T10:11:57.574847588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-65q6r,Uid:8dffd854-3556-4ce2-94b0-eb318a35c5e8,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:57.580965 kubelet[2590]: E0912 10:11:57.580920 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:57.582668 containerd[1490]: time="2025-09-12T10:11:57.581512677Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5l2kb,Uid:25ae4374-754c-4a3c-a7c5-b18eefd707e4,Namespace:kube-system,Attempt:0,}" Sep 12 10:11:57.939626 kubelet[2590]: E0912 10:11:57.937873 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:57.959580 kubelet[2590]: I0912 10:11:57.958563 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-s88lt" podStartSLOduration=5.602756365 podStartE2EDuration="12.958523723s" podCreationTimestamp="2025-09-12 10:11:45 +0000 UTC" firstStartedPulling="2025-09-12 10:11:45.556675649 +0000 UTC m=+6.949416759" lastFinishedPulling="2025-09-12 10:11:52.912443002 +0000 UTC m=+14.305184117" observedRunningTime="2025-09-12 10:11:57.958186116 +0000 UTC m=+19.350927262" watchObservedRunningTime="2025-09-12 10:11:57.958523723 +0000 UTC m=+19.351264860" Sep 12 10:11:58.936864 kubelet[2590]: E0912 10:11:58.936779 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:59.510608 systemd-networkd[1378]: cilium_host: Link UP Sep 12 10:11:59.512072 systemd-networkd[1378]: cilium_net: Link UP Sep 12 10:11:59.512244 systemd-networkd[1378]: cilium_net: Gained carrier Sep 12 10:11:59.512448 systemd-networkd[1378]: cilium_host: Gained carrier Sep 12 10:11:59.653086 systemd-networkd[1378]: cilium_vxlan: Link UP Sep 12 10:11:59.653244 systemd-networkd[1378]: cilium_vxlan: Gained carrier Sep 12 10:11:59.940651 kubelet[2590]: E0912 10:11:59.940451 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:11:59.982254 systemd-networkd[1378]: cilium_net: Gained IPv6LL Sep 12 10:12:00.035781 kernel: NET: Registered PF_ALG protocol family Sep 12 10:12:00.213881 systemd-networkd[1378]: cilium_host: Gained IPv6LL Sep 12 10:12:01.002510 systemd-networkd[1378]: lxc_health: Link UP Sep 12 10:12:01.021485 systemd-networkd[1378]: lxc_health: Gained carrier Sep 12 10:12:01.045752 systemd-networkd[1378]: cilium_vxlan: Gained IPv6LL Sep 12 10:12:01.191609 kernel: eth0: renamed from tmp7066e Sep 12 10:12:01.201154 systemd-networkd[1378]: lxcee79af63166e: Link UP Sep 12 10:12:01.201483 systemd-networkd[1378]: lxcee79af63166e: Gained carrier Sep 12 10:12:01.235637 kernel: eth0: renamed from tmpc731c Sep 12 10:12:01.237323 systemd-networkd[1378]: lxc418aed0e834a: Link UP Sep 12 10:12:01.245088 systemd-networkd[1378]: lxc418aed0e834a: Gained carrier Sep 12 10:12:01.447060 kubelet[2590]: E0912 10:12:01.447023 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:02.325970 systemd-networkd[1378]: lxc418aed0e834a: Gained IPv6LL Sep 12 10:12:02.582295 systemd-networkd[1378]: lxc_health: Gained IPv6LL Sep 12 10:12:02.965733 systemd-networkd[1378]: lxcee79af63166e: Gained IPv6LL Sep 12 10:12:05.853855 containerd[1490]: time="2025-09-12T10:12:05.852600147Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:12:05.853855 containerd[1490]: time="2025-09-12T10:12:05.853404852Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:12:05.853855 containerd[1490]: time="2025-09-12T10:12:05.853427673Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:12:05.853855 containerd[1490]: time="2025-09-12T10:12:05.853577074Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:12:05.907261 systemd[1]: Started cri-containerd-c731ce9c2af9a2d6a720a7ce570421cbd371dedf2489b15098ad149334c76676.scope - libcontainer container c731ce9c2af9a2d6a720a7ce570421cbd371dedf2489b15098ad149334c76676. Sep 12 10:12:05.942837 containerd[1490]: time="2025-09-12T10:12:05.942405434Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:12:05.942837 containerd[1490]: time="2025-09-12T10:12:05.942487570Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:12:05.942837 containerd[1490]: time="2025-09-12T10:12:05.942505281Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:12:05.942837 containerd[1490]: time="2025-09-12T10:12:05.942657270Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:12:05.977939 systemd[1]: Started cri-containerd-7066ef0351fa6aace8f6fce6d6415507d2c96fad9f94e8126f68321bb636bbf2.scope - libcontainer container 7066ef0351fa6aace8f6fce6d6415507d2c96fad9f94e8126f68321bb636bbf2. Sep 12 10:12:06.015873 containerd[1490]: time="2025-09-12T10:12:06.015821323Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-65q6r,Uid:8dffd854-3556-4ce2-94b0-eb318a35c5e8,Namespace:kube-system,Attempt:0,} returns sandbox id \"c731ce9c2af9a2d6a720a7ce570421cbd371dedf2489b15098ad149334c76676\"" Sep 12 10:12:06.019210 kubelet[2590]: E0912 10:12:06.018042 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:06.027753 containerd[1490]: time="2025-09-12T10:12:06.027248021Z" level=info msg="CreateContainer within sandbox \"c731ce9c2af9a2d6a720a7ce570421cbd371dedf2489b15098ad149334c76676\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 10:12:06.051146 containerd[1490]: time="2025-09-12T10:12:06.050989221Z" level=info msg="CreateContainer within sandbox \"c731ce9c2af9a2d6a720a7ce570421cbd371dedf2489b15098ad149334c76676\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ebce6889aa87f25e36b67579d585e0d8177c21dacb0216215c92b442f0c93320\"" Sep 12 10:12:06.052885 containerd[1490]: time="2025-09-12T10:12:06.052699375Z" level=info msg="StartContainer for \"ebce6889aa87f25e36b67579d585e0d8177c21dacb0216215c92b442f0c93320\"" Sep 12 10:12:06.099126 containerd[1490]: time="2025-09-12T10:12:06.098592192Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-5l2kb,Uid:25ae4374-754c-4a3c-a7c5-b18eefd707e4,Namespace:kube-system,Attempt:0,} returns sandbox id \"7066ef0351fa6aace8f6fce6d6415507d2c96fad9f94e8126f68321bb636bbf2\"" Sep 12 10:12:06.103719 kubelet[2590]: E0912 10:12:06.102776 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:06.112368 systemd[1]: Started cri-containerd-ebce6889aa87f25e36b67579d585e0d8177c21dacb0216215c92b442f0c93320.scope - libcontainer container ebce6889aa87f25e36b67579d585e0d8177c21dacb0216215c92b442f0c93320. Sep 12 10:12:06.123699 containerd[1490]: time="2025-09-12T10:12:06.123657495Z" level=info msg="CreateContainer within sandbox \"7066ef0351fa6aace8f6fce6d6415507d2c96fad9f94e8126f68321bb636bbf2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 10:12:06.140972 containerd[1490]: time="2025-09-12T10:12:06.140085687Z" level=info msg="CreateContainer within sandbox \"7066ef0351fa6aace8f6fce6d6415507d2c96fad9f94e8126f68321bb636bbf2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8f8a9736a3ab2f3b1648fb1096eaa6e2b5ca53872ece9d2d110108a995707d2e\"" Sep 12 10:12:06.142264 containerd[1490]: time="2025-09-12T10:12:06.141845642Z" level=info msg="StartContainer for \"8f8a9736a3ab2f3b1648fb1096eaa6e2b5ca53872ece9d2d110108a995707d2e\"" Sep 12 10:12:06.170609 containerd[1490]: time="2025-09-12T10:12:06.170118078Z" level=info msg="StartContainer for \"ebce6889aa87f25e36b67579d585e0d8177c21dacb0216215c92b442f0c93320\" returns successfully" Sep 12 10:12:06.200852 systemd[1]: Started cri-containerd-8f8a9736a3ab2f3b1648fb1096eaa6e2b5ca53872ece9d2d110108a995707d2e.scope - libcontainer container 8f8a9736a3ab2f3b1648fb1096eaa6e2b5ca53872ece9d2d110108a995707d2e. Sep 12 10:12:06.245153 containerd[1490]: time="2025-09-12T10:12:06.245095933Z" level=info msg="StartContainer for \"8f8a9736a3ab2f3b1648fb1096eaa6e2b5ca53872ece9d2d110108a995707d2e\" returns successfully" Sep 12 10:12:06.863167 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2092127840.mount: Deactivated successfully. Sep 12 10:12:06.966402 kubelet[2590]: E0912 10:12:06.966248 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:06.976862 kubelet[2590]: E0912 10:12:06.976233 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:06.995657 kubelet[2590]: I0912 10:12:06.994383 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-5l2kb" podStartSLOduration=21.99436133 podStartE2EDuration="21.99436133s" podCreationTimestamp="2025-09-12 10:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:12:06.988948641 +0000 UTC m=+28.381689759" watchObservedRunningTime="2025-09-12 10:12:06.99436133 +0000 UTC m=+28.387102452" Sep 12 10:12:07.052501 kubelet[2590]: I0912 10:12:07.052358 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-65q6r" podStartSLOduration=22.05233659 podStartE2EDuration="22.05233659s" podCreationTimestamp="2025-09-12 10:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:12:07.017859425 +0000 UTC m=+28.410600543" watchObservedRunningTime="2025-09-12 10:12:07.05233659 +0000 UTC m=+28.445077708" Sep 12 10:12:07.978615 kubelet[2590]: E0912 10:12:07.978558 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:07.979378 kubelet[2590]: E0912 10:12:07.979291 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:08.980974 kubelet[2590]: E0912 10:12:08.980739 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:08.980974 kubelet[2590]: E0912 10:12:08.980894 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:14.503116 kubelet[2590]: I0912 10:12:14.502836 2590 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 10:12:14.504214 kubelet[2590]: E0912 10:12:14.503805 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:14.993574 kubelet[2590]: E0912 10:12:14.993455 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:17.534934 systemd[1]: Started sshd@7-164.92.125.86:22-139.178.68.195:51772.service - OpenSSH per-connection server daemon (139.178.68.195:51772). Sep 12 10:12:17.634360 sshd[3984]: Accepted publickey for core from 139.178.68.195 port 51772 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:17.636215 sshd-session[3984]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:17.642943 systemd-logind[1466]: New session 8 of user core. Sep 12 10:12:17.650830 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 12 10:12:18.215433 sshd[3986]: Connection closed by 139.178.68.195 port 51772 Sep 12 10:12:18.216348 sshd-session[3984]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:18.221324 systemd[1]: sshd@7-164.92.125.86:22-139.178.68.195:51772.service: Deactivated successfully. Sep 12 10:12:18.225115 systemd[1]: session-8.scope: Deactivated successfully. Sep 12 10:12:18.226525 systemd-logind[1466]: Session 8 logged out. Waiting for processes to exit. Sep 12 10:12:18.227874 systemd-logind[1466]: Removed session 8. Sep 12 10:12:23.242109 systemd[1]: Started sshd@8-164.92.125.86:22-139.178.68.195:33802.service - OpenSSH per-connection server daemon (139.178.68.195:33802). Sep 12 10:12:23.295721 sshd[3999]: Accepted publickey for core from 139.178.68.195 port 33802 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:23.297485 sshd-session[3999]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:23.303154 systemd-logind[1466]: New session 9 of user core. Sep 12 10:12:23.310948 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 12 10:12:23.459022 sshd[4001]: Connection closed by 139.178.68.195 port 33802 Sep 12 10:12:23.459610 sshd-session[3999]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:23.464823 systemd[1]: sshd@8-164.92.125.86:22-139.178.68.195:33802.service: Deactivated successfully. Sep 12 10:12:23.467789 systemd[1]: session-9.scope: Deactivated successfully. Sep 12 10:12:23.469107 systemd-logind[1466]: Session 9 logged out. Waiting for processes to exit. Sep 12 10:12:23.470794 systemd-logind[1466]: Removed session 9. Sep 12 10:12:28.484078 systemd[1]: Started sshd@9-164.92.125.86:22-139.178.68.195:33808.service - OpenSSH per-connection server daemon (139.178.68.195:33808). Sep 12 10:12:28.540388 sshd[4014]: Accepted publickey for core from 139.178.68.195 port 33808 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:28.542468 sshd-session[4014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:28.549398 systemd-logind[1466]: New session 10 of user core. Sep 12 10:12:28.555873 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 12 10:12:28.707780 sshd[4016]: Connection closed by 139.178.68.195 port 33808 Sep 12 10:12:28.709888 sshd-session[4014]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:28.714647 systemd[1]: sshd@9-164.92.125.86:22-139.178.68.195:33808.service: Deactivated successfully. Sep 12 10:12:28.717718 systemd[1]: session-10.scope: Deactivated successfully. Sep 12 10:12:28.719515 systemd-logind[1466]: Session 10 logged out. Waiting for processes to exit. Sep 12 10:12:28.720842 systemd-logind[1466]: Removed session 10. Sep 12 10:12:33.731935 systemd[1]: Started sshd@10-164.92.125.86:22-139.178.68.195:35128.service - OpenSSH per-connection server daemon (139.178.68.195:35128). Sep 12 10:12:33.782429 sshd[4029]: Accepted publickey for core from 139.178.68.195 port 35128 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:33.784298 sshd-session[4029]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:33.792480 systemd-logind[1466]: New session 11 of user core. Sep 12 10:12:33.797797 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 12 10:12:33.950244 sshd[4031]: Connection closed by 139.178.68.195 port 35128 Sep 12 10:12:33.951048 sshd-session[4029]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:33.965398 systemd[1]: sshd@10-164.92.125.86:22-139.178.68.195:35128.service: Deactivated successfully. Sep 12 10:12:33.969067 systemd[1]: session-11.scope: Deactivated successfully. Sep 12 10:12:33.972221 systemd-logind[1466]: Session 11 logged out. Waiting for processes to exit. Sep 12 10:12:33.980112 systemd[1]: Started sshd@11-164.92.125.86:22-139.178.68.195:35136.service - OpenSSH per-connection server daemon (139.178.68.195:35136). Sep 12 10:12:33.982700 systemd-logind[1466]: Removed session 11. Sep 12 10:12:34.030011 sshd[4043]: Accepted publickey for core from 139.178.68.195 port 35136 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:34.031888 sshd-session[4043]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:34.039522 systemd-logind[1466]: New session 12 of user core. Sep 12 10:12:34.052865 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 12 10:12:34.276571 sshd[4046]: Connection closed by 139.178.68.195 port 35136 Sep 12 10:12:34.277306 sshd-session[4043]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:34.291459 systemd[1]: sshd@11-164.92.125.86:22-139.178.68.195:35136.service: Deactivated successfully. Sep 12 10:12:34.294067 systemd[1]: session-12.scope: Deactivated successfully. Sep 12 10:12:34.296285 systemd-logind[1466]: Session 12 logged out. Waiting for processes to exit. Sep 12 10:12:34.309370 systemd[1]: Started sshd@12-164.92.125.86:22-139.178.68.195:35140.service - OpenSSH per-connection server daemon (139.178.68.195:35140). Sep 12 10:12:34.313002 systemd-logind[1466]: Removed session 12. Sep 12 10:12:34.406230 sshd[4055]: Accepted publickey for core from 139.178.68.195 port 35140 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:34.408443 sshd-session[4055]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:34.414764 systemd-logind[1466]: New session 13 of user core. Sep 12 10:12:34.421042 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 12 10:12:34.579153 sshd[4058]: Connection closed by 139.178.68.195 port 35140 Sep 12 10:12:34.578822 sshd-session[4055]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:34.585015 systemd-logind[1466]: Session 13 logged out. Waiting for processes to exit. Sep 12 10:12:34.585939 systemd[1]: sshd@12-164.92.125.86:22-139.178.68.195:35140.service: Deactivated successfully. Sep 12 10:12:34.589894 systemd[1]: session-13.scope: Deactivated successfully. Sep 12 10:12:34.592998 systemd-logind[1466]: Removed session 13. Sep 12 10:12:39.603935 systemd[1]: Started sshd@13-164.92.125.86:22-139.178.68.195:35156.service - OpenSSH per-connection server daemon (139.178.68.195:35156). Sep 12 10:12:39.652086 sshd[4074]: Accepted publickey for core from 139.178.68.195 port 35156 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:39.654143 sshd-session[4074]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:39.660483 systemd-logind[1466]: New session 14 of user core. Sep 12 10:12:39.666054 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 12 10:12:39.808961 sshd[4076]: Connection closed by 139.178.68.195 port 35156 Sep 12 10:12:39.808794 sshd-session[4074]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:39.817000 systemd[1]: sshd@13-164.92.125.86:22-139.178.68.195:35156.service: Deactivated successfully. Sep 12 10:12:39.821341 systemd[1]: session-14.scope: Deactivated successfully. Sep 12 10:12:39.824465 systemd-logind[1466]: Session 14 logged out. Waiting for processes to exit. Sep 12 10:12:39.826311 systemd-logind[1466]: Removed session 14. Sep 12 10:12:44.834673 systemd[1]: Started sshd@14-164.92.125.86:22-139.178.68.195:44206.service - OpenSSH per-connection server daemon (139.178.68.195:44206). Sep 12 10:12:44.893207 sshd[4087]: Accepted publickey for core from 139.178.68.195 port 44206 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:44.894839 sshd-session[4087]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:44.900782 systemd-logind[1466]: New session 15 of user core. Sep 12 10:12:44.912837 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 12 10:12:45.074919 sshd[4089]: Connection closed by 139.178.68.195 port 44206 Sep 12 10:12:45.075798 sshd-session[4087]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:45.085381 systemd[1]: sshd@14-164.92.125.86:22-139.178.68.195:44206.service: Deactivated successfully. Sep 12 10:12:45.087930 systemd[1]: session-15.scope: Deactivated successfully. Sep 12 10:12:45.089523 systemd-logind[1466]: Session 15 logged out. Waiting for processes to exit. Sep 12 10:12:45.090690 systemd-logind[1466]: Removed session 15. Sep 12 10:12:50.102074 systemd[1]: Started sshd@15-164.92.125.86:22-139.178.68.195:39206.service - OpenSSH per-connection server daemon (139.178.68.195:39206). Sep 12 10:12:50.169464 sshd[4102]: Accepted publickey for core from 139.178.68.195 port 39206 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:50.171393 sshd-session[4102]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:50.176615 systemd-logind[1466]: New session 16 of user core. Sep 12 10:12:50.185853 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 12 10:12:50.333546 sshd[4104]: Connection closed by 139.178.68.195 port 39206 Sep 12 10:12:50.334197 sshd-session[4102]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:50.350350 systemd[1]: sshd@15-164.92.125.86:22-139.178.68.195:39206.service: Deactivated successfully. Sep 12 10:12:50.353063 systemd[1]: session-16.scope: Deactivated successfully. Sep 12 10:12:50.356159 systemd-logind[1466]: Session 16 logged out. Waiting for processes to exit. Sep 12 10:12:50.361225 systemd[1]: Started sshd@16-164.92.125.86:22-139.178.68.195:39214.service - OpenSSH per-connection server daemon (139.178.68.195:39214). Sep 12 10:12:50.363786 systemd-logind[1466]: Removed session 16. Sep 12 10:12:50.416009 sshd[4115]: Accepted publickey for core from 139.178.68.195 port 39214 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:50.417281 sshd-session[4115]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:50.422858 systemd-logind[1466]: New session 17 of user core. Sep 12 10:12:50.430918 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 12 10:12:50.751614 kubelet[2590]: E0912 10:12:50.751446 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:50.769910 sshd[4118]: Connection closed by 139.178.68.195 port 39214 Sep 12 10:12:50.767799 sshd-session[4115]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:50.780979 systemd[1]: sshd@16-164.92.125.86:22-139.178.68.195:39214.service: Deactivated successfully. Sep 12 10:12:50.784963 systemd[1]: session-17.scope: Deactivated successfully. Sep 12 10:12:50.786875 systemd-logind[1466]: Session 17 logged out. Waiting for processes to exit. Sep 12 10:12:50.795148 systemd[1]: Started sshd@17-164.92.125.86:22-139.178.68.195:39228.service - OpenSSH per-connection server daemon (139.178.68.195:39228). Sep 12 10:12:50.797839 systemd-logind[1466]: Removed session 17. Sep 12 10:12:50.882943 sshd[4127]: Accepted publickey for core from 139.178.68.195 port 39228 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:50.884501 sshd-session[4127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:50.890755 systemd-logind[1466]: New session 18 of user core. Sep 12 10:12:50.899778 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 12 10:12:51.603741 sshd[4130]: Connection closed by 139.178.68.195 port 39228 Sep 12 10:12:51.604258 sshd-session[4127]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:51.634358 systemd[1]: Started sshd@18-164.92.125.86:22-139.178.68.195:39240.service - OpenSSH per-connection server daemon (139.178.68.195:39240). Sep 12 10:12:51.636158 systemd[1]: sshd@17-164.92.125.86:22-139.178.68.195:39228.service: Deactivated successfully. Sep 12 10:12:51.649229 systemd[1]: session-18.scope: Deactivated successfully. Sep 12 10:12:51.654000 systemd-logind[1466]: Session 18 logged out. Waiting for processes to exit. Sep 12 10:12:51.659873 systemd-logind[1466]: Removed session 18. Sep 12 10:12:51.706600 sshd[4145]: Accepted publickey for core from 139.178.68.195 port 39240 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:51.708244 sshd-session[4145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:51.716648 systemd-logind[1466]: New session 19 of user core. Sep 12 10:12:51.725861 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 12 10:12:51.749703 kubelet[2590]: E0912 10:12:51.749608 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:12:52.047223 sshd[4150]: Connection closed by 139.178.68.195 port 39240 Sep 12 10:12:52.047742 sshd-session[4145]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:52.061072 systemd[1]: sshd@18-164.92.125.86:22-139.178.68.195:39240.service: Deactivated successfully. Sep 12 10:12:52.065303 systemd[1]: session-19.scope: Deactivated successfully. Sep 12 10:12:52.069286 systemd-logind[1466]: Session 19 logged out. Waiting for processes to exit. Sep 12 10:12:52.076015 systemd[1]: Started sshd@19-164.92.125.86:22-139.178.68.195:39242.service - OpenSSH per-connection server daemon (139.178.68.195:39242). Sep 12 10:12:52.078140 systemd-logind[1466]: Removed session 19. Sep 12 10:12:52.120824 sshd[4159]: Accepted publickey for core from 139.178.68.195 port 39242 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:52.122941 sshd-session[4159]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:52.128874 systemd-logind[1466]: New session 20 of user core. Sep 12 10:12:52.135819 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 12 10:12:52.265907 sshd[4162]: Connection closed by 139.178.68.195 port 39242 Sep 12 10:12:52.266732 sshd-session[4159]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:52.270858 systemd[1]: sshd@19-164.92.125.86:22-139.178.68.195:39242.service: Deactivated successfully. Sep 12 10:12:52.273648 systemd[1]: session-20.scope: Deactivated successfully. Sep 12 10:12:52.276164 systemd-logind[1466]: Session 20 logged out. Waiting for processes to exit. Sep 12 10:12:52.277807 systemd-logind[1466]: Removed session 20. Sep 12 10:12:57.291953 systemd[1]: Started sshd@20-164.92.125.86:22-139.178.68.195:39248.service - OpenSSH per-connection server daemon (139.178.68.195:39248). Sep 12 10:12:57.342695 sshd[4173]: Accepted publickey for core from 139.178.68.195 port 39248 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:12:57.344289 sshd-session[4173]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:12:57.350247 systemd-logind[1466]: New session 21 of user core. Sep 12 10:12:57.352789 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 12 10:12:57.489177 sshd[4175]: Connection closed by 139.178.68.195 port 39248 Sep 12 10:12:57.490017 sshd-session[4173]: pam_unix(sshd:session): session closed for user core Sep 12 10:12:57.496226 systemd[1]: sshd@20-164.92.125.86:22-139.178.68.195:39248.service: Deactivated successfully. Sep 12 10:12:57.501300 systemd[1]: session-21.scope: Deactivated successfully. Sep 12 10:12:57.503629 systemd-logind[1466]: Session 21 logged out. Waiting for processes to exit. Sep 12 10:12:57.505188 systemd-logind[1466]: Removed session 21. Sep 12 10:13:02.510075 systemd[1]: Started sshd@21-164.92.125.86:22-139.178.68.195:41982.service - OpenSSH per-connection server daemon (139.178.68.195:41982). Sep 12 10:13:02.558476 sshd[4189]: Accepted publickey for core from 139.178.68.195 port 41982 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:02.560512 sshd-session[4189]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:02.567906 systemd-logind[1466]: New session 22 of user core. Sep 12 10:13:02.577159 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 12 10:13:02.710063 sshd[4191]: Connection closed by 139.178.68.195 port 41982 Sep 12 10:13:02.711801 sshd-session[4189]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:02.717777 systemd[1]: sshd@21-164.92.125.86:22-139.178.68.195:41982.service: Deactivated successfully. Sep 12 10:13:02.721201 systemd[1]: session-22.scope: Deactivated successfully. Sep 12 10:13:02.723069 systemd-logind[1466]: Session 22 logged out. Waiting for processes to exit. Sep 12 10:13:02.725029 systemd-logind[1466]: Removed session 22. Sep 12 10:13:05.750422 kubelet[2590]: E0912 10:13:05.749393 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:07.734979 systemd[1]: Started sshd@22-164.92.125.86:22-139.178.68.195:41992.service - OpenSSH per-connection server daemon (139.178.68.195:41992). Sep 12 10:13:07.796429 sshd[4204]: Accepted publickey for core from 139.178.68.195 port 41992 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:07.798498 sshd-session[4204]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:07.804373 systemd-logind[1466]: New session 23 of user core. Sep 12 10:13:07.813262 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 12 10:13:07.976681 sshd[4206]: Connection closed by 139.178.68.195 port 41992 Sep 12 10:13:07.977758 sshd-session[4204]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:07.982832 systemd[1]: sshd@22-164.92.125.86:22-139.178.68.195:41992.service: Deactivated successfully. Sep 12 10:13:07.985588 systemd[1]: session-23.scope: Deactivated successfully. Sep 12 10:13:07.988054 systemd-logind[1466]: Session 23 logged out. Waiting for processes to exit. Sep 12 10:13:07.989783 systemd-logind[1466]: Removed session 23. Sep 12 10:13:12.749404 kubelet[2590]: E0912 10:13:12.748639 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:13.003971 systemd[1]: Started sshd@23-164.92.125.86:22-139.178.68.195:60194.service - OpenSSH per-connection server daemon (139.178.68.195:60194). Sep 12 10:13:13.046315 sshd[4218]: Accepted publickey for core from 139.178.68.195 port 60194 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:13.048116 sshd-session[4218]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:13.054220 systemd-logind[1466]: New session 24 of user core. Sep 12 10:13:13.060826 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 12 10:13:13.188317 sshd[4220]: Connection closed by 139.178.68.195 port 60194 Sep 12 10:13:13.190734 sshd-session[4218]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:13.203262 systemd[1]: sshd@23-164.92.125.86:22-139.178.68.195:60194.service: Deactivated successfully. Sep 12 10:13:13.206842 systemd[1]: session-24.scope: Deactivated successfully. Sep 12 10:13:13.207988 systemd-logind[1466]: Session 24 logged out. Waiting for processes to exit. Sep 12 10:13:13.215961 systemd[1]: Started sshd@24-164.92.125.86:22-139.178.68.195:60204.service - OpenSSH per-connection server daemon (139.178.68.195:60204). Sep 12 10:13:13.218773 systemd-logind[1466]: Removed session 24. Sep 12 10:13:13.263775 sshd[4231]: Accepted publickey for core from 139.178.68.195 port 60204 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:13.265634 sshd-session[4231]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:13.272793 systemd-logind[1466]: New session 25 of user core. Sep 12 10:13:13.277801 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 12 10:13:13.749581 kubelet[2590]: E0912 10:13:13.749402 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:13.750097 kubelet[2590]: E0912 10:13:13.749919 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:14.746616 containerd[1490]: time="2025-09-12T10:13:14.744870246Z" level=info msg="StopContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" with timeout 30 (s)" Sep 12 10:13:14.746616 containerd[1490]: time="2025-09-12T10:13:14.745729327Z" level=info msg="Stop container \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" with signal terminated" Sep 12 10:13:14.785102 containerd[1490]: time="2025-09-12T10:13:14.785031602Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 12 10:13:14.794494 containerd[1490]: time="2025-09-12T10:13:14.794310963Z" level=info msg="StopContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" with timeout 2 (s)" Sep 12 10:13:14.796254 systemd[1]: cri-containerd-7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f.scope: Deactivated successfully. Sep 12 10:13:14.797805 containerd[1490]: time="2025-09-12T10:13:14.797521237Z" level=info msg="Stop container \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" with signal terminated" Sep 12 10:13:14.809813 systemd-networkd[1378]: lxc_health: Link DOWN Sep 12 10:13:14.809824 systemd-networkd[1378]: lxc_health: Lost carrier Sep 12 10:13:14.842208 systemd[1]: cri-containerd-e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83.scope: Deactivated successfully. Sep 12 10:13:14.843966 systemd[1]: cri-containerd-e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83.scope: Consumed 8.286s CPU time, 193.6M memory peak, 69M read from disk, 13.3M written to disk. Sep 12 10:13:14.866237 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f-rootfs.mount: Deactivated successfully. Sep 12 10:13:14.869604 containerd[1490]: time="2025-09-12T10:13:14.869501727Z" level=info msg="shim disconnected" id=7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f namespace=k8s.io Sep 12 10:13:14.869800 containerd[1490]: time="2025-09-12T10:13:14.869781649Z" level=warning msg="cleaning up after shim disconnected" id=7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f namespace=k8s.io Sep 12 10:13:14.869860 containerd[1490]: time="2025-09-12T10:13:14.869849189Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:14.889261 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83-rootfs.mount: Deactivated successfully. Sep 12 10:13:14.895139 containerd[1490]: time="2025-09-12T10:13:14.894830583Z" level=info msg="shim disconnected" id=e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83 namespace=k8s.io Sep 12 10:13:14.895139 containerd[1490]: time="2025-09-12T10:13:14.894988513Z" level=warning msg="cleaning up after shim disconnected" id=e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83 namespace=k8s.io Sep 12 10:13:14.895139 containerd[1490]: time="2025-09-12T10:13:14.894999663Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:14.902940 containerd[1490]: time="2025-09-12T10:13:14.902511829Z" level=info msg="StopContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" returns successfully" Sep 12 10:13:14.904393 containerd[1490]: time="2025-09-12T10:13:14.904075654Z" level=info msg="StopPodSandbox for \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\"" Sep 12 10:13:14.911780 containerd[1490]: time="2025-09-12T10:13:14.909603559Z" level=info msg="Container to stop \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.914807 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353-shm.mount: Deactivated successfully. Sep 12 10:13:14.934135 containerd[1490]: time="2025-09-12T10:13:14.933795679Z" level=info msg="StopContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" returns successfully" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934565577Z" level=info msg="StopPodSandbox for \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\"" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934600876Z" level=info msg="Container to stop \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934636929Z" level=info msg="Container to stop \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934645843Z" level=info msg="Container to stop \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934654378Z" level=info msg="Container to stop \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.934748 containerd[1490]: time="2025-09-12T10:13:14.934661970Z" level=info msg="Container to stop \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 12 10:13:14.938397 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9-shm.mount: Deactivated successfully. Sep 12 10:13:14.940496 systemd[1]: cri-containerd-4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353.scope: Deactivated successfully. Sep 12 10:13:14.951001 systemd[1]: cri-containerd-0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9.scope: Deactivated successfully. Sep 12 10:13:14.979403 containerd[1490]: time="2025-09-12T10:13:14.979189813Z" level=info msg="shim disconnected" id=4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353 namespace=k8s.io Sep 12 10:13:14.979403 containerd[1490]: time="2025-09-12T10:13:14.979320674Z" level=warning msg="cleaning up after shim disconnected" id=4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353 namespace=k8s.io Sep 12 10:13:14.979403 containerd[1490]: time="2025-09-12T10:13:14.979331294Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:14.983272 containerd[1490]: time="2025-09-12T10:13:14.981157365Z" level=info msg="shim disconnected" id=0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9 namespace=k8s.io Sep 12 10:13:14.983272 containerd[1490]: time="2025-09-12T10:13:14.981207790Z" level=warning msg="cleaning up after shim disconnected" id=0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9 namespace=k8s.io Sep 12 10:13:14.983272 containerd[1490]: time="2025-09-12T10:13:14.981219703Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:14.999984 containerd[1490]: time="2025-09-12T10:13:14.999870494Z" level=info msg="TearDown network for sandbox \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" successfully" Sep 12 10:13:14.999984 containerd[1490]: time="2025-09-12T10:13:14.999902357Z" level=info msg="StopPodSandbox for \"0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9\" returns successfully" Sep 12 10:13:15.000583 containerd[1490]: time="2025-09-12T10:13:15.000551236Z" level=info msg="TearDown network for sandbox \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\" successfully" Sep 12 10:13:15.000919 containerd[1490]: time="2025-09-12T10:13:15.000883147Z" level=info msg="StopPodSandbox for \"4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353\" returns successfully" Sep 12 10:13:15.105301 kubelet[2590]: I0912 10:13:15.105240 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-hubble-tls\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.106014 kubelet[2590]: I0912 10:13:15.105891 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-cilium-config-path\") pod \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\" (UID: \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\") " Sep 12 10:13:15.106179 kubelet[2590]: I0912 10:13:15.106141 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98nzq\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-kube-api-access-98nzq\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.107610 kubelet[2590]: I0912 10:13:15.107580 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b636ea9b-fac2-43d4-abde-838ed8abdb63-clustermesh-secrets\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.107745 kubelet[2590]: I0912 10:13:15.107732 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-net\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107812 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-kernel\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107834 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-config-path\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107852 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-lib-modules\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107868 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-run\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107884 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cni-path\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108301 kubelet[2590]: I0912 10:13:15.107902 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-xtables-lock\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108551 kubelet[2590]: I0912 10:13:15.107918 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-hostproc\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108551 kubelet[2590]: I0912 10:13:15.107933 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-bpf-maps\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108551 kubelet[2590]: I0912 10:13:15.107950 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-cgroup\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108551 kubelet[2590]: I0912 10:13:15.107963 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-etc-cni-netd\") pod \"b636ea9b-fac2-43d4-abde-838ed8abdb63\" (UID: \"b636ea9b-fac2-43d4-abde-838ed8abdb63\") " Sep 12 10:13:15.108551 kubelet[2590]: I0912 10:13:15.107984 2590 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c4cm\" (UniqueName: \"kubernetes.io/projected/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-kube-api-access-8c4cm\") pod \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\" (UID: \"1bc90cd3-fbf9-4494-8740-a14bf04b19ca\") " Sep 12 10:13:15.109337 kubelet[2590]: I0912 10:13:15.109306 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109560 kubelet[2590]: I0912 10:13:15.109522 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cni-path" (OuterVolumeSpecName: "cni-path") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109674 kubelet[2590]: I0912 10:13:15.109662 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109753 kubelet[2590]: I0912 10:13:15.109743 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-hostproc" (OuterVolumeSpecName: "hostproc") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109825 kubelet[2590]: I0912 10:13:15.109814 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109957 kubelet[2590]: I0912 10:13:15.109898 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109957 kubelet[2590]: I0912 10:13:15.109917 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.109957 kubelet[2590]: I0912 10:13:15.109936 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.110855 kubelet[2590]: I0912 10:13:15.110746 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.111315 kubelet[2590]: I0912 10:13:15.111294 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 12 10:13:15.111706 kubelet[2590]: I0912 10:13:15.111464 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 12 10:13:15.115523 kubelet[2590]: I0912 10:13:15.115422 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1bc90cd3-fbf9-4494-8740-a14bf04b19ca" (UID: "1bc90cd3-fbf9-4494-8740-a14bf04b19ca"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 12 10:13:15.116363 kubelet[2590]: I0912 10:13:15.116247 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-kube-api-access-98nzq" (OuterVolumeSpecName: "kube-api-access-98nzq") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "kube-api-access-98nzq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 12 10:13:15.116891 kubelet[2590]: I0912 10:13:15.116859 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 12 10:13:15.117472 kubelet[2590]: I0912 10:13:15.117406 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-kube-api-access-8c4cm" (OuterVolumeSpecName: "kube-api-access-8c4cm") pod "1bc90cd3-fbf9-4494-8740-a14bf04b19ca" (UID: "1bc90cd3-fbf9-4494-8740-a14bf04b19ca"). InnerVolumeSpecName "kube-api-access-8c4cm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 12 10:13:15.117636 kubelet[2590]: I0912 10:13:15.117497 2590 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b636ea9b-fac2-43d4-abde-838ed8abdb63-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "b636ea9b-fac2-43d4-abde-838ed8abdb63" (UID: "b636ea9b-fac2-43d4-abde-838ed8abdb63"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 12 10:13:15.125771 kubelet[2590]: I0912 10:13:15.125733 2590 scope.go:117] "RemoveContainer" containerID="7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f" Sep 12 10:13:15.137962 containerd[1490]: time="2025-09-12T10:13:15.137404405Z" level=info msg="RemoveContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\"" Sep 12 10:13:15.138022 systemd[1]: Removed slice kubepods-besteffort-pod1bc90cd3_fbf9_4494_8740_a14bf04b19ca.slice - libcontainer container kubepods-besteffort-pod1bc90cd3_fbf9_4494_8740_a14bf04b19ca.slice. Sep 12 10:13:15.141368 systemd[1]: Removed slice kubepods-burstable-podb636ea9b_fac2_43d4_abde_838ed8abdb63.slice - libcontainer container kubepods-burstable-podb636ea9b_fac2_43d4_abde_838ed8abdb63.slice. Sep 12 10:13:15.141482 systemd[1]: kubepods-burstable-podb636ea9b_fac2_43d4_abde_838ed8abdb63.slice: Consumed 8.391s CPU time, 193.9M memory peak, 69M read from disk, 13.3M written to disk. Sep 12 10:13:15.145591 containerd[1490]: time="2025-09-12T10:13:15.145180459Z" level=info msg="RemoveContainer for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" returns successfully" Sep 12 10:13:15.148553 kubelet[2590]: I0912 10:13:15.148320 2590 scope.go:117] "RemoveContainer" containerID="7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f" Sep 12 10:13:15.149044 containerd[1490]: time="2025-09-12T10:13:15.148878892Z" level=error msg="ContainerStatus for \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\": not found" Sep 12 10:13:15.159685 kubelet[2590]: E0912 10:13:15.158524 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\": not found" containerID="7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f" Sep 12 10:13:15.159866 kubelet[2590]: I0912 10:13:15.159696 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f"} err="failed to get container status \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\": rpc error: code = NotFound desc = an error occurred when try to find container \"7dbac6c19defc72ad04d640151b40125964b276a5f3dbf3dd0d034b33387be6f\": not found" Sep 12 10:13:15.159866 kubelet[2590]: I0912 10:13:15.159782 2590 scope.go:117] "RemoveContainer" containerID="e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83" Sep 12 10:13:15.163084 containerd[1490]: time="2025-09-12T10:13:15.163037312Z" level=info msg="RemoveContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\"" Sep 12 10:13:15.169713 containerd[1490]: time="2025-09-12T10:13:15.169503256Z" level=info msg="RemoveContainer for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" returns successfully" Sep 12 10:13:15.169832 kubelet[2590]: I0912 10:13:15.169791 2590 scope.go:117] "RemoveContainer" containerID="97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b" Sep 12 10:13:15.173791 containerd[1490]: time="2025-09-12T10:13:15.173415997Z" level=info msg="RemoveContainer for \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\"" Sep 12 10:13:15.176095 containerd[1490]: time="2025-09-12T10:13:15.176042714Z" level=info msg="RemoveContainer for \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\" returns successfully" Sep 12 10:13:15.176515 kubelet[2590]: I0912 10:13:15.176485 2590 scope.go:117] "RemoveContainer" containerID="afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf" Sep 12 10:13:15.177921 containerd[1490]: time="2025-09-12T10:13:15.177804576Z" level=info msg="RemoveContainer for \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\"" Sep 12 10:13:15.180447 containerd[1490]: time="2025-09-12T10:13:15.180392466Z" level=info msg="RemoveContainer for \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\" returns successfully" Sep 12 10:13:15.180695 kubelet[2590]: I0912 10:13:15.180669 2590 scope.go:117] "RemoveContainer" containerID="aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e" Sep 12 10:13:15.182000 containerd[1490]: time="2025-09-12T10:13:15.181972326Z" level=info msg="RemoveContainer for \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\"" Sep 12 10:13:15.184112 containerd[1490]: time="2025-09-12T10:13:15.184056217Z" level=info msg="RemoveContainer for \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\" returns successfully" Sep 12 10:13:15.184240 kubelet[2590]: I0912 10:13:15.184214 2590 scope.go:117] "RemoveContainer" containerID="4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c" Sep 12 10:13:15.185348 containerd[1490]: time="2025-09-12T10:13:15.185295233Z" level=info msg="RemoveContainer for \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\"" Sep 12 10:13:15.187255 containerd[1490]: time="2025-09-12T10:13:15.187215450Z" level=info msg="RemoveContainer for \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\" returns successfully" Sep 12 10:13:15.187494 kubelet[2590]: I0912 10:13:15.187469 2590 scope.go:117] "RemoveContainer" containerID="e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83" Sep 12 10:13:15.187786 containerd[1490]: time="2025-09-12T10:13:15.187755969Z" level=error msg="ContainerStatus for \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\": not found" Sep 12 10:13:15.187898 kubelet[2590]: E0912 10:13:15.187879 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\": not found" containerID="e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83" Sep 12 10:13:15.187963 kubelet[2590]: I0912 10:13:15.187937 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83"} err="failed to get container status \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\": rpc error: code = NotFound desc = an error occurred when try to find container \"e4a8c3dd091eca3904adaca5f54e41cdc3f20aa041058cb734617d7993fb8d83\": not found" Sep 12 10:13:15.187995 kubelet[2590]: I0912 10:13:15.187965 2590 scope.go:117] "RemoveContainer" containerID="97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b" Sep 12 10:13:15.188170 containerd[1490]: time="2025-09-12T10:13:15.188134499Z" level=error msg="ContainerStatus for \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\": not found" Sep 12 10:13:15.188267 kubelet[2590]: E0912 10:13:15.188243 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\": not found" containerID="97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b" Sep 12 10:13:15.188307 kubelet[2590]: I0912 10:13:15.188265 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b"} err="failed to get container status \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\": rpc error: code = NotFound desc = an error occurred when try to find container \"97716d38aeb4dbf7767f3b3a24f139377c03620a0cf437838ecdf846ecf3262b\": not found" Sep 12 10:13:15.188307 kubelet[2590]: I0912 10:13:15.188282 2590 scope.go:117] "RemoveContainer" containerID="afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf" Sep 12 10:13:15.188458 containerd[1490]: time="2025-09-12T10:13:15.188434152Z" level=error msg="ContainerStatus for \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\": not found" Sep 12 10:13:15.188680 kubelet[2590]: E0912 10:13:15.188620 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\": not found" containerID="afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf" Sep 12 10:13:15.188737 kubelet[2590]: I0912 10:13:15.188687 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf"} err="failed to get container status \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\": rpc error: code = NotFound desc = an error occurred when try to find container \"afd18d95652ba1e7f985a0fb5cb66d128531754ff5496189c3ffb383459a5ecf\": not found" Sep 12 10:13:15.188737 kubelet[2590]: I0912 10:13:15.188706 2590 scope.go:117] "RemoveContainer" containerID="aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e" Sep 12 10:13:15.188898 containerd[1490]: time="2025-09-12T10:13:15.188869298Z" level=error msg="ContainerStatus for \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\": not found" Sep 12 10:13:15.191569 kubelet[2590]: E0912 10:13:15.189608 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\": not found" containerID="aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e" Sep 12 10:13:15.191569 kubelet[2590]: I0912 10:13:15.189637 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e"} err="failed to get container status \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\": rpc error: code = NotFound desc = an error occurred when try to find container \"aa805180ec623f4476750cda898f51b5b3d492163bb831b156d3f620867b8e6e\": not found" Sep 12 10:13:15.191569 kubelet[2590]: I0912 10:13:15.189656 2590 scope.go:117] "RemoveContainer" containerID="4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c" Sep 12 10:13:15.191569 kubelet[2590]: E0912 10:13:15.189939 2590 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\": not found" containerID="4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c" Sep 12 10:13:15.191569 kubelet[2590]: I0912 10:13:15.189962 2590 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c"} err="failed to get container status \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\": rpc error: code = NotFound desc = an error occurred when try to find container \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\": not found" Sep 12 10:13:15.191878 containerd[1490]: time="2025-09-12T10:13:15.189842968Z" level=error msg="ContainerStatus for \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4c4ef7d92c9039316764ee5ae6b1a8ed704c5cc60b9bd3e299626cf07656573c\": not found" Sep 12 10:13:15.208417 kubelet[2590]: I0912 10:13:15.208356 2590 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8c4cm\" (UniqueName: \"kubernetes.io/projected/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-kube-api-access-8c4cm\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.208762 kubelet[2590]: I0912 10:13:15.208728 2590 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-hubble-tls\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.208892 kubelet[2590]: I0912 10:13:15.208874 2590 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1bc90cd3-fbf9-4494-8740-a14bf04b19ca-cilium-config-path\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.208962 kubelet[2590]: I0912 10:13:15.208951 2590 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-98nzq\" (UniqueName: \"kubernetes.io/projected/b636ea9b-fac2-43d4-abde-838ed8abdb63-kube-api-access-98nzq\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209030 kubelet[2590]: I0912 10:13:15.209020 2590 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b636ea9b-fac2-43d4-abde-838ed8abdb63-clustermesh-secrets\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209077 kubelet[2590]: I0912 10:13:15.209069 2590 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-net\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209120 kubelet[2590]: I0912 10:13:15.209112 2590 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-host-proc-sys-kernel\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209175 kubelet[2590]: I0912 10:13:15.209166 2590 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-config-path\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209222 kubelet[2590]: I0912 10:13:15.209214 2590 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-lib-modules\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209268 kubelet[2590]: I0912 10:13:15.209261 2590 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-run\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209312 kubelet[2590]: I0912 10:13:15.209305 2590 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cni-path\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209379 kubelet[2590]: I0912 10:13:15.209367 2590 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-xtables-lock\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209436 kubelet[2590]: I0912 10:13:15.209428 2590 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-hostproc\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209482 kubelet[2590]: I0912 10:13:15.209475 2590 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-bpf-maps\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209524 kubelet[2590]: I0912 10:13:15.209517 2590 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-cilium-cgroup\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.209586 kubelet[2590]: I0912 10:13:15.209578 2590 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b636ea9b-fac2-43d4-abde-838ed8abdb63-etc-cni-netd\") on node \"ci-4230.2.2-n-dc4800d201\" DevicePath \"\"" Sep 12 10:13:15.745502 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4fb61e50c1b007a574d54f4a2348a853057f1a46f2e3498919c48f8455dac353-rootfs.mount: Deactivated successfully. Sep 12 10:13:15.745655 systemd[1]: var-lib-kubelet-pods-1bc90cd3\x2dfbf9\x2d4494\x2d8740\x2da14bf04b19ca-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8c4cm.mount: Deactivated successfully. Sep 12 10:13:15.745739 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0ee0ef055e143bebf655a9aa2eb2fd7d21fb2a67c8bdcecf862784ffc84561f9-rootfs.mount: Deactivated successfully. Sep 12 10:13:15.745857 systemd[1]: var-lib-kubelet-pods-b636ea9b\x2dfac2\x2d43d4\x2dabde\x2d838ed8abdb63-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d98nzq.mount: Deactivated successfully. Sep 12 10:13:15.745932 systemd[1]: var-lib-kubelet-pods-b636ea9b\x2dfac2\x2d43d4\x2dabde\x2d838ed8abdb63-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 12 10:13:15.746012 systemd[1]: var-lib-kubelet-pods-b636ea9b\x2dfac2\x2d43d4\x2dabde\x2d838ed8abdb63-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 12 10:13:16.651841 sshd[4234]: Connection closed by 139.178.68.195 port 60204 Sep 12 10:13:16.654583 sshd-session[4231]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:16.664017 systemd[1]: sshd@24-164.92.125.86:22-139.178.68.195:60204.service: Deactivated successfully. Sep 12 10:13:16.667186 systemd[1]: session-25.scope: Deactivated successfully. Sep 12 10:13:16.669735 systemd-logind[1466]: Session 25 logged out. Waiting for processes to exit. Sep 12 10:13:16.676117 systemd[1]: Started sshd@25-164.92.125.86:22-139.178.68.195:60220.service - OpenSSH per-connection server daemon (139.178.68.195:60220). Sep 12 10:13:16.678117 systemd-logind[1466]: Removed session 25. Sep 12 10:13:16.763695 kubelet[2590]: I0912 10:13:16.762940 2590 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc90cd3-fbf9-4494-8740-a14bf04b19ca" path="/var/lib/kubelet/pods/1bc90cd3-fbf9-4494-8740-a14bf04b19ca/volumes" Sep 12 10:13:16.763695 kubelet[2590]: I0912 10:13:16.763472 2590 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b636ea9b-fac2-43d4-abde-838ed8abdb63" path="/var/lib/kubelet/pods/b636ea9b-fac2-43d4-abde-838ed8abdb63/volumes" Sep 12 10:13:16.772373 sshd[4397]: Accepted publickey for core from 139.178.68.195 port 60220 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:16.774570 sshd-session[4397]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:16.780952 systemd-logind[1466]: New session 26 of user core. Sep 12 10:13:16.787930 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 12 10:13:17.986228 sshd[4400]: Connection closed by 139.178.68.195 port 60220 Sep 12 10:13:17.987929 sshd-session[4397]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:18.007880 systemd[1]: sshd@25-164.92.125.86:22-139.178.68.195:60220.service: Deactivated successfully. Sep 12 10:13:18.015312 systemd[1]: session-26.scope: Deactivated successfully. Sep 12 10:13:18.015670 systemd[1]: session-26.scope: Consumed 1.080s CPU time, 25.2M memory peak. Sep 12 10:13:18.017570 systemd-logind[1466]: Session 26 logged out. Waiting for processes to exit. Sep 12 10:13:18.030612 systemd[1]: Started sshd@26-164.92.125.86:22-139.178.68.195:60228.service - OpenSSH per-connection server daemon (139.178.68.195:60228). Sep 12 10:13:18.032778 systemd-logind[1466]: Removed session 26. Sep 12 10:13:18.039392 kubelet[2590]: I0912 10:13:18.037714 2590 memory_manager.go:355] "RemoveStaleState removing state" podUID="b636ea9b-fac2-43d4-abde-838ed8abdb63" containerName="cilium-agent" Sep 12 10:13:18.039392 kubelet[2590]: I0912 10:13:18.038632 2590 memory_manager.go:355] "RemoveStaleState removing state" podUID="1bc90cd3-fbf9-4494-8740-a14bf04b19ca" containerName="cilium-operator" Sep 12 10:13:18.059652 systemd[1]: Created slice kubepods-burstable-pod5371b3b2_beee_409b_9905_b2561c984ad4.slice - libcontainer container kubepods-burstable-pod5371b3b2_beee_409b_9905_b2561c984ad4.slice. Sep 12 10:13:18.070256 kubelet[2590]: W0912 10:13:18.070149 2590 reflector.go:569] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-4230.2.2-n-dc4800d201" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object Sep 12 10:13:18.070256 kubelet[2590]: E0912 10:13:18.070202 2590 reflector.go:166] "Unhandled Error" err="object-\"kube-system\"/\"cilium-clustermesh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-clustermesh\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" logger="UnhandledError" Sep 12 10:13:18.070875 kubelet[2590]: I0912 10:13:18.070459 2590 status_manager.go:890] "Failed to get status for pod" podUID="5371b3b2-beee-409b-9905-b2561c984ad4" pod="kube-system/cilium-pgmnl" err="pods \"cilium-pgmnl\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" Sep 12 10:13:18.073098 kubelet[2590]: W0912 10:13:18.072840 2590 reflector.go:569] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-4230.2.2-n-dc4800d201" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object Sep 12 10:13:18.073098 kubelet[2590]: E0912 10:13:18.072904 2590 reflector.go:166] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" logger="UnhandledError" Sep 12 10:13:18.073098 kubelet[2590]: W0912 10:13:18.073025 2590 reflector.go:569] object-"kube-system"/"cilium-ipsec-keys": failed to list *v1.Secret: secrets "cilium-ipsec-keys" is forbidden: User "system:node:ci-4230.2.2-n-dc4800d201" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object Sep 12 10:13:18.073098 kubelet[2590]: E0912 10:13:18.073040 2590 reflector.go:166] "Unhandled Error" err="object-\"kube-system\"/\"cilium-ipsec-keys\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-ipsec-keys\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" logger="UnhandledError" Sep 12 10:13:18.073391 kubelet[2590]: W0912 10:13:18.073347 2590 reflector.go:569] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-4230.2.2-n-dc4800d201" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object Sep 12 10:13:18.073391 kubelet[2590]: E0912 10:13:18.073373 2590 reflector.go:166] "Unhandled Error" err="object-\"kube-system\"/\"cilium-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cilium-config\" is forbidden: User \"system:node:ci-4230.2.2-n-dc4800d201\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230.2.2-n-dc4800d201' and this object" logger="UnhandledError" Sep 12 10:13:18.091160 sshd[4410]: Accepted publickey for core from 139.178.68.195 port 60228 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:18.094609 sshd-session[4410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:18.103107 systemd-logind[1466]: New session 27 of user core. Sep 12 10:13:18.106787 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125784 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-cgroup\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125837 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-clustermesh-secrets\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125855 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-config-path\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125876 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j8k2\" (UniqueName: \"kubernetes.io/projected/5371b3b2-beee-409b-9905-b2561c984ad4-kube-api-access-5j8k2\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125898 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-run\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126177 kubelet[2590]: I0912 10:13:18.125917 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-cni-path\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.125933 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-lib-modules\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.125950 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-bpf-maps\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.125966 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-hostproc\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.125982 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-host-proc-sys-net\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.126001 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/5371b3b2-beee-409b-9905-b2561c984ad4-hubble-tls\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126489 kubelet[2590]: I0912 10:13:18.126019 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-host-proc-sys-kernel\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126693 kubelet[2590]: I0912 10:13:18.126038 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-etc-cni-netd\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126693 kubelet[2590]: I0912 10:13:18.126058 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5371b3b2-beee-409b-9905-b2561c984ad4-xtables-lock\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.126693 kubelet[2590]: I0912 10:13:18.126076 2590 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-ipsec-secrets\") pod \"cilium-pgmnl\" (UID: \"5371b3b2-beee-409b-9905-b2561c984ad4\") " pod="kube-system/cilium-pgmnl" Sep 12 10:13:18.171303 sshd[4413]: Connection closed by 139.178.68.195 port 60228 Sep 12 10:13:18.172204 sshd-session[4410]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:18.187543 systemd[1]: sshd@26-164.92.125.86:22-139.178.68.195:60228.service: Deactivated successfully. Sep 12 10:13:18.191006 systemd[1]: session-27.scope: Deactivated successfully. Sep 12 10:13:18.194630 systemd-logind[1466]: Session 27 logged out. Waiting for processes to exit. Sep 12 10:13:18.200881 systemd[1]: Started sshd@27-164.92.125.86:22-139.178.68.195:60234.service - OpenSSH per-connection server daemon (139.178.68.195:60234). Sep 12 10:13:18.202135 systemd-logind[1466]: Removed session 27. Sep 12 10:13:18.261571 sshd[4419]: Accepted publickey for core from 139.178.68.195 port 60234 ssh2: RSA SHA256:2VqWZqk4hMH9H5AhbP/0AQtkzByPETmNCvQEl/0/v6I Sep 12 10:13:18.263352 sshd-session[4419]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 10:13:18.269912 systemd-logind[1466]: New session 28 of user core. Sep 12 10:13:18.276878 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 12 10:13:18.863623 kubelet[2590]: E0912 10:13:18.863422 2590 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 12 10:13:19.227821 kubelet[2590]: E0912 10:13:19.227668 2590 secret.go:189] Couldn't get secret kube-system/cilium-ipsec-keys: failed to sync secret cache: timed out waiting for the condition Sep 12 10:13:19.228269 kubelet[2590]: E0912 10:13:19.227842 2590 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-ipsec-secrets podName:5371b3b2-beee-409b-9905-b2561c984ad4 nodeName:}" failed. No retries permitted until 2025-09-12 10:13:19.727817918 +0000 UTC m=+101.120559017 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-ipsec-secrets" (UniqueName: "kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-cilium-ipsec-secrets") pod "cilium-pgmnl" (UID: "5371b3b2-beee-409b-9905-b2561c984ad4") : failed to sync secret cache: timed out waiting for the condition Sep 12 10:13:19.228269 kubelet[2590]: E0912 10:13:19.227666 2590 secret.go:189] Couldn't get secret kube-system/cilium-clustermesh: failed to sync secret cache: timed out waiting for the condition Sep 12 10:13:19.228269 kubelet[2590]: E0912 10:13:19.227901 2590 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-clustermesh-secrets podName:5371b3b2-beee-409b-9905-b2561c984ad4 nodeName:}" failed. No retries permitted until 2025-09-12 10:13:19.727893576 +0000 UTC m=+101.120634671 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "clustermesh-secrets" (UniqueName: "kubernetes.io/secret/5371b3b2-beee-409b-9905-b2561c984ad4-clustermesh-secrets") pod "cilium-pgmnl" (UID: "5371b3b2-beee-409b-9905-b2561c984ad4") : failed to sync secret cache: timed out waiting for the condition Sep 12 10:13:19.869818 kubelet[2590]: E0912 10:13:19.869729 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:19.870508 containerd[1490]: time="2025-09-12T10:13:19.870454056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-pgmnl,Uid:5371b3b2-beee-409b-9905-b2561c984ad4,Namespace:kube-system,Attempt:0,}" Sep 12 10:13:19.898435 containerd[1490]: time="2025-09-12T10:13:19.898290504Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 12 10:13:19.899285 containerd[1490]: time="2025-09-12T10:13:19.899221615Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 12 10:13:19.899484 containerd[1490]: time="2025-09-12T10:13:19.899408930Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:13:19.899768 containerd[1490]: time="2025-09-12T10:13:19.899726765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 12 10:13:19.931848 systemd[1]: Started cri-containerd-4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5.scope - libcontainer container 4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5. Sep 12 10:13:19.960912 containerd[1490]: time="2025-09-12T10:13:19.960873037Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-pgmnl,Uid:5371b3b2-beee-409b-9905-b2561c984ad4,Namespace:kube-system,Attempt:0,} returns sandbox id \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\"" Sep 12 10:13:19.961894 kubelet[2590]: E0912 10:13:19.961856 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:19.966576 containerd[1490]: time="2025-09-12T10:13:19.966510342Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 12 10:13:19.981685 containerd[1490]: time="2025-09-12T10:13:19.981550068Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6\"" Sep 12 10:13:19.983467 containerd[1490]: time="2025-09-12T10:13:19.983431194Z" level=info msg="StartContainer for \"05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6\"" Sep 12 10:13:20.026811 systemd[1]: Started cri-containerd-05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6.scope - libcontainer container 05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6. Sep 12 10:13:20.061365 containerd[1490]: time="2025-09-12T10:13:20.061303874Z" level=info msg="StartContainer for \"05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6\" returns successfully" Sep 12 10:13:20.076982 systemd[1]: cri-containerd-05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6.scope: Deactivated successfully. Sep 12 10:13:20.114669 containerd[1490]: time="2025-09-12T10:13:20.114594690Z" level=info msg="shim disconnected" id=05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6 namespace=k8s.io Sep 12 10:13:20.114669 containerd[1490]: time="2025-09-12T10:13:20.114663109Z" level=warning msg="cleaning up after shim disconnected" id=05284b208e952aa5a0df4452a34883a05755577b5f8a6ef6a9b410b067d5dce6 namespace=k8s.io Sep 12 10:13:20.114669 containerd[1490]: time="2025-09-12T10:13:20.114674480Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:20.130262 containerd[1490]: time="2025-09-12T10:13:20.130119432Z" level=warning msg="cleanup warnings time=\"2025-09-12T10:13:20Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 12 10:13:20.143990 kubelet[2590]: E0912 10:13:20.142899 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:20.145993 containerd[1490]: time="2025-09-12T10:13:20.145955533Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 12 10:13:20.157661 containerd[1490]: time="2025-09-12T10:13:20.157238828Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f\"" Sep 12 10:13:20.159744 containerd[1490]: time="2025-09-12T10:13:20.159673575Z" level=info msg="StartContainer for \"fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f\"" Sep 12 10:13:20.195782 systemd[1]: Started cri-containerd-fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f.scope - libcontainer container fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f. Sep 12 10:13:20.227251 containerd[1490]: time="2025-09-12T10:13:20.227186246Z" level=info msg="StartContainer for \"fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f\" returns successfully" Sep 12 10:13:20.238047 systemd[1]: cri-containerd-fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f.scope: Deactivated successfully. Sep 12 10:13:20.270505 containerd[1490]: time="2025-09-12T10:13:20.270422835Z" level=info msg="shim disconnected" id=fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f namespace=k8s.io Sep 12 10:13:20.270505 containerd[1490]: time="2025-09-12T10:13:20.270506902Z" level=warning msg="cleaning up after shim disconnected" id=fe27d1d0664cbca6b845068cdf2d65d72dfa3b721197c24880318fa41527ea4f namespace=k8s.io Sep 12 10:13:20.270894 containerd[1490]: time="2025-09-12T10:13:20.270522420Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:20.288963 containerd[1490]: time="2025-09-12T10:13:20.287513202Z" level=warning msg="cleanup warnings time=\"2025-09-12T10:13:20Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 12 10:13:20.891796 kubelet[2590]: I0912 10:13:20.891735 2590 setters.go:602] "Node became not ready" node="ci-4230.2.2-n-dc4800d201" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-12T10:13:20Z","lastTransitionTime":"2025-09-12T10:13:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 12 10:13:21.153607 kubelet[2590]: E0912 10:13:21.151643 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:21.155984 containerd[1490]: time="2025-09-12T10:13:21.155946354Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 12 10:13:21.178882 containerd[1490]: time="2025-09-12T10:13:21.178802858Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce\"" Sep 12 10:13:21.181513 containerd[1490]: time="2025-09-12T10:13:21.180824009Z" level=info msg="StartContainer for \"cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce\"" Sep 12 10:13:21.225818 systemd[1]: Started cri-containerd-cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce.scope - libcontainer container cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce. Sep 12 10:13:21.268592 containerd[1490]: time="2025-09-12T10:13:21.268543662Z" level=info msg="StartContainer for \"cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce\" returns successfully" Sep 12 10:13:21.278732 systemd[1]: cri-containerd-cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce.scope: Deactivated successfully. Sep 12 10:13:21.306489 containerd[1490]: time="2025-09-12T10:13:21.306422981Z" level=info msg="shim disconnected" id=cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce namespace=k8s.io Sep 12 10:13:21.306489 containerd[1490]: time="2025-09-12T10:13:21.306481287Z" level=warning msg="cleaning up after shim disconnected" id=cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce namespace=k8s.io Sep 12 10:13:21.306489 containerd[1490]: time="2025-09-12T10:13:21.306489837Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:21.742632 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cd615517f66a93265cd45d42b3873d6449f70aa4a6652b218ee60dbaf1e06bce-rootfs.mount: Deactivated successfully. Sep 12 10:13:21.748946 kubelet[2590]: E0912 10:13:21.748823 2590 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-65q6r" podUID="8dffd854-3556-4ce2-94b0-eb318a35c5e8" Sep 12 10:13:22.155429 kubelet[2590]: E0912 10:13:22.155070 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:22.159720 containerd[1490]: time="2025-09-12T10:13:22.157945300Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 12 10:13:22.178646 containerd[1490]: time="2025-09-12T10:13:22.177094135Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499\"" Sep 12 10:13:22.179696 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3286678329.mount: Deactivated successfully. Sep 12 10:13:22.181133 containerd[1490]: time="2025-09-12T10:13:22.179882161Z" level=info msg="StartContainer for \"eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499\"" Sep 12 10:13:22.229907 systemd[1]: Started cri-containerd-eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499.scope - libcontainer container eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499. Sep 12 10:13:22.262560 containerd[1490]: time="2025-09-12T10:13:22.262368962Z" level=info msg="StartContainer for \"eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499\" returns successfully" Sep 12 10:13:22.263957 systemd[1]: cri-containerd-eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499.scope: Deactivated successfully. Sep 12 10:13:22.291359 containerd[1490]: time="2025-09-12T10:13:22.291261181Z" level=info msg="shim disconnected" id=eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499 namespace=k8s.io Sep 12 10:13:22.291359 containerd[1490]: time="2025-09-12T10:13:22.291359312Z" level=warning msg="cleaning up after shim disconnected" id=eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499 namespace=k8s.io Sep 12 10:13:22.291359 containerd[1490]: time="2025-09-12T10:13:22.291368535Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 12 10:13:22.308767 containerd[1490]: time="2025-09-12T10:13:22.308689693Z" level=warning msg="cleanup warnings time=\"2025-09-12T10:13:22Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 12 10:13:22.742522 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-eee13ffa08e676363cbbdfecf1265e32d0f297f072ba8fb5b85b0512d244f499-rootfs.mount: Deactivated successfully. Sep 12 10:13:23.160036 kubelet[2590]: E0912 10:13:23.159985 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:23.163756 containerd[1490]: time="2025-09-12T10:13:23.163694101Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 12 10:13:23.186709 containerd[1490]: time="2025-09-12T10:13:23.186644272Z" level=info msg="CreateContainer within sandbox \"4fe3ebb044f3843b4c6ccf5a51efd902c8ab93cc67d1fd8757e54e23ca099db5\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5\"" Sep 12 10:13:23.188028 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3852889612.mount: Deactivated successfully. Sep 12 10:13:23.189596 containerd[1490]: time="2025-09-12T10:13:23.188435050Z" level=info msg="StartContainer for \"af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5\"" Sep 12 10:13:23.231791 systemd[1]: Started cri-containerd-af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5.scope - libcontainer container af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5. Sep 12 10:13:23.269657 containerd[1490]: time="2025-09-12T10:13:23.269439828Z" level=info msg="StartContainer for \"af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5\" returns successfully" Sep 12 10:13:23.724240 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 12 10:13:23.742736 systemd[1]: run-containerd-runc-k8s.io-af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5-runc.ufzgZK.mount: Deactivated successfully. Sep 12 10:13:23.749668 kubelet[2590]: E0912 10:13:23.748685 2590 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-65q6r" podUID="8dffd854-3556-4ce2-94b0-eb318a35c5e8" Sep 12 10:13:24.165740 kubelet[2590]: E0912 10:13:24.165299 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:24.185363 kubelet[2590]: I0912 10:13:24.184771 2590 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-pgmnl" podStartSLOduration=7.184749742 podStartE2EDuration="7.184749742s" podCreationTimestamp="2025-09-12 10:13:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 10:13:24.182862869 +0000 UTC m=+105.575603987" watchObservedRunningTime="2025-09-12 10:13:24.184749742 +0000 UTC m=+105.577490869" Sep 12 10:13:25.749917 kubelet[2590]: E0912 10:13:25.748548 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:25.872581 kubelet[2590]: E0912 10:13:25.872071 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:26.824652 systemd[1]: run-containerd-runc-k8s.io-af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5-runc.PvS3VH.mount: Deactivated successfully. Sep 12 10:13:27.057237 systemd-networkd[1378]: lxc_health: Link UP Sep 12 10:13:27.057507 systemd-networkd[1378]: lxc_health: Gained carrier Sep 12 10:13:27.875330 kubelet[2590]: E0912 10:13:27.872784 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:28.149752 systemd-networkd[1378]: lxc_health: Gained IPv6LL Sep 12 10:13:28.179155 kubelet[2590]: E0912 10:13:28.178942 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:29.180445 kubelet[2590]: E0912 10:13:29.180396 2590 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Sep 12 10:13:29.358613 systemd[1]: run-containerd-runc-k8s.io-af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5-runc.VMTJOw.mount: Deactivated successfully. Sep 12 10:13:29.428712 kubelet[2590]: E0912 10:13:29.428548 2590 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 127.0.0.1:51718->127.0.0.1:44765: read tcp 127.0.0.1:51718->127.0.0.1:44765: read: connection reset by peer Sep 12 10:13:33.729295 systemd[1]: run-containerd-runc-k8s.io-af11d1e2c5ca482c336306a3313fa11d46a69c17a2923a2cd4ad8bc6031424f5-runc.hNhyzL.mount: Deactivated successfully. Sep 12 10:13:33.864875 sshd[4423]: Connection closed by 139.178.68.195 port 60234 Sep 12 10:13:33.865883 sshd-session[4419]: pam_unix(sshd:session): session closed for user core Sep 12 10:13:33.870760 systemd-logind[1466]: Session 28 logged out. Waiting for processes to exit. Sep 12 10:13:33.871869 systemd[1]: sshd@27-164.92.125.86:22-139.178.68.195:60234.service: Deactivated successfully. Sep 12 10:13:33.875834 systemd[1]: session-28.scope: Deactivated successfully. Sep 12 10:13:33.878397 systemd-logind[1466]: Removed session 28.