Sep 13 00:49:15.997862 kernel: Linux version 5.15.192-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri Sep 12 23:13:49 -00 2025 Sep 13 00:49:15.997905 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:49:16.002044 kernel: BIOS-provided physical RAM map: Sep 13 00:49:16.002080 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 13 00:49:16.002091 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 13 00:49:16.002100 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 13 00:49:16.002112 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Sep 13 00:49:16.002123 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Sep 13 00:49:16.002143 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Sep 13 00:49:16.002153 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 13 00:49:16.002164 kernel: NX (Execute Disable) protection: active Sep 13 00:49:16.002173 kernel: SMBIOS 2.8 present. Sep 13 00:49:16.002182 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Sep 13 00:49:16.002193 kernel: Hypervisor detected: KVM Sep 13 00:49:16.002207 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 13 00:49:16.002222 kernel: kvm-clock: cpu 0, msr 6819f001, primary cpu clock Sep 13 00:49:16.002233 kernel: kvm-clock: using sched offset of 3900017588 cycles Sep 13 00:49:16.002246 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 13 00:49:16.002266 kernel: tsc: Detected 2000.000 MHz processor Sep 13 00:49:16.002278 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 13 00:49:16.002291 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 13 00:49:16.002301 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Sep 13 00:49:16.002312 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 13 00:49:16.002328 kernel: ACPI: Early table checksum verification disabled Sep 13 00:49:16.002339 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Sep 13 00:49:16.002351 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002363 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002375 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002398 kernel: ACPI: FACS 0x000000007FFE0000 000040 Sep 13 00:49:16.002409 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002421 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002434 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002450 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 13 00:49:16.002462 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Sep 13 00:49:16.002474 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Sep 13 00:49:16.002485 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Sep 13 00:49:16.002499 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Sep 13 00:49:16.002511 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Sep 13 00:49:16.002522 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Sep 13 00:49:16.002533 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Sep 13 00:49:16.002555 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 13 00:49:16.002568 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 13 00:49:16.002580 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Sep 13 00:49:16.002592 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Sep 13 00:49:16.002604 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Sep 13 00:49:16.002616 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Sep 13 00:49:16.002633 kernel: Zone ranges: Sep 13 00:49:16.002644 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 13 00:49:16.002656 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Sep 13 00:49:16.002668 kernel: Normal empty Sep 13 00:49:16.002680 kernel: Movable zone start for each node Sep 13 00:49:16.002691 kernel: Early memory node ranges Sep 13 00:49:16.002702 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 13 00:49:16.002713 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Sep 13 00:49:16.002725 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Sep 13 00:49:16.002741 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 13 00:49:16.002760 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 13 00:49:16.002771 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Sep 13 00:49:16.002783 kernel: ACPI: PM-Timer IO Port: 0x608 Sep 13 00:49:16.002796 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 13 00:49:16.002808 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 13 00:49:16.002819 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Sep 13 00:49:16.002830 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 13 00:49:16.002841 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 13 00:49:16.002857 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 13 00:49:16.002874 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 13 00:49:16.002887 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 13 00:49:16.002900 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 13 00:49:16.002912 kernel: TSC deadline timer available Sep 13 00:49:16.002941 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 13 00:49:16.002954 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Sep 13 00:49:16.002966 kernel: Booting paravirtualized kernel on KVM Sep 13 00:49:16.002979 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 13 00:49:16.002996 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Sep 13 00:49:16.003008 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 Sep 13 00:49:16.003020 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 Sep 13 00:49:16.003032 kernel: pcpu-alloc: [0] 0 1 Sep 13 00:49:16.003044 kernel: kvm-guest: stealtime: cpu 0, msr 7dc1c0c0 Sep 13 00:49:16.003057 kernel: kvm-guest: PV spinlocks disabled, no host support Sep 13 00:49:16.003069 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Sep 13 00:49:16.003196 kernel: Policy zone: DMA32 Sep 13 00:49:16.003220 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:49:16.003239 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 00:49:16.003252 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 00:49:16.003266 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 13 00:49:16.003277 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 00:49:16.003289 kernel: Memory: 1973276K/2096612K available (12295K kernel code, 2276K rwdata, 13732K rodata, 47492K init, 4088K bss, 123076K reserved, 0K cma-reserved) Sep 13 00:49:16.003300 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 13 00:49:16.003312 kernel: Kernel/User page tables isolation: enabled Sep 13 00:49:16.003325 kernel: ftrace: allocating 34614 entries in 136 pages Sep 13 00:49:16.003354 kernel: ftrace: allocated 136 pages with 2 groups Sep 13 00:49:16.003367 kernel: rcu: Hierarchical RCU implementation. Sep 13 00:49:16.003380 kernel: rcu: RCU event tracing is enabled. Sep 13 00:49:16.003391 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 13 00:49:16.003403 kernel: Rude variant of Tasks RCU enabled. Sep 13 00:49:16.003415 kernel: Tracing variant of Tasks RCU enabled. Sep 13 00:49:16.003428 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 00:49:16.003441 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 13 00:49:16.003453 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 13 00:49:16.003474 kernel: random: crng init done Sep 13 00:49:16.003493 kernel: Console: colour VGA+ 80x25 Sep 13 00:49:16.003504 kernel: printk: console [tty0] enabled Sep 13 00:49:16.003517 kernel: printk: console [ttyS0] enabled Sep 13 00:49:16.003529 kernel: ACPI: Core revision 20210730 Sep 13 00:49:16.003541 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Sep 13 00:49:16.003552 kernel: APIC: Switch to symmetric I/O mode setup Sep 13 00:49:16.003563 kernel: x2apic enabled Sep 13 00:49:16.003575 kernel: Switched APIC routing to physical x2apic. Sep 13 00:49:16.003586 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Sep 13 00:49:16.003603 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x39a85c9bff6, max_idle_ns: 881590591483 ns Sep 13 00:49:16.003615 kernel: Calibrating delay loop (skipped) preset value.. 4000.00 BogoMIPS (lpj=2000000) Sep 13 00:49:16.003640 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Sep 13 00:49:16.003652 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Sep 13 00:49:16.003664 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 13 00:49:16.003676 kernel: Spectre V2 : Mitigation: Retpolines Sep 13 00:49:16.003688 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 13 00:49:16.003700 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Sep 13 00:49:16.003717 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 13 00:49:16.003741 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Sep 13 00:49:16.003755 kernel: MDS: Mitigation: Clear CPU buffers Sep 13 00:49:16.003771 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 13 00:49:16.003784 kernel: active return thunk: its_return_thunk Sep 13 00:49:16.003797 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 13 00:49:16.003810 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 13 00:49:16.003824 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 13 00:49:16.003837 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 13 00:49:16.003849 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 13 00:49:16.003866 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Sep 13 00:49:16.003879 kernel: Freeing SMP alternatives memory: 32K Sep 13 00:49:16.003890 kernel: pid_max: default: 32768 minimum: 301 Sep 13 00:49:16.003903 kernel: LSM: Security Framework initializing Sep 13 00:49:16.003933 kernel: SELinux: Initializing. Sep 13 00:49:16.003964 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 00:49:16.003977 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 00:49:16.003994 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Sep 13 00:49:16.004006 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Sep 13 00:49:16.004018 kernel: signal: max sigframe size: 1776 Sep 13 00:49:16.004029 kernel: rcu: Hierarchical SRCU implementation. Sep 13 00:49:16.004042 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 13 00:49:16.004054 kernel: smp: Bringing up secondary CPUs ... Sep 13 00:49:16.004066 kernel: x86: Booting SMP configuration: Sep 13 00:49:16.004078 kernel: .... node #0, CPUs: #1 Sep 13 00:49:16.004091 kernel: kvm-clock: cpu 1, msr 6819f041, secondary cpu clock Sep 13 00:49:16.004107 kernel: kvm-guest: stealtime: cpu 1, msr 7dd1c0c0 Sep 13 00:49:16.004122 kernel: smp: Brought up 1 node, 2 CPUs Sep 13 00:49:16.004135 kernel: smpboot: Max logical packages: 1 Sep 13 00:49:16.004147 kernel: smpboot: Total of 2 processors activated (8000.00 BogoMIPS) Sep 13 00:49:16.004159 kernel: devtmpfs: initialized Sep 13 00:49:16.004172 kernel: x86/mm: Memory block size: 128MB Sep 13 00:49:16.004185 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 00:49:16.004199 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 13 00:49:16.004213 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 00:49:16.004230 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 00:49:16.004243 kernel: audit: initializing netlink subsys (disabled) Sep 13 00:49:16.004255 kernel: audit: type=2000 audit(1757724555.178:1): state=initialized audit_enabled=0 res=1 Sep 13 00:49:16.004269 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 00:49:16.004283 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 13 00:49:16.004296 kernel: cpuidle: using governor menu Sep 13 00:49:16.004318 kernel: ACPI: bus type PCI registered Sep 13 00:49:16.004332 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 00:49:16.004344 kernel: dca service started, version 1.12.1 Sep 13 00:49:16.004359 kernel: PCI: Using configuration type 1 for base access Sep 13 00:49:16.004371 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 13 00:49:16.004384 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 00:49:16.004395 kernel: ACPI: Added _OSI(Module Device) Sep 13 00:49:16.004407 kernel: ACPI: Added _OSI(Processor Device) Sep 13 00:49:16.004419 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 00:49:16.004430 kernel: ACPI: Added _OSI(Linux-Dell-Video) Sep 13 00:49:16.004443 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Sep 13 00:49:16.004455 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Sep 13 00:49:16.004472 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 13 00:49:16.004484 kernel: ACPI: Interpreter enabled Sep 13 00:49:16.004499 kernel: ACPI: PM: (supports S0 S5) Sep 13 00:49:16.004512 kernel: ACPI: Using IOAPIC for interrupt routing Sep 13 00:49:16.004526 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 13 00:49:16.004538 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Sep 13 00:49:16.004552 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 13 00:49:16.004898 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 13 00:49:16.005054 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Sep 13 00:49:16.005074 kernel: acpiphp: Slot [3] registered Sep 13 00:49:16.005088 kernel: acpiphp: Slot [4] registered Sep 13 00:49:16.005100 kernel: acpiphp: Slot [5] registered Sep 13 00:49:16.005114 kernel: acpiphp: Slot [6] registered Sep 13 00:49:16.005127 kernel: acpiphp: Slot [7] registered Sep 13 00:49:16.005140 kernel: acpiphp: Slot [8] registered Sep 13 00:49:16.005153 kernel: acpiphp: Slot [9] registered Sep 13 00:49:16.005166 kernel: acpiphp: Slot [10] registered Sep 13 00:49:16.005184 kernel: acpiphp: Slot [11] registered Sep 13 00:49:16.005196 kernel: acpiphp: Slot [12] registered Sep 13 00:49:16.005210 kernel: acpiphp: Slot [13] registered Sep 13 00:49:16.005224 kernel: acpiphp: Slot [14] registered Sep 13 00:49:16.005237 kernel: acpiphp: Slot [15] registered Sep 13 00:49:16.005252 kernel: acpiphp: Slot [16] registered Sep 13 00:49:16.005265 kernel: acpiphp: Slot [17] registered Sep 13 00:49:16.005278 kernel: acpiphp: Slot [18] registered Sep 13 00:49:16.005294 kernel: acpiphp: Slot [19] registered Sep 13 00:49:16.005309 kernel: acpiphp: Slot [20] registered Sep 13 00:49:16.005321 kernel: acpiphp: Slot [21] registered Sep 13 00:49:16.005332 kernel: acpiphp: Slot [22] registered Sep 13 00:49:16.005345 kernel: acpiphp: Slot [23] registered Sep 13 00:49:16.005357 kernel: acpiphp: Slot [24] registered Sep 13 00:49:16.005369 kernel: acpiphp: Slot [25] registered Sep 13 00:49:16.005380 kernel: acpiphp: Slot [26] registered Sep 13 00:49:16.005393 kernel: acpiphp: Slot [27] registered Sep 13 00:49:16.005407 kernel: acpiphp: Slot [28] registered Sep 13 00:49:16.005421 kernel: acpiphp: Slot [29] registered Sep 13 00:49:16.005434 kernel: acpiphp: Slot [30] registered Sep 13 00:49:16.005443 kernel: acpiphp: Slot [31] registered Sep 13 00:49:16.005451 kernel: PCI host bridge to bus 0000:00 Sep 13 00:49:16.005587 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 13 00:49:16.005677 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 13 00:49:16.005762 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 13 00:49:16.005847 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 13 00:49:16.008162 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Sep 13 00:49:16.008293 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 13 00:49:16.008453 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Sep 13 00:49:16.008568 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Sep 13 00:49:16.008673 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Sep 13 00:49:16.008774 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Sep 13 00:49:16.008876 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Sep 13 00:49:16.009038 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Sep 13 00:49:16.009178 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Sep 13 00:49:16.009315 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Sep 13 00:49:16.009480 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Sep 13 00:49:16.009578 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Sep 13 00:49:16.009682 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Sep 13 00:49:16.009780 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Sep 13 00:49:16.009874 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Sep 13 00:49:16.010025 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Sep 13 00:49:16.010155 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Sep 13 00:49:16.010298 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Sep 13 00:49:16.010442 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Sep 13 00:49:16.010575 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Sep 13 00:49:16.010676 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 13 00:49:16.010800 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Sep 13 00:49:16.010894 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Sep 13 00:49:16.011039 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Sep 13 00:49:16.011176 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Sep 13 00:49:16.011312 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 13 00:49:16.011413 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Sep 13 00:49:16.011508 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Sep 13 00:49:16.011602 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Sep 13 00:49:16.011716 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Sep 13 00:49:16.011830 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Sep 13 00:49:16.015000 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Sep 13 00:49:16.015292 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Sep 13 00:49:16.015471 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Sep 13 00:49:16.015629 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Sep 13 00:49:16.015768 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Sep 13 00:49:16.015905 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Sep 13 00:49:16.017192 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Sep 13 00:49:16.017350 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Sep 13 00:49:16.017497 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Sep 13 00:49:16.017648 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Sep 13 00:49:16.017821 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Sep 13 00:49:16.017993 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Sep 13 00:49:16.018143 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Sep 13 00:49:16.018161 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 13 00:49:16.018176 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 13 00:49:16.018202 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 13 00:49:16.018228 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 13 00:49:16.018242 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 13 00:49:16.018257 kernel: iommu: Default domain type: Translated Sep 13 00:49:16.018272 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 13 00:49:16.018421 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Sep 13 00:49:16.018607 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 13 00:49:16.018760 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Sep 13 00:49:16.018780 kernel: vgaarb: loaded Sep 13 00:49:16.018795 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 13 00:49:16.018815 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 13 00:49:16.018831 kernel: PTP clock support registered Sep 13 00:49:16.018845 kernel: PCI: Using ACPI for IRQ routing Sep 13 00:49:16.018860 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 13 00:49:16.018874 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 13 00:49:16.018889 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Sep 13 00:49:16.018903 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Sep 13 00:49:16.018933 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Sep 13 00:49:16.021999 kernel: clocksource: Switched to clocksource kvm-clock Sep 13 00:49:16.022017 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 00:49:16.022028 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 00:49:16.022037 kernel: pnp: PnP ACPI init Sep 13 00:49:16.022046 kernel: pnp: PnP ACPI: found 4 devices Sep 13 00:49:16.022055 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 13 00:49:16.022065 kernel: NET: Registered PF_INET protocol family Sep 13 00:49:16.022074 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 13 00:49:16.022083 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 13 00:49:16.022094 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 00:49:16.022103 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 13 00:49:16.022112 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) Sep 13 00:49:16.022121 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 13 00:49:16.022129 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 00:49:16.022138 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 00:49:16.022147 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 00:49:16.022156 kernel: NET: Registered PF_XDP protocol family Sep 13 00:49:16.022309 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 13 00:49:16.022404 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 13 00:49:16.022486 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 13 00:49:16.022569 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 13 00:49:16.022651 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Sep 13 00:49:16.022754 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Sep 13 00:49:16.022852 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 13 00:49:16.022996 kernel: pci 0000:00:01.0: Activating ISA DMA hang workarounds Sep 13 00:49:16.023008 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 13 00:49:16.023129 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x740 took 37855 usecs Sep 13 00:49:16.023146 kernel: PCI: CLS 0 bytes, default 64 Sep 13 00:49:16.023157 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 13 00:49:16.023166 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x39a85c9bff6, max_idle_ns: 881590591483 ns Sep 13 00:49:16.023179 kernel: Initialise system trusted keyrings Sep 13 00:49:16.023193 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 13 00:49:16.023206 kernel: Key type asymmetric registered Sep 13 00:49:16.023218 kernel: Asymmetric key parser 'x509' registered Sep 13 00:49:16.023230 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 13 00:49:16.023247 kernel: io scheduler mq-deadline registered Sep 13 00:49:16.023259 kernel: io scheduler kyber registered Sep 13 00:49:16.023270 kernel: io scheduler bfq registered Sep 13 00:49:16.023282 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 13 00:49:16.023296 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Sep 13 00:49:16.023310 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Sep 13 00:49:16.023324 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Sep 13 00:49:16.023337 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 00:49:16.023346 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 13 00:49:16.023358 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 13 00:49:16.023367 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 13 00:49:16.023376 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 13 00:49:16.023512 kernel: rtc_cmos 00:03: RTC can wake from S4 Sep 13 00:49:16.023525 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 13 00:49:16.023610 kernel: rtc_cmos 00:03: registered as rtc0 Sep 13 00:49:16.023695 kernel: rtc_cmos 00:03: setting system clock to 2025-09-13T00:49:15 UTC (1757724555) Sep 13 00:49:16.023811 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Sep 13 00:49:16.023828 kernel: intel_pstate: CPU model not supported Sep 13 00:49:16.023841 kernel: NET: Registered PF_INET6 protocol family Sep 13 00:49:16.023854 kernel: Segment Routing with IPv6 Sep 13 00:49:16.023866 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 00:49:16.023878 kernel: NET: Registered PF_PACKET protocol family Sep 13 00:49:16.023891 kernel: Key type dns_resolver registered Sep 13 00:49:16.023903 kernel: IPI shorthand broadcast: enabled Sep 13 00:49:16.023933 kernel: sched_clock: Marking stable (764009135, 173716089)->(1096312628, -158587404) Sep 13 00:49:16.023946 kernel: registered taskstats version 1 Sep 13 00:49:16.023963 kernel: Loading compiled-in X.509 certificates Sep 13 00:49:16.023974 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.192-flatcar: d4931373bb0d9b9f95da11f02ae07d3649cc6c37' Sep 13 00:49:16.023987 kernel: Key type .fscrypt registered Sep 13 00:49:16.023999 kernel: Key type fscrypt-provisioning registered Sep 13 00:49:16.024010 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 13 00:49:16.024022 kernel: ima: Allocated hash algorithm: sha1 Sep 13 00:49:16.024035 kernel: ima: No architecture policies found Sep 13 00:49:16.024049 kernel: clk: Disabling unused clocks Sep 13 00:49:16.024067 kernel: Freeing unused kernel image (initmem) memory: 47492K Sep 13 00:49:16.024078 kernel: Write protecting the kernel read-only data: 28672k Sep 13 00:49:16.024090 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Sep 13 00:49:16.024102 kernel: Freeing unused kernel image (rodata/data gap) memory: 604K Sep 13 00:49:16.024113 kernel: Run /init as init process Sep 13 00:49:16.024125 kernel: with arguments: Sep 13 00:49:16.024166 kernel: /init Sep 13 00:49:16.024185 kernel: with environment: Sep 13 00:49:16.024198 kernel: HOME=/ Sep 13 00:49:16.024210 kernel: TERM=linux Sep 13 00:49:16.024225 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 00:49:16.024243 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:49:16.024260 systemd[1]: Detected virtualization kvm. Sep 13 00:49:16.024278 systemd[1]: Detected architecture x86-64. Sep 13 00:49:16.024290 systemd[1]: Running in initrd. Sep 13 00:49:16.024300 systemd[1]: No hostname configured, using default hostname. Sep 13 00:49:16.024309 systemd[1]: Hostname set to . Sep 13 00:49:16.024322 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:49:16.024334 systemd[1]: Queued start job for default target initrd.target. Sep 13 00:49:16.024348 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:49:16.024363 systemd[1]: Reached target cryptsetup.target. Sep 13 00:49:16.024376 systemd[1]: Reached target paths.target. Sep 13 00:49:16.024388 systemd[1]: Reached target slices.target. Sep 13 00:49:16.024400 systemd[1]: Reached target swap.target. Sep 13 00:49:16.024412 systemd[1]: Reached target timers.target. Sep 13 00:49:16.024430 systemd[1]: Listening on iscsid.socket. Sep 13 00:49:16.024444 systemd[1]: Listening on iscsiuio.socket. Sep 13 00:49:16.024459 systemd[1]: Listening on systemd-journald-audit.socket. Sep 13 00:49:16.024472 systemd[1]: Listening on systemd-journald-dev-log.socket. Sep 13 00:49:16.024485 systemd[1]: Listening on systemd-journald.socket. Sep 13 00:49:16.024501 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:49:16.024516 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:49:16.024531 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:49:16.024549 systemd[1]: Reached target sockets.target. Sep 13 00:49:16.024564 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:49:16.024582 systemd[1]: Finished network-cleanup.service. Sep 13 00:49:16.024597 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 00:49:16.024609 systemd[1]: Starting systemd-journald.service... Sep 13 00:49:16.024621 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:49:16.024631 systemd[1]: Starting systemd-resolved.service... Sep 13 00:49:16.024640 systemd[1]: Starting systemd-vconsole-setup.service... Sep 13 00:49:16.024650 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:49:16.024659 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 00:49:16.024676 systemd-journald[185]: Journal started Sep 13 00:49:16.024752 systemd-journald[185]: Runtime Journal (/run/log/journal/9c81bef432b74dc0818e41e91c9d4566) is 4.9M, max 39.5M, 34.5M free. Sep 13 00:49:16.010988 systemd-modules-load[186]: Inserted module 'overlay' Sep 13 00:49:16.068157 systemd[1]: Started systemd-journald.service. Sep 13 00:49:16.068188 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 00:49:16.068202 kernel: audit: type=1130 audit(1757724556.059:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.059000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.030865 systemd-resolved[187]: Positive Trust Anchors: Sep 13 00:49:16.030876 systemd-resolved[187]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:49:16.030912 systemd-resolved[187]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:49:16.077151 kernel: audit: type=1130 audit(1757724556.072:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.072000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.034321 systemd-resolved[187]: Defaulting to hostname 'linux'. Sep 13 00:49:16.079529 kernel: Bridge firewalling registered Sep 13 00:49:16.073229 systemd[1]: Started systemd-resolved.service. Sep 13 00:49:16.077550 systemd-modules-load[186]: Inserted module 'br_netfilter' Sep 13 00:49:16.080000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.085172 kernel: audit: type=1130 audit(1757724556.080:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.085142 systemd[1]: Finished systemd-vconsole-setup.service. Sep 13 00:49:16.085000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.090430 systemd[1]: Reached target nss-lookup.target. Sep 13 00:49:16.091530 kernel: audit: type=1130 audit(1757724556.085:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.092517 systemd[1]: Starting dracut-cmdline-ask.service... Sep 13 00:49:16.094761 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 13 00:49:16.104280 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 13 00:49:16.109392 kernel: audit: type=1130 audit(1757724556.103:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.103000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.112019 kernel: SCSI subsystem initialized Sep 13 00:49:16.119779 systemd[1]: Finished dracut-cmdline-ask.service. Sep 13 00:49:16.125163 kernel: audit: type=1130 audit(1757724556.119:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.119000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.125485 systemd[1]: Starting dracut-cmdline.service... Sep 13 00:49:16.136306 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 00:49:16.136386 kernel: device-mapper: uevent: version 1.0.3 Sep 13 00:49:16.136417 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Sep 13 00:49:16.141826 dracut-cmdline[202]: dracut-dracut-053 Sep 13 00:49:16.142518 systemd-modules-load[186]: Inserted module 'dm_multipath' Sep 13 00:49:16.143275 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:49:16.156035 kernel: audit: type=1130 audit(1757724556.143:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.143000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.156195 dracut-cmdline[202]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:49:16.145014 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:49:16.167890 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:49:16.174157 kernel: audit: type=1130 audit(1757724556.167:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.167000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.243974 kernel: Loading iSCSI transport class v2.0-870. Sep 13 00:49:16.267988 kernel: iscsi: registered transport (tcp) Sep 13 00:49:16.309412 kernel: iscsi: registered transport (qla4xxx) Sep 13 00:49:16.309513 kernel: QLogic iSCSI HBA Driver Sep 13 00:49:16.367732 systemd[1]: Finished dracut-cmdline.service. Sep 13 00:49:16.367000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.375995 kernel: audit: type=1130 audit(1757724556.367:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.374290 systemd[1]: Starting dracut-pre-udev.service... Sep 13 00:49:16.438991 kernel: raid6: avx2x4 gen() 26844 MB/s Sep 13 00:49:16.456009 kernel: raid6: avx2x4 xor() 8986 MB/s Sep 13 00:49:16.474005 kernel: raid6: avx2x2 gen() 24073 MB/s Sep 13 00:49:16.490988 kernel: raid6: avx2x2 xor() 15936 MB/s Sep 13 00:49:16.508999 kernel: raid6: avx2x1 gen() 19382 MB/s Sep 13 00:49:16.525993 kernel: raid6: avx2x1 xor() 12634 MB/s Sep 13 00:49:16.543005 kernel: raid6: sse2x4 gen() 11868 MB/s Sep 13 00:49:16.559990 kernel: raid6: sse2x4 xor() 5737 MB/s Sep 13 00:49:16.576988 kernel: raid6: sse2x2 gen() 9500 MB/s Sep 13 00:49:16.595001 kernel: raid6: sse2x2 xor() 6414 MB/s Sep 13 00:49:16.611994 kernel: raid6: sse2x1 gen() 7835 MB/s Sep 13 00:49:16.630022 kernel: raid6: sse2x1 xor() 4933 MB/s Sep 13 00:49:16.630107 kernel: raid6: using algorithm avx2x4 gen() 26844 MB/s Sep 13 00:49:16.630126 kernel: raid6: .... xor() 8986 MB/s, rmw enabled Sep 13 00:49:16.630996 kernel: raid6: using avx2x2 recovery algorithm Sep 13 00:49:16.648993 kernel: xor: automatically using best checksumming function avx Sep 13 00:49:16.780955 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Sep 13 00:49:16.792982 systemd[1]: Finished dracut-pre-udev.service. Sep 13 00:49:16.792000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.793000 audit: BPF prog-id=7 op=LOAD Sep 13 00:49:16.793000 audit: BPF prog-id=8 op=LOAD Sep 13 00:49:16.795874 systemd[1]: Starting systemd-udevd.service... Sep 13 00:49:16.812184 systemd-udevd[384]: Using default interface naming scheme 'v252'. Sep 13 00:49:16.817353 systemd[1]: Started systemd-udevd.service. Sep 13 00:49:16.817000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.819198 systemd[1]: Starting dracut-pre-trigger.service... Sep 13 00:49:16.835958 dracut-pre-trigger[385]: rd.md=0: removing MD RAID activation Sep 13 00:49:16.877760 systemd[1]: Finished dracut-pre-trigger.service. Sep 13 00:49:16.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.879514 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:49:16.930000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:16.931214 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:49:16.995318 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Sep 13 00:49:17.078400 kernel: scsi host0: Virtio SCSI HBA Sep 13 00:49:17.078599 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 00:49:17.078613 kernel: GPT:9289727 != 125829119 Sep 13 00:49:17.078624 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 00:49:17.078635 kernel: GPT:9289727 != 125829119 Sep 13 00:49:17.078646 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 00:49:17.078657 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:49:17.078673 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 00:49:17.078684 kernel: AVX2 version of gcm_enc/dec engaged. Sep 13 00:49:17.083471 kernel: virtio_blk virtio5: [vdb] 976 512-byte logical blocks (500 kB/488 KiB) Sep 13 00:49:17.090685 kernel: AES CTR mode by8 optimization enabled Sep 13 00:49:17.116519 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (440) Sep 13 00:49:17.115942 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Sep 13 00:49:17.117678 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Sep 13 00:49:17.129976 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Sep 13 00:49:17.218428 kernel: libata version 3.00 loaded. Sep 13 00:49:17.218454 kernel: ACPI: bus type USB registered Sep 13 00:49:17.218466 kernel: usbcore: registered new interface driver usbfs Sep 13 00:49:17.218486 kernel: usbcore: registered new interface driver hub Sep 13 00:49:17.218497 kernel: usbcore: registered new device driver usb Sep 13 00:49:17.218508 kernel: ata_piix 0000:00:01.1: version 2.13 Sep 13 00:49:17.218677 kernel: ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver Sep 13 00:49:17.218689 kernel: ehci-pci: EHCI PCI platform driver Sep 13 00:49:17.218700 kernel: scsi host1: ata_piix Sep 13 00:49:17.218826 kernel: scsi host2: ata_piix Sep 13 00:49:17.218975 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Sep 13 00:49:17.218992 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Sep 13 00:49:17.219004 kernel: uhci_hcd: USB Universal Host Controller Interface driver Sep 13 00:49:17.219015 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Sep 13 00:49:17.219213 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Sep 13 00:49:17.219321 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Sep 13 00:49:17.219424 kernel: uhci_hcd 0000:00:01.2: irq 11, io base 0x0000c180 Sep 13 00:49:17.219527 kernel: hub 1-0:1.0: USB hub found Sep 13 00:49:17.219668 kernel: hub 1-0:1.0: 2 ports detected Sep 13 00:49:17.228221 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Sep 13 00:49:17.237116 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:49:17.240399 systemd[1]: Starting disk-uuid.service... Sep 13 00:49:17.251518 disk-uuid[503]: Primary Header is updated. Sep 13 00:49:17.251518 disk-uuid[503]: Secondary Entries is updated. Sep 13 00:49:17.251518 disk-uuid[503]: Secondary Header is updated. Sep 13 00:49:17.257973 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:49:17.264991 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:49:18.265953 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 13 00:49:18.266845 disk-uuid[505]: The operation has completed successfully. Sep 13 00:49:18.317557 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 00:49:18.318955 systemd[1]: Finished disk-uuid.service. Sep 13 00:49:18.319000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.319000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.327061 systemd[1]: Starting verity-setup.service... Sep 13 00:49:18.349961 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 13 00:49:18.404482 systemd[1]: Found device dev-mapper-usr.device. Sep 13 00:49:18.406827 systemd[1]: Mounting sysusr-usr.mount... Sep 13 00:49:18.409757 systemd[1]: Finished verity-setup.service. Sep 13 00:49:18.409000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.514958 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Sep 13 00:49:18.515901 systemd[1]: Mounted sysusr-usr.mount. Sep 13 00:49:18.516541 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Sep 13 00:49:18.517442 systemd[1]: Starting ignition-setup.service... Sep 13 00:49:18.519220 systemd[1]: Starting parse-ip-for-networkd.service... Sep 13 00:49:18.534943 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 13 00:49:18.535010 kernel: BTRFS info (device vda6): using free space tree Sep 13 00:49:18.535024 kernel: BTRFS info (device vda6): has skinny extents Sep 13 00:49:18.550388 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 13 00:49:18.558668 systemd[1]: Finished ignition-setup.service. Sep 13 00:49:18.558000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.560839 systemd[1]: Starting ignition-fetch-offline.service... Sep 13 00:49:18.678354 systemd[1]: Finished parse-ip-for-networkd.service. Sep 13 00:49:18.678000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.679000 audit: BPF prog-id=9 op=LOAD Sep 13 00:49:18.680823 systemd[1]: Starting systemd-networkd.service... Sep 13 00:49:18.708035 ignition[604]: Ignition 2.14.0 Sep 13 00:49:18.708052 ignition[604]: Stage: fetch-offline Sep 13 00:49:18.708129 ignition[604]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:18.708169 ignition[604]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:18.716119 ignition[604]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:18.716302 ignition[604]: parsed url from cmdline: "" Sep 13 00:49:18.716309 ignition[604]: no config URL provided Sep 13 00:49:18.716319 ignition[604]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 00:49:18.716334 ignition[604]: no config at "/usr/lib/ignition/user.ign" Sep 13 00:49:18.716343 ignition[604]: failed to fetch config: resource requires networking Sep 13 00:49:18.716496 ignition[604]: Ignition finished successfully Sep 13 00:49:18.719000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.719344 systemd[1]: Finished ignition-fetch-offline.service. Sep 13 00:49:18.728373 systemd-networkd[688]: lo: Link UP Sep 13 00:49:18.729319 systemd-networkd[688]: lo: Gained carrier Sep 13 00:49:18.730799 systemd-networkd[688]: Enumeration completed Sep 13 00:49:18.731681 systemd[1]: Started systemd-networkd.service. Sep 13 00:49:18.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.732223 systemd-networkd[688]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 00:49:18.732517 systemd[1]: Reached target network.target. Sep 13 00:49:18.734861 systemd[1]: Starting ignition-fetch.service... Sep 13 00:49:18.737165 systemd-networkd[688]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Sep 13 00:49:18.737539 systemd[1]: Starting iscsiuio.service... Sep 13 00:49:18.757788 ignition[690]: Ignition 2.14.0 Sep 13 00:49:18.757446 systemd-networkd[688]: eth1: Link UP Sep 13 00:49:18.757795 ignition[690]: Stage: fetch Sep 13 00:49:18.757452 systemd-networkd[688]: eth1: Gained carrier Sep 13 00:49:18.757957 ignition[690]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:18.761866 systemd-networkd[688]: eth0: Link UP Sep 13 00:49:18.757977 ignition[690]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:18.761872 systemd-networkd[688]: eth0: Gained carrier Sep 13 00:49:18.760528 ignition[690]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:18.760653 ignition[690]: parsed url from cmdline: "" Sep 13 00:49:18.760657 ignition[690]: no config URL provided Sep 13 00:49:18.760663 ignition[690]: reading system config file "/usr/lib/ignition/user.ign" Sep 13 00:49:18.760671 ignition[690]: no config at "/usr/lib/ignition/user.ign" Sep 13 00:49:18.760702 ignition[690]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Sep 13 00:49:18.777000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.777508 systemd[1]: Started iscsiuio.service. Sep 13 00:49:18.777182 ignition[690]: GET error: Get "http://169.254.169.254/metadata/v1/user-data": dial tcp 169.254.169.254:80: connect: network is unreachable Sep 13 00:49:18.779013 systemd[1]: Starting iscsid.service... Sep 13 00:49:18.783063 systemd-networkd[688]: eth1: DHCPv4 address 10.124.0.35/20 acquired from 169.254.169.253 Sep 13 00:49:18.784389 iscsid[698]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:49:18.784389 iscsid[698]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Sep 13 00:49:18.784389 iscsid[698]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Sep 13 00:49:18.784389 iscsid[698]: If using hardware iscsi like qla4xxx this message can be ignored. Sep 13 00:49:18.784389 iscsid[698]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:49:18.788000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.791856 iscsid[698]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Sep 13 00:49:18.787098 systemd-networkd[688]: eth0: DHCPv4 address 24.144.95.58/20, gateway 24.144.80.1 acquired from 169.254.169.253 Sep 13 00:49:18.788996 systemd[1]: Started iscsid.service. Sep 13 00:49:18.790908 systemd[1]: Starting dracut-initqueue.service... Sep 13 00:49:18.808329 systemd[1]: Finished dracut-initqueue.service. Sep 13 00:49:18.808000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.809091 systemd[1]: Reached target remote-fs-pre.target. Sep 13 00:49:18.810204 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:49:18.811397 systemd[1]: Reached target remote-fs.target. Sep 13 00:49:18.813447 systemd[1]: Starting dracut-pre-mount.service... Sep 13 00:49:18.825000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:18.825653 systemd[1]: Finished dracut-pre-mount.service. Sep 13 00:49:18.977338 ignition[690]: GET http://169.254.169.254/metadata/v1/user-data: attempt #2 Sep 13 00:49:18.992357 ignition[690]: GET result: OK Sep 13 00:49:18.992596 ignition[690]: parsing config with SHA512: 6beebed25e9e13ce609e6689ba4d4fba2f48e0348597a32142ca2f9b6ed881afa2af94b3dca58147437a8bd34f6daeecb5e00bc892105ecc14df6019210451cd Sep 13 00:49:19.001309 unknown[690]: fetched base config from "system" Sep 13 00:49:19.001330 unknown[690]: fetched base config from "system" Sep 13 00:49:19.002064 ignition[690]: fetch: fetch complete Sep 13 00:49:19.001340 unknown[690]: fetched user config from "digitalocean" Sep 13 00:49:19.002072 ignition[690]: fetch: fetch passed Sep 13 00:49:19.004000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.004729 systemd[1]: Finished ignition-fetch.service. Sep 13 00:49:19.002144 ignition[690]: Ignition finished successfully Sep 13 00:49:19.006527 systemd[1]: Starting ignition-kargs.service... Sep 13 00:49:19.021463 ignition[713]: Ignition 2.14.0 Sep 13 00:49:19.021481 ignition[713]: Stage: kargs Sep 13 00:49:19.021665 ignition[713]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:19.021691 ignition[713]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:19.024769 ignition[713]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:19.027586 ignition[713]: kargs: kargs passed Sep 13 00:49:19.027725 ignition[713]: Ignition finished successfully Sep 13 00:49:19.028913 systemd[1]: Finished ignition-kargs.service. Sep 13 00:49:19.028000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.030527 systemd[1]: Starting ignition-disks.service... Sep 13 00:49:19.048904 ignition[719]: Ignition 2.14.0 Sep 13 00:49:19.048940 ignition[719]: Stage: disks Sep 13 00:49:19.049106 ignition[719]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:19.049126 ignition[719]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:19.051293 ignition[719]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:19.053686 ignition[719]: disks: disks passed Sep 13 00:49:19.053760 ignition[719]: Ignition finished successfully Sep 13 00:49:19.055016 systemd[1]: Finished ignition-disks.service. Sep 13 00:49:19.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.056078 systemd[1]: Reached target initrd-root-device.target. Sep 13 00:49:19.056978 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:49:19.058130 systemd[1]: Reached target local-fs.target. Sep 13 00:49:19.059203 systemd[1]: Reached target sysinit.target. Sep 13 00:49:19.060158 systemd[1]: Reached target basic.target. Sep 13 00:49:19.062287 systemd[1]: Starting systemd-fsck-root.service... Sep 13 00:49:19.081238 systemd-fsck[727]: ROOT: clean, 629/553520 files, 56028/553472 blocks Sep 13 00:49:19.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.084950 systemd[1]: Finished systemd-fsck-root.service. Sep 13 00:49:19.086621 systemd[1]: Mounting sysroot.mount... Sep 13 00:49:19.102943 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Sep 13 00:49:19.103556 systemd[1]: Mounted sysroot.mount. Sep 13 00:49:19.105235 systemd[1]: Reached target initrd-root-fs.target. Sep 13 00:49:19.108713 systemd[1]: Mounting sysroot-usr.mount... Sep 13 00:49:19.110818 systemd[1]: Starting flatcar-digitalocean-network.service... Sep 13 00:49:19.113779 systemd[1]: Starting flatcar-metadata-hostname.service... Sep 13 00:49:19.115042 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 00:49:19.116139 systemd[1]: Reached target ignition-diskful.target. Sep 13 00:49:19.121931 systemd[1]: Mounted sysroot-usr.mount. Sep 13 00:49:19.124547 systemd[1]: Starting initrd-setup-root.service... Sep 13 00:49:19.134950 initrd-setup-root[739]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 00:49:19.148214 initrd-setup-root[747]: cut: /sysroot/etc/group: No such file or directory Sep 13 00:49:19.156385 initrd-setup-root[755]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 00:49:19.167700 initrd-setup-root[765]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 00:49:19.252272 coreos-metadata[733]: Sep 13 00:49:19.252 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 13 00:49:19.263645 systemd[1]: Finished initrd-setup-root.service. Sep 13 00:49:19.264000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.266345 systemd[1]: Starting ignition-mount.service... Sep 13 00:49:19.269472 systemd[1]: Starting sysroot-boot.service... Sep 13 00:49:19.277874 coreos-metadata[733]: Sep 13 00:49:19.277 INFO Fetch successful Sep 13 00:49:19.290079 bash[785]: umount: /sysroot/usr/share/oem: not mounted. Sep 13 00:49:19.293213 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Sep 13 00:49:19.293322 systemd[1]: Finished flatcar-digitalocean-network.service. Sep 13 00:49:19.294000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.294000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.303731 coreos-metadata[734]: Sep 13 00:49:19.303 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 13 00:49:19.311107 ignition[786]: INFO : Ignition 2.14.0 Sep 13 00:49:19.312966 ignition[786]: INFO : Stage: mount Sep 13 00:49:19.312966 ignition[786]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:19.312966 ignition[786]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:19.315557 ignition[786]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:19.317309 ignition[786]: INFO : mount: mount passed Sep 13 00:49:19.317309 ignition[786]: INFO : Ignition finished successfully Sep 13 00:49:19.319560 coreos-metadata[734]: Sep 13 00:49:19.318 INFO Fetch successful Sep 13 00:49:19.318718 systemd[1]: Finished ignition-mount.service. Sep 13 00:49:19.318000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.322959 coreos-metadata[734]: Sep 13 00:49:19.322 INFO wrote hostname ci-3510.3.8-n-1fe98121a2 to /sysroot/etc/hostname Sep 13 00:49:19.325169 systemd[1]: Finished flatcar-metadata-hostname.service. Sep 13 00:49:19.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.325950 systemd[1]: Finished sysroot-boot.service. Sep 13 00:49:19.325000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:19.428533 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 13 00:49:19.438963 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (794) Sep 13 00:49:19.453756 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Sep 13 00:49:19.453841 kernel: BTRFS info (device vda6): using free space tree Sep 13 00:49:19.453859 kernel: BTRFS info (device vda6): has skinny extents Sep 13 00:49:19.458842 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 13 00:49:19.460828 systemd[1]: Starting ignition-files.service... Sep 13 00:49:19.484375 ignition[814]: INFO : Ignition 2.14.0 Sep 13 00:49:19.484375 ignition[814]: INFO : Stage: files Sep 13 00:49:19.486104 ignition[814]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:19.486104 ignition[814]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:19.488279 ignition[814]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:19.489834 ignition[814]: DEBUG : files: compiled without relabeling support, skipping Sep 13 00:49:19.490958 ignition[814]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 00:49:19.490958 ignition[814]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 00:49:19.494424 ignition[814]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 00:49:19.495956 ignition[814]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 00:49:19.497288 ignition[814]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 00:49:19.497288 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 13 00:49:19.497288 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 13 00:49:19.496013 unknown[814]: wrote ssh authorized keys file for user: core Sep 13 00:49:19.551049 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 13 00:49:19.756338 systemd-networkd[688]: eth1: Gained IPv6LL Sep 13 00:49:19.954101 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 13 00:49:19.955729 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:49:19.956900 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 13 00:49:20.037357 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 13 00:49:20.076482 systemd-networkd[688]: eth0: Gained IPv6LL Sep 13 00:49:20.167643 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:49:20.168909 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 13 00:49:20.176557 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 13 00:49:20.388493 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Sep 13 00:49:21.041506 ignition[814]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 13 00:49:21.043036 ignition[814]: INFO : files: op(c): [started] processing unit "coreos-metadata-sshkeys@.service" Sep 13 00:49:21.043036 ignition[814]: INFO : files: op(c): [finished] processing unit "coreos-metadata-sshkeys@.service" Sep 13 00:49:21.043036 ignition[814]: INFO : files: op(d): [started] processing unit "prepare-helm.service" Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(d): op(e): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(d): op(e): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(d): [finished] processing unit "prepare-helm.service" Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(f): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(f): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Sep 13 00:49:21.046189 ignition[814]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 00:49:21.054526 ignition[814]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:49:21.054526 ignition[814]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:49:21.054526 ignition[814]: INFO : files: files passed Sep 13 00:49:21.054526 ignition[814]: INFO : Ignition finished successfully Sep 13 00:49:21.070894 kernel: kauditd_printk_skb: 28 callbacks suppressed Sep 13 00:49:21.070951 kernel: audit: type=1130 audit(1757724561.055:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.055000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.054997 systemd[1]: Finished ignition-files.service. Sep 13 00:49:21.079212 kernel: audit: type=1130 audit(1757724561.070:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.079246 kernel: audit: type=1131 audit(1757724561.070:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.070000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.070000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.057322 systemd[1]: Starting initrd-setup-root-after-ignition.service... Sep 13 00:49:21.084107 kernel: audit: type=1130 audit(1757724561.078:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.078000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.064043 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Sep 13 00:49:21.085552 initrd-setup-root-after-ignition[839]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 00:49:21.065353 systemd[1]: Starting ignition-quench.service... Sep 13 00:49:21.070300 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 00:49:21.070457 systemd[1]: Finished ignition-quench.service. Sep 13 00:49:21.072555 systemd[1]: Finished initrd-setup-root-after-ignition.service. Sep 13 00:49:21.080006 systemd[1]: Reached target ignition-complete.target. Sep 13 00:49:21.085709 systemd[1]: Starting initrd-parse-etc.service... Sep 13 00:49:21.103837 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 00:49:21.112338 kernel: audit: type=1130 audit(1757724561.103:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.112369 kernel: audit: type=1131 audit(1757724561.103:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.103000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.103000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.103975 systemd[1]: Finished initrd-parse-etc.service. Sep 13 00:49:21.104679 systemd[1]: Reached target initrd-fs.target. Sep 13 00:49:21.112802 systemd[1]: Reached target initrd.target. Sep 13 00:49:21.113797 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Sep 13 00:49:21.114750 systemd[1]: Starting dracut-pre-pivot.service... Sep 13 00:49:21.129172 systemd[1]: Finished dracut-pre-pivot.service. Sep 13 00:49:21.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.134978 kernel: audit: type=1130 audit(1757724561.129:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.135909 systemd[1]: Starting initrd-cleanup.service... Sep 13 00:49:21.146585 systemd[1]: Stopped target nss-lookup.target. Sep 13 00:49:21.147898 systemd[1]: Stopped target remote-cryptsetup.target. Sep 13 00:49:21.155972 systemd[1]: Stopped target timers.target. Sep 13 00:49:21.157332 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 00:49:21.158242 systemd[1]: Stopped dracut-pre-pivot.service. Sep 13 00:49:21.163408 kernel: audit: type=1131 audit(1757724561.158:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.158000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.159340 systemd[1]: Stopped target initrd.target. Sep 13 00:49:21.164863 systemd[1]: Stopped target basic.target. Sep 13 00:49:21.165722 systemd[1]: Stopped target ignition-complete.target. Sep 13 00:49:21.166974 systemd[1]: Stopped target ignition-diskful.target. Sep 13 00:49:21.168372 systemd[1]: Stopped target initrd-root-device.target. Sep 13 00:49:21.169507 systemd[1]: Stopped target remote-fs.target. Sep 13 00:49:21.170608 systemd[1]: Stopped target remote-fs-pre.target. Sep 13 00:49:21.172056 systemd[1]: Stopped target sysinit.target. Sep 13 00:49:21.173219 systemd[1]: Stopped target local-fs.target. Sep 13 00:49:21.174404 systemd[1]: Stopped target local-fs-pre.target. Sep 13 00:49:21.175586 systemd[1]: Stopped target swap.target. Sep 13 00:49:21.176715 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 00:49:21.182270 kernel: audit: type=1131 audit(1757724561.176:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.176000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.176981 systemd[1]: Stopped dracut-pre-mount.service. Sep 13 00:49:21.177886 systemd[1]: Stopped target cryptsetup.target. Sep 13 00:49:21.188437 kernel: audit: type=1131 audit(1757724561.183:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.183000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.182891 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 00:49:21.188000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.183174 systemd[1]: Stopped dracut-initqueue.service. Sep 13 00:49:21.189000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.184292 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 00:49:21.190000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.184523 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Sep 13 00:49:21.189307 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 00:49:21.189515 systemd[1]: Stopped ignition-files.service. Sep 13 00:49:21.190416 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 13 00:49:21.190564 systemd[1]: Stopped flatcar-metadata-hostname.service. Sep 13 00:49:21.192802 systemd[1]: Stopping ignition-mount.service... Sep 13 00:49:21.198581 systemd[1]: Stopping iscsiuio.service... Sep 13 00:49:21.203421 systemd[1]: Stopping sysroot-boot.service... Sep 13 00:49:21.204668 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 00:49:21.205797 systemd[1]: Stopped systemd-udev-trigger.service. Sep 13 00:49:21.206000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.207843 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 00:49:21.209049 systemd[1]: Stopped dracut-pre-trigger.service. Sep 13 00:49:21.209716 ignition[852]: INFO : Ignition 2.14.0 Sep 13 00:49:21.209716 ignition[852]: INFO : Stage: umount Sep 13 00:49:21.209716 ignition[852]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:49:21.209716 ignition[852]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Sep 13 00:49:21.213165 ignition[852]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Sep 13 00:49:21.221742 ignition[852]: INFO : umount: umount passed Sep 13 00:49:21.221742 ignition[852]: INFO : Ignition finished successfully Sep 13 00:49:21.222000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.225175 systemd[1]: iscsiuio.service: Deactivated successfully. Sep 13 00:49:21.226000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.225311 systemd[1]: Stopped iscsiuio.service. Sep 13 00:49:21.227742 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 00:49:21.228000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.228532 systemd[1]: Stopped ignition-mount.service. Sep 13 00:49:21.231000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.231404 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 00:49:21.232000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.231517 systemd[1]: Stopped ignition-disks.service. Sep 13 00:49:21.233000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.234000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.232054 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 00:49:21.232860 systemd[1]: Stopped ignition-kargs.service. Sep 13 00:49:21.233368 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 13 00:49:21.233414 systemd[1]: Stopped ignition-fetch.service. Sep 13 00:49:21.234316 systemd[1]: Stopped target network.target. Sep 13 00:49:21.234724 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 00:49:21.234866 systemd[1]: Stopped ignition-fetch-offline.service. Sep 13 00:49:21.235499 systemd[1]: Stopped target paths.target. Sep 13 00:49:21.244076 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 00:49:21.274000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.250051 systemd[1]: Stopped systemd-ask-password-console.path. Sep 13 00:49:21.251344 systemd[1]: Stopped target slices.target. Sep 13 00:49:21.265068 systemd[1]: Stopped target sockets.target. Sep 13 00:49:21.265622 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 00:49:21.265663 systemd[1]: Closed iscsid.socket. Sep 13 00:49:21.266164 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 00:49:21.304000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.266211 systemd[1]: Closed iscsiuio.socket. Sep 13 00:49:21.266698 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 00:49:21.266766 systemd[1]: Stopped ignition-setup.service. Sep 13 00:49:21.275540 systemd[1]: Stopping systemd-networkd.service... Sep 13 00:49:21.318000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.277301 systemd[1]: Stopping systemd-resolved.service... Sep 13 00:49:21.284046 systemd-networkd[688]: eth1: DHCPv6 lease lost Sep 13 00:49:21.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.320000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.292983 systemd-networkd[688]: eth0: DHCPv6 lease lost Sep 13 00:49:21.296479 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 00:49:21.323000 audit: BPF prog-id=9 op=UNLOAD Sep 13 00:49:21.323000 audit: BPF prog-id=6 op=UNLOAD Sep 13 00:49:21.297444 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 00:49:21.297622 systemd[1]: Stopped systemd-resolved.service. Sep 13 00:49:21.307289 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 00:49:21.307465 systemd[1]: Stopped systemd-networkd.service. Sep 13 00:49:21.347000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.320610 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 00:49:21.348000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.320748 systemd[1]: Finished initrd-cleanup.service. Sep 13 00:49:21.327491 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 00:49:21.327549 systemd[1]: Closed systemd-networkd.socket. Sep 13 00:49:21.346178 systemd[1]: Stopping network-cleanup.service... Sep 13 00:49:21.346907 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 00:49:21.347254 systemd[1]: Stopped parse-ip-for-networkd.service. Sep 13 00:49:21.348512 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:49:21.348609 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:49:21.351362 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 00:49:21.351421 systemd[1]: Stopped systemd-modules-load.service. Sep 13 00:49:21.369000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.370661 systemd[1]: Stopping systemd-udevd.service... Sep 13 00:49:21.378559 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 13 00:49:21.380801 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 00:49:21.381000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.380931 systemd[1]: Stopped sysroot-boot.service. Sep 13 00:49:21.385000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.386000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.383264 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 00:49:21.383331 systemd[1]: Stopped initrd-setup-root.service. Sep 13 00:49:21.387000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.386542 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 00:49:21.386693 systemd[1]: Stopped systemd-udevd.service. Sep 13 00:49:21.387684 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 00:49:21.391000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.387776 systemd[1]: Stopped network-cleanup.service. Sep 13 00:49:21.392000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.389104 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 00:49:21.393000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.389153 systemd[1]: Closed systemd-udevd-control.socket. Sep 13 00:49:21.390090 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 00:49:21.407000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.408000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.390126 systemd[1]: Closed systemd-udevd-kernel.socket. Sep 13 00:49:21.408000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.391355 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 00:49:21.391412 systemd[1]: Stopped dracut-pre-udev.service. Sep 13 00:49:21.411000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.411000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.392596 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 00:49:21.392646 systemd[1]: Stopped dracut-cmdline.service. Sep 13 00:49:21.393759 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 00:49:21.393802 systemd[1]: Stopped dracut-cmdline-ask.service. Sep 13 00:49:21.395968 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Sep 13 00:49:21.407271 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 13 00:49:21.407403 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Sep 13 00:49:21.408294 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 00:49:21.408358 systemd[1]: Stopped kmod-static-nodes.service. Sep 13 00:49:21.409051 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 00:49:21.409103 systemd[1]: Stopped systemd-vconsole-setup.service. Sep 13 00:49:21.410949 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 13 00:49:21.411512 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 00:49:21.411617 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Sep 13 00:49:21.412377 systemd[1]: Reached target initrd-switch-root.target. Sep 13 00:49:21.414318 systemd[1]: Starting initrd-switch-root.service... Sep 13 00:49:21.431907 systemd[1]: Switching root. Sep 13 00:49:21.453360 iscsid[698]: iscsid shutting down. Sep 13 00:49:21.454005 systemd-journald[185]: Received SIGTERM from PID 1 (n/a). Sep 13 00:49:21.454058 systemd-journald[185]: Journal stopped Sep 13 00:49:25.606774 kernel: SELinux: Class mctp_socket not defined in policy. Sep 13 00:49:25.606881 kernel: SELinux: Class anon_inode not defined in policy. Sep 13 00:49:25.606903 kernel: SELinux: the above unknown classes and permissions will be allowed Sep 13 00:49:25.610104 kernel: SELinux: policy capability network_peer_controls=1 Sep 13 00:49:25.610159 kernel: SELinux: policy capability open_perms=1 Sep 13 00:49:25.610185 kernel: SELinux: policy capability extended_socket_class=1 Sep 13 00:49:25.610210 kernel: SELinux: policy capability always_check_network=0 Sep 13 00:49:25.610229 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 13 00:49:25.610248 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 13 00:49:25.610272 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 13 00:49:25.610291 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 13 00:49:25.610313 systemd[1]: Successfully loaded SELinux policy in 59.841ms. Sep 13 00:49:25.610354 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.960ms. Sep 13 00:49:25.610382 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:49:25.610405 systemd[1]: Detected virtualization kvm. Sep 13 00:49:25.610426 systemd[1]: Detected architecture x86-64. Sep 13 00:49:25.610445 systemd[1]: Detected first boot. Sep 13 00:49:25.610469 systemd[1]: Hostname set to . Sep 13 00:49:25.610505 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:49:25.610526 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Sep 13 00:49:25.610547 systemd[1]: Populated /etc with preset unit settings. Sep 13 00:49:25.610568 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:49:25.610597 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:49:25.610620 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:49:25.610646 systemd[1]: iscsid.service: Deactivated successfully. Sep 13 00:49:25.610673 systemd[1]: Stopped iscsid.service. Sep 13 00:49:25.610693 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 13 00:49:25.610721 systemd[1]: Stopped initrd-switch-root.service. Sep 13 00:49:25.610741 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 13 00:49:25.610759 systemd[1]: Created slice system-addon\x2dconfig.slice. Sep 13 00:49:25.610809 systemd[1]: Created slice system-addon\x2drun.slice. Sep 13 00:49:25.610828 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Sep 13 00:49:25.610853 systemd[1]: Created slice system-getty.slice. Sep 13 00:49:25.610870 systemd[1]: Created slice system-modprobe.slice. Sep 13 00:49:25.610888 systemd[1]: Created slice system-serial\x2dgetty.slice. Sep 13 00:49:25.610906 systemd[1]: Created slice system-system\x2dcloudinit.slice. Sep 13 00:49:25.612365 systemd[1]: Created slice system-systemd\x2dfsck.slice. Sep 13 00:49:25.612400 systemd[1]: Created slice user.slice. Sep 13 00:49:25.612422 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:49:25.612443 systemd[1]: Started systemd-ask-password-wall.path. Sep 13 00:49:25.612472 systemd[1]: Set up automount boot.automount. Sep 13 00:49:25.612492 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Sep 13 00:49:25.612519 systemd[1]: Stopped target initrd-switch-root.target. Sep 13 00:49:25.612540 systemd[1]: Stopped target initrd-fs.target. Sep 13 00:49:25.612559 systemd[1]: Stopped target initrd-root-fs.target. Sep 13 00:49:25.612578 systemd[1]: Reached target integritysetup.target. Sep 13 00:49:25.612598 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:49:25.612619 systemd[1]: Reached target remote-fs.target. Sep 13 00:49:25.612644 systemd[1]: Reached target slices.target. Sep 13 00:49:25.612664 systemd[1]: Reached target swap.target. Sep 13 00:49:25.612685 systemd[1]: Reached target torcx.target. Sep 13 00:49:25.612737 systemd[1]: Reached target veritysetup.target. Sep 13 00:49:25.612756 systemd[1]: Listening on systemd-coredump.socket. Sep 13 00:49:25.612775 systemd[1]: Listening on systemd-initctl.socket. Sep 13 00:49:25.612793 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:49:25.612814 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:49:25.612836 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:49:25.612856 systemd[1]: Listening on systemd-userdbd.socket. Sep 13 00:49:25.612880 systemd[1]: Mounting dev-hugepages.mount... Sep 13 00:49:25.612900 systemd[1]: Mounting dev-mqueue.mount... Sep 13 00:49:25.612938 systemd[1]: Mounting media.mount... Sep 13 00:49:25.612959 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:25.612978 systemd[1]: Mounting sys-kernel-debug.mount... Sep 13 00:49:25.612998 systemd[1]: Mounting sys-kernel-tracing.mount... Sep 13 00:49:25.613017 systemd[1]: Mounting tmp.mount... Sep 13 00:49:25.613036 systemd[1]: Starting flatcar-tmpfiles.service... Sep 13 00:49:25.613059 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:25.613078 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:49:25.613095 systemd[1]: Starting modprobe@configfs.service... Sep 13 00:49:25.613113 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:25.613135 systemd[1]: Starting modprobe@drm.service... Sep 13 00:49:25.613155 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:25.613192 systemd[1]: Starting modprobe@fuse.service... Sep 13 00:49:25.613210 systemd[1]: Starting modprobe@loop.service... Sep 13 00:49:25.613228 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 00:49:25.613250 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 13 00:49:25.613268 systemd[1]: Stopped systemd-fsck-root.service. Sep 13 00:49:25.613289 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 13 00:49:25.613311 systemd[1]: Stopped systemd-fsck-usr.service. Sep 13 00:49:25.613332 systemd[1]: Stopped systemd-journald.service. Sep 13 00:49:25.613353 systemd[1]: Starting systemd-journald.service... Sep 13 00:49:25.613375 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:49:25.613397 systemd[1]: Starting systemd-network-generator.service... Sep 13 00:49:25.613417 systemd[1]: Starting systemd-remount-fs.service... Sep 13 00:49:25.613442 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:49:25.613462 systemd[1]: verity-setup.service: Deactivated successfully. Sep 13 00:49:25.613483 systemd[1]: Stopped verity-setup.service. Sep 13 00:49:25.613506 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:25.613525 systemd[1]: Mounted dev-hugepages.mount. Sep 13 00:49:25.613547 systemd[1]: Mounted dev-mqueue.mount. Sep 13 00:49:25.613567 systemd[1]: Mounted media.mount. Sep 13 00:49:25.613587 systemd[1]: Mounted sys-kernel-debug.mount. Sep 13 00:49:25.613607 systemd[1]: Mounted sys-kernel-tracing.mount. Sep 13 00:49:25.613642 kernel: loop: module loaded Sep 13 00:49:25.613667 systemd[1]: Mounted tmp.mount. Sep 13 00:49:25.613689 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:49:25.613709 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 13 00:49:25.613731 systemd[1]: Finished modprobe@configfs.service. Sep 13 00:49:25.613752 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:25.614756 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:25.614791 kernel: fuse: init (API version 7.34) Sep 13 00:49:25.614813 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:49:25.614834 systemd[1]: Finished modprobe@drm.service. Sep 13 00:49:25.614861 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:25.614882 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:25.614910 systemd-journald[956]: Journal started Sep 13 00:49:25.617317 systemd-journald[956]: Runtime Journal (/run/log/journal/9c81bef432b74dc0818e41e91c9d4566) is 4.9M, max 39.5M, 34.5M free. Sep 13 00:49:25.617396 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 13 00:49:21.605000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 13 00:49:21.669000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:49:21.669000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:49:21.669000 audit: BPF prog-id=10 op=LOAD Sep 13 00:49:21.669000 audit: BPF prog-id=10 op=UNLOAD Sep 13 00:49:21.669000 audit: BPF prog-id=11 op=LOAD Sep 13 00:49:21.669000 audit: BPF prog-id=11 op=UNLOAD Sep 13 00:49:21.782000 audit[885]: AVC avc: denied { associate } for pid=885 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Sep 13 00:49:21.782000 audit[885]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0001058d4 a1=c00002ae58 a2=c000029100 a3=32 items=0 ppid=868 pid=885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:49:21.782000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 13 00:49:21.784000 audit[885]: AVC avc: denied { associate } for pid=885 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Sep 13 00:49:21.784000 audit[885]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0001059b9 a2=1ed a3=0 items=2 ppid=868 pid=885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:49:21.784000 audit: CWD cwd="/" Sep 13 00:49:21.784000 audit: PATH item=0 name=(null) inode=2 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:21.784000 audit: PATH item=1 name=(null) inode=3 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:21.784000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 13 00:49:25.304000 audit: BPF prog-id=12 op=LOAD Sep 13 00:49:25.304000 audit: BPF prog-id=3 op=UNLOAD Sep 13 00:49:25.304000 audit: BPF prog-id=13 op=LOAD Sep 13 00:49:25.304000 audit: BPF prog-id=14 op=LOAD Sep 13 00:49:25.304000 audit: BPF prog-id=4 op=UNLOAD Sep 13 00:49:25.304000 audit: BPF prog-id=5 op=UNLOAD Sep 13 00:49:25.306000 audit: BPF prog-id=15 op=LOAD Sep 13 00:49:25.306000 audit: BPF prog-id=12 op=UNLOAD Sep 13 00:49:25.306000 audit: BPF prog-id=16 op=LOAD Sep 13 00:49:25.306000 audit: BPF prog-id=17 op=LOAD Sep 13 00:49:25.306000 audit: BPF prog-id=13 op=UNLOAD Sep 13 00:49:25.307000 audit: BPF prog-id=14 op=UNLOAD Sep 13 00:49:25.308000 audit: BPF prog-id=18 op=LOAD Sep 13 00:49:25.308000 audit: BPF prog-id=15 op=UNLOAD Sep 13 00:49:25.308000 audit: BPF prog-id=19 op=LOAD Sep 13 00:49:25.308000 audit: BPF prog-id=20 op=LOAD Sep 13 00:49:25.308000 audit: BPF prog-id=16 op=UNLOAD Sep 13 00:49:25.308000 audit: BPF prog-id=17 op=UNLOAD Sep 13 00:49:25.309000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.314000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.317000 audit: BPF prog-id=18 op=UNLOAD Sep 13 00:49:25.319000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.319000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.502000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.504000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.504000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.506000 audit: BPF prog-id=21 op=LOAD Sep 13 00:49:25.506000 audit: BPF prog-id=22 op=LOAD Sep 13 00:49:25.506000 audit: BPF prog-id=23 op=LOAD Sep 13 00:49:25.506000 audit: BPF prog-id=19 op=UNLOAD Sep 13 00:49:25.506000 audit: BPF prog-id=20 op=UNLOAD Sep 13 00:49:25.542000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.586000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.594000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.594000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.599000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.599000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.603000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Sep 13 00:49:25.603000 audit[956]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffe1f7894a0 a2=4000 a3=7ffe1f78953c items=0 ppid=1 pid=956 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:49:25.603000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Sep 13 00:49:25.606000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.606000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.613000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.613000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.302385 systemd[1]: Queued start job for default target multi-user.target. Sep 13 00:49:21.778519 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:49:25.302412 systemd[1]: Unnecessary job was removed for dev-vda6.device. Sep 13 00:49:21.779096 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 13 00:49:25.310364 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 13 00:49:21.779133 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 13 00:49:25.624176 systemd[1]: Finished modprobe@fuse.service. Sep 13 00:49:25.624230 systemd[1]: Started systemd-journald.service. Sep 13 00:49:25.621000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.621000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.779183 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Sep 13 00:49:21.779202 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="skipped missing lower profile" missing profile=oem Sep 13 00:49:21.779263 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Sep 13 00:49:21.779285 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Sep 13 00:49:25.624000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:21.779640 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Sep 13 00:49:21.779717 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 13 00:49:21.779742 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 13 00:49:25.626186 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:49:21.781808 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Sep 13 00:49:21.781873 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Sep 13 00:49:21.781905 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.8: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.8 Sep 13 00:49:21.781971 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Sep 13 00:49:21.782002 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.8: no such file or directory" path=/var/lib/torcx/store/3510.3.8 Sep 13 00:49:21.782017 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:21Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Sep 13 00:49:24.752734 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:49:24.753122 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:49:24.753356 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:49:24.753608 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 13 00:49:24.753666 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Sep 13 00:49:24.753746 /usr/lib/systemd/system-generators/torcx-generator[885]: time="2025-09-13T00:49:24Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Sep 13 00:49:25.628000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.628000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.628530 systemd[1]: Finished modprobe@loop.service. Sep 13 00:49:25.631000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.631415 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:49:25.634202 systemd[1]: Finished systemd-network-generator.service. Sep 13 00:49:25.634000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.635475 systemd[1]: Finished systemd-remount-fs.service. Sep 13 00:49:25.635000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.637265 systemd[1]: Reached target network-pre.target. Sep 13 00:49:25.642511 systemd[1]: Mounting sys-fs-fuse-connections.mount... Sep 13 00:49:25.647206 systemd[1]: Mounting sys-kernel-config.mount... Sep 13 00:49:25.648028 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 00:49:25.653191 systemd[1]: Starting systemd-hwdb-update.service... Sep 13 00:49:25.656304 systemd[1]: Starting systemd-journal-flush.service... Sep 13 00:49:25.657230 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:49:25.660153 systemd[1]: Starting systemd-random-seed.service... Sep 13 00:49:25.661275 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:49:25.664456 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:49:25.671547 systemd[1]: Mounted sys-fs-fuse-connections.mount. Sep 13 00:49:25.672513 systemd[1]: Mounted sys-kernel-config.mount. Sep 13 00:49:25.679400 systemd-journald[956]: Time spent on flushing to /var/log/journal/9c81bef432b74dc0818e41e91c9d4566 is 53.236ms for 1160 entries. Sep 13 00:49:25.679400 systemd-journald[956]: System Journal (/var/log/journal/9c81bef432b74dc0818e41e91c9d4566) is 8.0M, max 195.6M, 187.6M free. Sep 13 00:49:25.739354 systemd-journald[956]: Received client request to flush runtime journal. Sep 13 00:49:25.683000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.709000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.710000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.683193 systemd[1]: Finished systemd-random-seed.service. Sep 13 00:49:25.740000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.684315 systemd[1]: Reached target first-boot-complete.target. Sep 13 00:49:25.709201 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:49:25.710514 systemd[1]: Finished flatcar-tmpfiles.service. Sep 13 00:49:25.713293 systemd[1]: Starting systemd-sysusers.service... Sep 13 00:49:25.741086 systemd[1]: Finished systemd-journal-flush.service. Sep 13 00:49:25.758410 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:49:25.758000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.761049 systemd[1]: Starting systemd-udev-settle.service... Sep 13 00:49:25.767525 systemd[1]: Finished systemd-sysusers.service. Sep 13 00:49:25.767000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.770239 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 13 00:49:25.784759 udevadm[997]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 13 00:49:25.808000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:25.808216 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 13 00:49:26.375219 systemd[1]: Finished systemd-hwdb-update.service. Sep 13 00:49:26.382698 kernel: kauditd_printk_skb: 109 callbacks suppressed Sep 13 00:49:26.382856 kernel: audit: type=1130 audit(1757724566.375:149): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.382902 kernel: audit: type=1334 audit(1757724566.377:150): prog-id=24 op=LOAD Sep 13 00:49:26.375000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.377000 audit: BPF prog-id=24 op=LOAD Sep 13 00:49:26.382404 systemd[1]: Starting systemd-udevd.service... Sep 13 00:49:26.381000 audit: BPF prog-id=25 op=LOAD Sep 13 00:49:26.381000 audit: BPF prog-id=7 op=UNLOAD Sep 13 00:49:26.381000 audit: BPF prog-id=8 op=UNLOAD Sep 13 00:49:26.383983 kernel: audit: type=1334 audit(1757724566.381:151): prog-id=25 op=LOAD Sep 13 00:49:26.384074 kernel: audit: type=1334 audit(1757724566.381:152): prog-id=7 op=UNLOAD Sep 13 00:49:26.384101 kernel: audit: type=1334 audit(1757724566.381:153): prog-id=8 op=UNLOAD Sep 13 00:49:26.419115 systemd-udevd[1000]: Using default interface naming scheme 'v252'. Sep 13 00:49:26.455154 systemd[1]: Started systemd-udevd.service. Sep 13 00:49:26.467647 kernel: audit: type=1130 audit(1757724566.455:154): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.467752 kernel: audit: type=1334 audit(1757724566.462:155): prog-id=26 op=LOAD Sep 13 00:49:26.455000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.462000 audit: BPF prog-id=26 op=LOAD Sep 13 00:49:26.465695 systemd[1]: Starting systemd-networkd.service... Sep 13 00:49:26.479384 kernel: audit: type=1334 audit(1757724566.472:156): prog-id=27 op=LOAD Sep 13 00:49:26.479481 kernel: audit: type=1334 audit(1757724566.472:157): prog-id=28 op=LOAD Sep 13 00:49:26.479502 kernel: audit: type=1334 audit(1757724566.472:158): prog-id=29 op=LOAD Sep 13 00:49:26.472000 audit: BPF prog-id=27 op=LOAD Sep 13 00:49:26.472000 audit: BPF prog-id=28 op=LOAD Sep 13 00:49:26.472000 audit: BPF prog-id=29 op=LOAD Sep 13 00:49:26.477420 systemd[1]: Starting systemd-userdbd.service... Sep 13 00:49:26.546000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.546676 systemd[1]: Started systemd-userdbd.service. Sep 13 00:49:26.580664 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Sep 13 00:49:26.593512 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:26.593681 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:26.594999 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:26.598356 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:26.601303 systemd[1]: Starting modprobe@loop.service... Sep 13 00:49:26.606000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.606000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.607000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.607000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.605078 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 00:49:26.605158 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 00:49:26.605272 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:26.606020 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:26.606241 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:26.607654 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:49:26.607865 systemd[1]: Finished modprobe@loop.service. Sep 13 00:49:26.610105 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:49:26.612000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.612000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.612334 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:26.612476 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:26.613143 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:49:26.667678 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 13 00:49:26.678768 systemd-networkd[1014]: lo: Link UP Sep 13 00:49:26.679199 systemd-networkd[1014]: lo: Gained carrier Sep 13 00:49:26.680174 systemd-networkd[1014]: Enumeration completed Sep 13 00:49:26.680000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.680510 systemd[1]: Started systemd-networkd.service. Sep 13 00:49:26.681898 systemd-networkd[1014]: eth1: Configuring with /run/systemd/network/10-8a:77:91:07:ab:9b.network. Sep 13 00:49:26.684783 systemd-networkd[1014]: eth0: Configuring with /run/systemd/network/10-9a:42:e3:cd:54:71.network. Sep 13 00:49:26.686219 systemd-networkd[1014]: eth1: Link UP Sep 13 00:49:26.686369 systemd-networkd[1014]: eth1: Gained carrier Sep 13 00:49:26.691326 systemd-networkd[1014]: eth0: Link UP Sep 13 00:49:26.691336 systemd-networkd[1014]: eth0: Gained carrier Sep 13 00:49:26.695954 kernel: ACPI: button: Power Button [PWRF] Sep 13 00:49:26.722707 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:49:26.725000 audit[1010]: AVC avc: denied { confidentiality } for pid=1010 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Sep 13 00:49:26.725000 audit[1010]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=55dffbfe5660 a1=338ec a2=7f816500fbc5 a3=5 items=110 ppid=1000 pid=1010 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:49:26.725000 audit: CWD cwd="/" Sep 13 00:49:26.725000 audit: PATH item=0 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=1 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=2 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=3 name=(null) inode=14725 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=4 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=5 name=(null) inode=14726 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=6 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=7 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=8 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=9 name=(null) inode=14728 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=10 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=11 name=(null) inode=14729 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=12 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=13 name=(null) inode=14730 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=14 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=15 name=(null) inode=14731 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=16 name=(null) inode=14727 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=17 name=(null) inode=14732 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=18 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=19 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=20 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=21 name=(null) inode=14734 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=22 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=23 name=(null) inode=14735 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=24 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=25 name=(null) inode=14736 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=26 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=27 name=(null) inode=14737 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=28 name=(null) inode=14733 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=29 name=(null) inode=14738 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=30 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=31 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=32 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=33 name=(null) inode=14740 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=34 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=35 name=(null) inode=14741 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=36 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=37 name=(null) inode=14742 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=38 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=39 name=(null) inode=14743 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=40 name=(null) inode=14739 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=41 name=(null) inode=14744 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=42 name=(null) inode=14724 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=43 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=44 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=45 name=(null) inode=14746 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=46 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=47 name=(null) inode=14747 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=48 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=49 name=(null) inode=14748 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=50 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=51 name=(null) inode=14749 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=52 name=(null) inode=14745 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=53 name=(null) inode=14750 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=54 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=55 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=56 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=57 name=(null) inode=14752 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=58 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=59 name=(null) inode=14753 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=60 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=61 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=62 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=63 name=(null) inode=14755 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=64 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=65 name=(null) inode=14756 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=66 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=67 name=(null) inode=14757 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=68 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=69 name=(null) inode=14758 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=70 name=(null) inode=14754 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=71 name=(null) inode=14759 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=72 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=73 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=74 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=75 name=(null) inode=14761 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=76 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=77 name=(null) inode=14762 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=78 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=79 name=(null) inode=14763 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=80 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=81 name=(null) inode=14764 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=82 name=(null) inode=14760 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=83 name=(null) inode=14765 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=84 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=85 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=86 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=87 name=(null) inode=14767 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=88 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=89 name=(null) inode=14768 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=90 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=91 name=(null) inode=14769 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=92 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=93 name=(null) inode=14770 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=94 name=(null) inode=14766 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=95 name=(null) inode=14771 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=96 name=(null) inode=14751 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=97 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=98 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=99 name=(null) inode=14773 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=100 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=101 name=(null) inode=14774 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=102 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=103 name=(null) inode=14775 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=104 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=105 name=(null) inode=14776 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=106 name=(null) inode=14772 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=107 name=(null) inode=14777 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PATH item=109 name=(null) inode=14778 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:49:26.725000 audit: PROCTITLE proctitle="(udev-worker)" Sep 13 00:49:26.781969 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Sep 13 00:49:26.798948 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Sep 13 00:49:26.825388 kernel: mousedev: PS/2 mouse device common for all mice Sep 13 00:49:26.957949 kernel: EDAC MC: Ver: 3.0.0 Sep 13 00:49:26.977736 systemd[1]: Finished systemd-udev-settle.service. Sep 13 00:49:26.977000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:26.980766 systemd[1]: Starting lvm2-activation-early.service... Sep 13 00:49:27.003701 lvm[1038]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:49:27.040357 systemd[1]: Finished lvm2-activation-early.service. Sep 13 00:49:27.040000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.041090 systemd[1]: Reached target cryptsetup.target. Sep 13 00:49:27.043434 systemd[1]: Starting lvm2-activation.service... Sep 13 00:49:27.051622 lvm[1039]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:49:27.083635 systemd[1]: Finished lvm2-activation.service. Sep 13 00:49:27.083000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.084372 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:49:27.086381 systemd[1]: Mounting media-configdrive.mount... Sep 13 00:49:27.086930 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 13 00:49:27.086981 systemd[1]: Reached target machines.target. Sep 13 00:49:27.088932 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Sep 13 00:49:27.103020 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Sep 13 00:49:27.102000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.113955 kernel: ISO 9660 Extensions: RRIP_1991A Sep 13 00:49:27.115883 systemd[1]: Mounted media-configdrive.mount. Sep 13 00:49:27.116594 systemd[1]: Reached target local-fs.target. Sep 13 00:49:27.118453 systemd[1]: Starting ldconfig.service... Sep 13 00:49:27.119837 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:49:27.119934 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:27.122421 systemd[1]: Starting systemd-boot-update.service... Sep 13 00:49:27.125160 systemd[1]: Starting systemd-machine-id-commit.service... Sep 13 00:49:27.130086 systemd[1]: Starting systemd-sysext.service... Sep 13 00:49:27.132723 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1045 (bootctl) Sep 13 00:49:27.136796 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Sep 13 00:49:27.154850 systemd[1]: Unmounting usr-share-oem.mount... Sep 13 00:49:27.165580 systemd[1]: usr-share-oem.mount: Deactivated successfully. Sep 13 00:49:27.165866 systemd[1]: Unmounted usr-share-oem.mount. Sep 13 00:49:27.195096 kernel: loop0: detected capacity change from 0 to 229808 Sep 13 00:49:27.278893 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 13 00:49:27.280000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.281228 systemd[1]: Finished systemd-machine-id-commit.service. Sep 13 00:49:27.303400 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 13 00:49:27.305724 systemd-fsck[1051]: fsck.fat 4.2 (2021-01-31) Sep 13 00:49:27.305724 systemd-fsck[1051]: /dev/vda1: 790 files, 120761/258078 clusters Sep 13 00:49:27.315000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.315387 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Sep 13 00:49:27.317988 systemd[1]: Mounting boot.mount... Sep 13 00:49:27.326958 kernel: loop1: detected capacity change from 0 to 229808 Sep 13 00:49:27.331958 systemd[1]: Mounted boot.mount. Sep 13 00:49:27.352000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.352775 systemd[1]: Finished systemd-boot-update.service. Sep 13 00:49:27.360427 (sd-sysext)[1056]: Using extensions 'kubernetes'. Sep 13 00:49:27.364378 (sd-sysext)[1056]: Merged extensions into '/usr'. Sep 13 00:49:27.412992 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:27.416104 systemd[1]: Mounting usr-share-oem.mount... Sep 13 00:49:27.419407 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:27.422835 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:27.427113 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:27.432032 systemd[1]: Starting modprobe@loop.service... Sep 13 00:49:27.434510 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:49:27.435025 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:27.435463 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:27.442851 systemd[1]: Mounted usr-share-oem.mount. Sep 13 00:49:27.445033 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:27.445196 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:27.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.445000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.446615 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:27.446801 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:27.447000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.447000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.448910 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:49:27.449235 systemd[1]: Finished modprobe@loop.service. Sep 13 00:49:27.450000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.450000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.451972 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:49:27.452084 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:49:27.456711 systemd[1]: Finished systemd-sysext.service. Sep 13 00:49:27.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.459478 systemd[1]: Starting ensure-sysext.service... Sep 13 00:49:27.461877 systemd[1]: Starting systemd-tmpfiles-setup.service... Sep 13 00:49:27.484117 systemd[1]: Reloading. Sep 13 00:49:27.485885 systemd-tmpfiles[1063]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Sep 13 00:49:27.490127 systemd-tmpfiles[1063]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 13 00:49:27.504516 systemd-tmpfiles[1063]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 13 00:49:27.667500 ldconfig[1044]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 13 00:49:27.688809 /usr/lib/systemd/system-generators/torcx-generator[1083]: time="2025-09-13T00:49:27Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:49:27.688842 /usr/lib/systemd/system-generators/torcx-generator[1083]: time="2025-09-13T00:49:27Z" level=info msg="torcx already run" Sep 13 00:49:27.830270 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:49:27.830296 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:49:27.849945 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:49:27.939000 audit: BPF prog-id=30 op=LOAD Sep 13 00:49:27.939000 audit: BPF prog-id=31 op=LOAD Sep 13 00:49:27.939000 audit: BPF prog-id=24 op=UNLOAD Sep 13 00:49:27.939000 audit: BPF prog-id=25 op=UNLOAD Sep 13 00:49:27.941000 audit: BPF prog-id=32 op=LOAD Sep 13 00:49:27.941000 audit: BPF prog-id=26 op=UNLOAD Sep 13 00:49:27.944000 audit: BPF prog-id=33 op=LOAD Sep 13 00:49:27.944000 audit: BPF prog-id=21 op=UNLOAD Sep 13 00:49:27.944000 audit: BPF prog-id=34 op=LOAD Sep 13 00:49:27.944000 audit: BPF prog-id=35 op=LOAD Sep 13 00:49:27.944000 audit: BPF prog-id=22 op=UNLOAD Sep 13 00:49:27.944000 audit: BPF prog-id=23 op=UNLOAD Sep 13 00:49:27.945000 audit: BPF prog-id=36 op=LOAD Sep 13 00:49:27.945000 audit: BPF prog-id=27 op=UNLOAD Sep 13 00:49:27.946000 audit: BPF prog-id=37 op=LOAD Sep 13 00:49:27.946000 audit: BPF prog-id=38 op=LOAD Sep 13 00:49:27.946000 audit: BPF prog-id=28 op=UNLOAD Sep 13 00:49:27.946000 audit: BPF prog-id=29 op=UNLOAD Sep 13 00:49:27.956221 systemd[1]: Finished ldconfig.service. Sep 13 00:49:27.956000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.958716 systemd[1]: Finished systemd-tmpfiles-setup.service. Sep 13 00:49:27.958000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.963875 systemd[1]: Starting audit-rules.service... Sep 13 00:49:27.966650 systemd[1]: Starting clean-ca-certificates.service... Sep 13 00:49:27.976000 audit: BPF prog-id=39 op=LOAD Sep 13 00:49:27.969657 systemd[1]: Starting systemd-journal-catalog-update.service... Sep 13 00:49:27.980000 audit: BPF prog-id=40 op=LOAD Sep 13 00:49:27.989000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:27.979368 systemd[1]: Starting systemd-resolved.service... Sep 13 00:49:27.982698 systemd[1]: Starting systemd-timesyncd.service... Sep 13 00:49:27.986166 systemd[1]: Starting systemd-update-utmp.service... Sep 13 00:49:27.988516 systemd[1]: Finished clean-ca-certificates.service. Sep 13 00:49:27.994046 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:49:27.999812 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.001467 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:28.008000 audit[1138]: SYSTEM_BOOT pid=1138 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.007880 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:28.012000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.012000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.013000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.013000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.015000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.015000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.010562 systemd[1]: Starting modprobe@loop.service... Sep 13 00:49:28.011194 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.011371 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:28.011509 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:49:28.012675 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:28.012856 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:28.013902 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:28.014068 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:28.015171 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:49:28.015298 systemd[1]: Finished modprobe@loop.service. Sep 13 00:49:28.020259 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:49:28.020446 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.036000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.023886 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.026507 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:28.030126 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:28.033757 systemd[1]: Starting modprobe@loop.service... Sep 13 00:49:28.034365 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.034524 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:28.034654 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:49:28.035788 systemd[1]: Finished systemd-update-utmp.service. Sep 13 00:49:28.037819 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:28.038228 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:28.039000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.039000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.040000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.040000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.040583 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:28.040749 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:28.048407 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.051981 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:49:28.056700 systemd[1]: Starting modprobe@drm.service... Sep 13 00:49:28.072310 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:49:28.073252 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.073584 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:28.076352 systemd[1]: Starting systemd-networkd-wait-online.service... Sep 13 00:49:28.077137 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:49:28.079447 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:49:28.079704 systemd[1]: Finished modprobe@loop.service. Sep 13 00:49:28.081000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.081000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.083135 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:49:28.083376 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:49:28.083000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.083000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.085094 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.086695 systemd[1]: Finished ensure-sysext.service. Sep 13 00:49:28.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.094084 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:49:28.094244 systemd[1]: Finished modprobe@drm.service. Sep 13 00:49:28.093000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.093000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.096456 systemd[1]: Finished systemd-journal-catalog-update.service. Sep 13 00:49:28.096000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.099177 systemd[1]: Starting systemd-update-done.service... Sep 13 00:49:28.103655 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:49:28.103828 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:49:28.104604 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:49:28.103000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.103000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.112340 systemd[1]: Finished systemd-update-done.service. Sep 13 00:49:28.111000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:49:28.116083 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:28.116113 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:49:28.150000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Sep 13 00:49:28.150000 audit[1161]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc04f17250 a2=420 a3=0 items=0 ppid=1130 pid=1161 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:49:28.150000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Sep 13 00:49:28.152153 augenrules[1161]: No rules Sep 13 00:49:28.153777 systemd[1]: Finished audit-rules.service. Sep 13 00:49:28.168411 systemd-resolved[1134]: Positive Trust Anchors: Sep 13 00:49:28.168432 systemd-resolved[1134]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:49:28.168476 systemd-resolved[1134]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:49:28.177008 systemd[1]: Started systemd-timesyncd.service. Sep 13 00:49:28.177948 systemd[1]: Reached target time-set.target. Sep 13 00:49:28.178674 systemd-resolved[1134]: Using system hostname 'ci-3510.3.8-n-1fe98121a2'. Sep 13 00:49:28.181883 systemd[1]: Started systemd-resolved.service. Sep 13 00:49:28.182696 systemd[1]: Reached target network.target. Sep 13 00:49:28.183340 systemd[1]: Reached target nss-lookup.target. Sep 13 00:49:28.183985 systemd[1]: Reached target sysinit.target. Sep 13 00:49:28.184750 systemd[1]: Started motdgen.path. Sep 13 00:49:28.185411 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Sep 13 00:49:28.186413 systemd[1]: Started logrotate.timer. Sep 13 00:49:28.187170 systemd[1]: Started mdadm.timer. Sep 13 00:49:28.187749 systemd[1]: Started systemd-tmpfiles-clean.timer. Sep 13 00:49:28.188528 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 13 00:49:28.188582 systemd[1]: Reached target paths.target. Sep 13 00:49:28.189220 systemd[1]: Reached target timers.target. Sep 13 00:49:28.190302 systemd[1]: Listening on dbus.socket. Sep 13 00:49:28.192712 systemd[1]: Starting docker.socket... Sep 13 00:49:28.197733 systemd[1]: Listening on sshd.socket. Sep 13 00:49:28.198659 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:28.199529 systemd[1]: Listening on docker.socket. Sep 13 00:49:28.200269 systemd[1]: Reached target sockets.target. Sep 13 00:49:28.200720 systemd[1]: Reached target basic.target. Sep 13 00:49:28.201201 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.201231 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:49:28.202572 systemd[1]: Starting containerd.service... Sep 13 00:49:28.204335 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Sep 13 00:49:28.206976 systemd[1]: Starting dbus.service... Sep 13 00:49:28.213011 systemd[1]: Starting enable-oem-cloudinit.service... Sep 13 00:49:28.215811 systemd[1]: Starting extend-filesystems.service... Sep 13 00:49:28.888985 jq[1174]: false Sep 13 00:49:28.216649 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Sep 13 00:49:28.218056 systemd[1]: Starting motdgen.service... Sep 13 00:49:28.221901 systemd[1]: Starting prepare-helm.service... Sep 13 00:49:28.226688 systemd[1]: Starting ssh-key-proc-cmdline.service... Sep 13 00:49:28.229703 systemd[1]: Starting sshd-keygen.service... Sep 13 00:49:28.235574 systemd[1]: Starting systemd-logind.service... Sep 13 00:49:28.237053 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:49:28.237157 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 13 00:49:28.237743 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 13 00:49:28.240172 systemd[1]: Starting update-engine.service... Sep 13 00:49:28.244988 systemd[1]: Starting update-ssh-keys-after-ignition.service... Sep 13 00:49:28.890426 systemd-timesyncd[1137]: Contacted time server 216.229.0.50:123 (0.flatcar.pool.ntp.org). Sep 13 00:49:28.890527 systemd-timesyncd[1137]: Initial clock synchronization to Sat 2025-09-13 00:49:28.888743 UTC. Sep 13 00:49:28.891974 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 13 00:49:28.892233 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Sep 13 00:49:28.894989 jq[1184]: true Sep 13 00:49:28.895275 systemd-resolved[1134]: Clock change detected. Flushing caches. Sep 13 00:49:28.917168 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 13 00:49:28.917437 systemd[1]: Finished ssh-key-proc-cmdline.service. Sep 13 00:49:28.921277 jq[1187]: true Sep 13 00:49:28.931660 tar[1186]: linux-amd64/LICENSE Sep 13 00:49:28.935768 tar[1186]: linux-amd64/helm Sep 13 00:49:28.942127 extend-filesystems[1175]: Found loop1 Sep 13 00:49:28.943825 extend-filesystems[1175]: Found vda Sep 13 00:49:28.953422 extend-filesystems[1175]: Found vda1 Sep 13 00:49:28.954556 extend-filesystems[1175]: Found vda2 Sep 13 00:49:28.956903 dbus-daemon[1171]: [system] SELinux support is enabled Sep 13 00:49:28.957424 systemd[1]: Started dbus.service. Sep 13 00:49:28.960334 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 13 00:49:28.960375 systemd[1]: Reached target system-config.target. Sep 13 00:49:28.960939 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 13 00:49:28.960962 systemd[1]: Reached target user-config.target. Sep 13 00:49:28.961213 extend-filesystems[1175]: Found vda3 Sep 13 00:49:28.962636 extend-filesystems[1175]: Found usr Sep 13 00:49:28.962636 extend-filesystems[1175]: Found vda4 Sep 13 00:49:28.962636 extend-filesystems[1175]: Found vda6 Sep 13 00:49:28.962636 extend-filesystems[1175]: Found vda7 Sep 13 00:49:28.962636 extend-filesystems[1175]: Found vda9 Sep 13 00:49:28.962636 extend-filesystems[1175]: Checking size of /dev/vda9 Sep 13 00:49:29.008147 systemd[1]: motdgen.service: Deactivated successfully. Sep 13 00:49:29.008407 systemd[1]: Finished motdgen.service. Sep 13 00:49:29.017005 update_engine[1183]: I0913 00:49:29.016335 1183 main.cc:92] Flatcar Update Engine starting Sep 13 00:49:29.025974 systemd[1]: Started update-engine.service. Sep 13 00:49:29.026388 update_engine[1183]: I0913 00:49:29.026084 1183 update_check_scheduler.cc:74] Next update check in 7m23s Sep 13 00:49:29.028449 systemd[1]: Started locksmithd.service. Sep 13 00:49:29.035071 extend-filesystems[1175]: Resized partition /dev/vda9 Sep 13 00:49:29.044586 extend-filesystems[1225]: resize2fs 1.46.5 (30-Dec-2021) Sep 13 00:49:29.053637 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Sep 13 00:49:29.094109 bash[1224]: Updated "/home/core/.ssh/authorized_keys" Sep 13 00:49:29.095003 systemd[1]: Finished update-ssh-keys-after-ignition.service. Sep 13 00:49:29.103940 systemd-networkd[1014]: eth0: Gained IPv6LL Sep 13 00:49:29.106145 systemd[1]: Finished systemd-networkd-wait-online.service. Sep 13 00:49:29.106995 systemd[1]: Reached target network-online.target. Sep 13 00:49:29.109371 systemd[1]: Starting kubelet.service... Sep 13 00:49:29.152948 systemd-logind[1182]: Watching system buttons on /dev/input/event1 (Power Button) Sep 13 00:49:29.157761 systemd-logind[1182]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 13 00:49:29.163181 systemd-logind[1182]: New seat seat0. Sep 13 00:49:29.171854 systemd[1]: Started systemd-logind.service. Sep 13 00:49:29.177899 env[1189]: time="2025-09-13T00:49:29.177826994Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Sep 13 00:49:29.184934 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Sep 13 00:49:29.196997 coreos-metadata[1170]: Sep 13 00:49:29.196 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 13 00:49:29.214481 extend-filesystems[1225]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 13 00:49:29.214481 extend-filesystems[1225]: old_desc_blocks = 1, new_desc_blocks = 8 Sep 13 00:49:29.214481 extend-filesystems[1225]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Sep 13 00:49:29.218732 extend-filesystems[1175]: Resized filesystem in /dev/vda9 Sep 13 00:49:29.218732 extend-filesystems[1175]: Found vdb Sep 13 00:49:29.216866 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 13 00:49:29.217126 systemd[1]: Finished extend-filesystems.service. Sep 13 00:49:29.222847 coreos-metadata[1170]: Sep 13 00:49:29.221 INFO Fetch successful Sep 13 00:49:29.227924 unknown[1170]: wrote ssh authorized keys file for user: core Sep 13 00:49:29.244570 update-ssh-keys[1231]: Updated "/home/core/.ssh/authorized_keys" Sep 13 00:49:29.244903 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Sep 13 00:49:29.255330 env[1189]: time="2025-09-13T00:49:29.255257067Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 13 00:49:29.255497 env[1189]: time="2025-09-13T00:49:29.255443874Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.259676 env[1189]: time="2025-09-13T00:49:29.259285549Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.192-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:49:29.259676 env[1189]: time="2025-09-13T00:49:29.259659801Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.259981 env[1189]: time="2025-09-13T00:49:29.259951978Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:49:29.259981 env[1189]: time="2025-09-13T00:49:29.259976044Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.260075 env[1189]: time="2025-09-13T00:49:29.259990229Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Sep 13 00:49:29.260075 env[1189]: time="2025-09-13T00:49:29.260000491Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.260123 env[1189]: time="2025-09-13T00:49:29.260088115Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.260362 env[1189]: time="2025-09-13T00:49:29.260338955Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:49:29.260502 env[1189]: time="2025-09-13T00:49:29.260479174Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:49:29.260502 env[1189]: time="2025-09-13T00:49:29.260498366Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 13 00:49:29.260581 env[1189]: time="2025-09-13T00:49:29.260556421Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Sep 13 00:49:29.260581 env[1189]: time="2025-09-13T00:49:29.260568615Z" level=info msg="metadata content store policy set" policy=shared Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270215041Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270283094Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270298056Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270432749Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270466184Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270487617Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270507980Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270529385Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270549826Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270574556Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270597034Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270627758Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270820493Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 13 00:49:29.272633 env[1189]: time="2025-09-13T00:49:29.270910191Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271148930Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271173215Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271188818Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271242913Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271256545Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271270053Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271282018Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271293897Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271306327Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271317796Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271329012Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271342551Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271502495Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271517969Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273092 env[1189]: time="2025-09-13T00:49:29.271530386Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273492 env[1189]: time="2025-09-13T00:49:29.271542579Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 13 00:49:29.273492 env[1189]: time="2025-09-13T00:49:29.271558501Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Sep 13 00:49:29.273492 env[1189]: time="2025-09-13T00:49:29.271571417Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 13 00:49:29.273492 env[1189]: time="2025-09-13T00:49:29.271595007Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Sep 13 00:49:29.273492 env[1189]: time="2025-09-13T00:49:29.271672134Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 13 00:49:29.273662 env[1189]: time="2025-09-13T00:49:29.271911319Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 13 00:49:29.273662 env[1189]: time="2025-09-13T00:49:29.271967689Z" level=info msg="Connect containerd service" Sep 13 00:49:29.273662 env[1189]: time="2025-09-13T00:49:29.272016555Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 13 00:49:29.273662 env[1189]: time="2025-09-13T00:49:29.272562882Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:49:29.277417 env[1189]: time="2025-09-13T00:49:29.274163853Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 13 00:49:29.277417 env[1189]: time="2025-09-13T00:49:29.274233278Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 13 00:49:29.277417 env[1189]: time="2025-09-13T00:49:29.274305980Z" level=info msg="containerd successfully booted in 0.102058s" Sep 13 00:49:29.274483 systemd[1]: Started containerd.service. Sep 13 00:49:29.284053 env[1189]: time="2025-09-13T00:49:29.283976723Z" level=info msg="Start subscribing containerd event" Sep 13 00:49:29.295920 systemd-networkd[1014]: eth1: Gained IPv6LL Sep 13 00:49:29.302587 env[1189]: time="2025-09-13T00:49:29.302508451Z" level=info msg="Start recovering state" Sep 13 00:49:29.303041 env[1189]: time="2025-09-13T00:49:29.303005716Z" level=info msg="Start event monitor" Sep 13 00:49:29.303267 env[1189]: time="2025-09-13T00:49:29.303245302Z" level=info msg="Start snapshots syncer" Sep 13 00:49:29.303371 env[1189]: time="2025-09-13T00:49:29.303348973Z" level=info msg="Start cni network conf syncer for default" Sep 13 00:49:29.303475 env[1189]: time="2025-09-13T00:49:29.303459131Z" level=info msg="Start streaming server" Sep 13 00:49:29.543242 systemd[1]: Created slice system-sshd.slice. Sep 13 00:49:30.456820 locksmithd[1222]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 13 00:49:30.587257 tar[1186]: linux-amd64/README.md Sep 13 00:49:30.597023 systemd[1]: Finished prepare-helm.service. Sep 13 00:49:30.784261 sshd_keygen[1201]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 13 00:49:30.821385 systemd[1]: Finished sshd-keygen.service. Sep 13 00:49:30.825340 systemd[1]: Starting issuegen.service... Sep 13 00:49:30.829689 systemd[1]: Started sshd@0-24.144.95.58:22-147.75.109.163:49814.service. Sep 13 00:49:30.868113 systemd[1]: issuegen.service: Deactivated successfully. Sep 13 00:49:30.868323 systemd[1]: Finished issuegen.service. Sep 13 00:49:30.873030 systemd[1]: Starting systemd-user-sessions.service... Sep 13 00:49:30.938523 systemd[1]: Finished systemd-user-sessions.service. Sep 13 00:49:30.942518 systemd[1]: Started getty@tty1.service. Sep 13 00:49:30.947594 systemd[1]: Started serial-getty@ttyS0.service. Sep 13 00:49:30.948951 systemd[1]: Reached target getty.target. Sep 13 00:49:31.019451 sshd[1249]: Accepted publickey for core from 147.75.109.163 port 49814 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:49:31.021304 sshd[1249]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.042670 systemd[1]: Started kubelet.service. Sep 13 00:49:31.048734 systemd-logind[1182]: New session 1 of user core. Sep 13 00:49:31.049090 systemd[1]: Created slice user-500.slice. Sep 13 00:49:31.049760 systemd[1]: Reached target multi-user.target. Sep 13 00:49:31.052637 systemd[1]: Starting systemd-update-utmp-runlevel.service... Sep 13 00:49:31.062758 systemd[1]: Starting user-runtime-dir@500.service... Sep 13 00:49:31.076883 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Sep 13 00:49:31.077113 systemd[1]: Finished systemd-update-utmp-runlevel.service. Sep 13 00:49:31.082509 systemd[1]: Finished user-runtime-dir@500.service. Sep 13 00:49:31.085312 systemd[1]: Starting user@500.service... Sep 13 00:49:31.092352 (systemd)[1262]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.211513 systemd[1262]: Queued start job for default target default.target. Sep 13 00:49:31.212734 systemd[1262]: Reached target paths.target. Sep 13 00:49:31.212762 systemd[1262]: Reached target sockets.target. Sep 13 00:49:31.212783 systemd[1262]: Reached target timers.target. Sep 13 00:49:31.212802 systemd[1262]: Reached target basic.target. Sep 13 00:49:31.212957 systemd[1]: Started user@500.service. Sep 13 00:49:31.215238 systemd[1]: Started session-1.scope. Sep 13 00:49:31.216034 systemd[1]: Startup finished in 1.009s (kernel) + 5.781s (initrd) + 9.035s (userspace) = 15.826s. Sep 13 00:49:31.236897 systemd[1262]: Reached target default.target. Sep 13 00:49:31.237022 systemd[1262]: Startup finished in 134ms. Sep 13 00:49:31.312879 systemd[1]: Started sshd@1-24.144.95.58:22-147.75.109.163:47208.service. Sep 13 00:49:31.405243 sshd[1276]: Accepted publickey for core from 147.75.109.163 port 47208 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:49:31.407565 sshd[1276]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.416821 systemd[1]: Started session-2.scope. Sep 13 00:49:31.419071 systemd-logind[1182]: New session 2 of user core. Sep 13 00:49:31.490551 sshd[1276]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:31.493904 systemd[1]: sshd@1-24.144.95.58:22-147.75.109.163:47208.service: Deactivated successfully. Sep 13 00:49:31.498246 systemd[1]: session-2.scope: Deactivated successfully. Sep 13 00:49:31.499932 systemd-logind[1182]: Session 2 logged out. Waiting for processes to exit. Sep 13 00:49:31.502211 systemd[1]: Started sshd@2-24.144.95.58:22-147.75.109.163:47224.service. Sep 13 00:49:31.506200 systemd-logind[1182]: Removed session 2. Sep 13 00:49:31.558693 sshd[1282]: Accepted publickey for core from 147.75.109.163 port 47224 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:49:31.561243 sshd[1282]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.569284 systemd[1]: Started session-3.scope. Sep 13 00:49:31.570487 systemd-logind[1182]: New session 3 of user core. Sep 13 00:49:31.638662 sshd[1282]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:31.644097 systemd[1]: Started sshd@3-24.144.95.58:22-147.75.109.163:47240.service. Sep 13 00:49:31.648576 systemd[1]: sshd@2-24.144.95.58:22-147.75.109.163:47224.service: Deactivated successfully. Sep 13 00:49:31.649543 systemd[1]: session-3.scope: Deactivated successfully. Sep 13 00:49:31.650019 systemd-logind[1182]: Session 3 logged out. Waiting for processes to exit. Sep 13 00:49:31.652220 systemd-logind[1182]: Removed session 3. Sep 13 00:49:31.709748 sshd[1287]: Accepted publickey for core from 147.75.109.163 port 47240 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:49:31.711184 sshd[1287]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.718436 systemd[1]: Started session-4.scope. Sep 13 00:49:31.719215 systemd-logind[1182]: New session 4 of user core. Sep 13 00:49:31.789908 sshd[1287]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:31.796462 systemd[1]: sshd@3-24.144.95.58:22-147.75.109.163:47240.service: Deactivated successfully. Sep 13 00:49:31.797182 systemd[1]: session-4.scope: Deactivated successfully. Sep 13 00:49:31.798347 systemd-logind[1182]: Session 4 logged out. Waiting for processes to exit. Sep 13 00:49:31.800074 systemd[1]: Started sshd@4-24.144.95.58:22-147.75.109.163:47244.service. Sep 13 00:49:31.802142 systemd-logind[1182]: Removed session 4. Sep 13 00:49:31.859148 sshd[1294]: Accepted publickey for core from 147.75.109.163 port 47244 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:49:31.860813 sshd[1294]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:31.867286 systemd[1]: Started session-5.scope. Sep 13 00:49:31.867820 systemd-logind[1182]: New session 5 of user core. Sep 13 00:49:31.947072 kubelet[1259]: E0913 00:49:31.947017 1259 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:49:31.949404 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:49:31.949639 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:49:31.950007 systemd[1]: kubelet.service: Consumed 1.693s CPU time. Sep 13 00:49:31.954798 sudo[1297]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 13 00:49:31.955218 sudo[1297]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 13 00:49:31.986225 systemd[1]: Starting docker.service... Sep 13 00:49:32.056582 env[1307]: time="2025-09-13T00:49:32.056520746Z" level=info msg="Starting up" Sep 13 00:49:32.058772 env[1307]: time="2025-09-13T00:49:32.058670370Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:49:32.058772 env[1307]: time="2025-09-13T00:49:32.058767292Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:49:32.058971 env[1307]: time="2025-09-13T00:49:32.058799531Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:49:32.058971 env[1307]: time="2025-09-13T00:49:32.058816330Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:49:32.061392 env[1307]: time="2025-09-13T00:49:32.061358555Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:49:32.061593 env[1307]: time="2025-09-13T00:49:32.061567508Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:49:32.061772 env[1307]: time="2025-09-13T00:49:32.061740354Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:49:32.061901 env[1307]: time="2025-09-13T00:49:32.061875799Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:49:32.070653 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2131478452-merged.mount: Deactivated successfully. Sep 13 00:49:32.206948 env[1307]: time="2025-09-13T00:49:32.206782141Z" level=info msg="Loading containers: start." Sep 13 00:49:32.406641 kernel: Initializing XFRM netlink socket Sep 13 00:49:32.463486 env[1307]: time="2025-09-13T00:49:32.463340082Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Sep 13 00:49:32.568032 systemd-networkd[1014]: docker0: Link UP Sep 13 00:49:32.587210 env[1307]: time="2025-09-13T00:49:32.587157982Z" level=info msg="Loading containers: done." Sep 13 00:49:32.603548 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1119469884-merged.mount: Deactivated successfully. Sep 13 00:49:32.606529 env[1307]: time="2025-09-13T00:49:32.606483449Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 13 00:49:32.607165 env[1307]: time="2025-09-13T00:49:32.607135020Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Sep 13 00:49:32.607383 env[1307]: time="2025-09-13T00:49:32.607365225Z" level=info msg="Daemon has completed initialization" Sep 13 00:49:32.626280 systemd[1]: Started docker.service. Sep 13 00:49:32.635161 env[1307]: time="2025-09-13T00:49:32.635089571Z" level=info msg="API listen on /run/docker.sock" Sep 13 00:49:32.657039 systemd[1]: Starting coreos-metadata.service... Sep 13 00:49:32.721797 coreos-metadata[1425]: Sep 13 00:49:32.721 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Sep 13 00:49:32.733643 coreos-metadata[1425]: Sep 13 00:49:32.733 INFO Fetch successful Sep 13 00:49:32.752084 systemd[1]: Finished coreos-metadata.service. Sep 13 00:49:33.853272 env[1189]: time="2025-09-13T00:49:33.853196880Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\"" Sep 13 00:49:34.362834 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount653943800.mount: Deactivated successfully. Sep 13 00:49:36.401963 env[1189]: time="2025-09-13T00:49:36.401881551Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:36.405554 env[1189]: time="2025-09-13T00:49:36.405496722Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:36.411457 env[1189]: time="2025-09-13T00:49:36.411405598Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:36.413845 env[1189]: time="2025-09-13T00:49:36.413783138Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:36.417341 env[1189]: time="2025-09-13T00:49:36.417290044Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\" returns image reference \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\"" Sep 13 00:49:36.418040 env[1189]: time="2025-09-13T00:49:36.417999213Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\"" Sep 13 00:49:38.661472 env[1189]: time="2025-09-13T00:49:38.661404993Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:38.663292 env[1189]: time="2025-09-13T00:49:38.663242166Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:38.665934 env[1189]: time="2025-09-13T00:49:38.665885541Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:38.667707 env[1189]: time="2025-09-13T00:49:38.667655545Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:38.670097 env[1189]: time="2025-09-13T00:49:38.670037327Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\" returns image reference \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\"" Sep 13 00:49:38.670623 env[1189]: time="2025-09-13T00:49:38.670569500Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\"" Sep 13 00:49:40.354081 env[1189]: time="2025-09-13T00:49:40.354001356Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:40.357785 env[1189]: time="2025-09-13T00:49:40.357716013Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:40.361598 env[1189]: time="2025-09-13T00:49:40.360083046Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:40.364203 env[1189]: time="2025-09-13T00:49:40.364132451Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\" returns image reference \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\"" Sep 13 00:49:40.364683 env[1189]: time="2025-09-13T00:49:40.364649970Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\"" Sep 13 00:49:40.364916 env[1189]: time="2025-09-13T00:49:40.363254102Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:41.653913 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2344411034.mount: Deactivated successfully. Sep 13 00:49:42.200747 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 13 00:49:42.200974 systemd[1]: Stopped kubelet.service. Sep 13 00:49:42.201028 systemd[1]: kubelet.service: Consumed 1.693s CPU time. Sep 13 00:49:42.202975 systemd[1]: Starting kubelet.service... Sep 13 00:49:42.360530 systemd[1]: Started kubelet.service. Sep 13 00:49:42.430866 kubelet[1448]: E0913 00:49:42.430815 1448 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:49:42.435637 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:49:42.435780 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:49:42.880276 env[1189]: time="2025-09-13T00:49:42.880184220Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:42.881851 env[1189]: time="2025-09-13T00:49:42.881811731Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:42.883257 env[1189]: time="2025-09-13T00:49:42.883215795Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.33.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:42.885274 env[1189]: time="2025-09-13T00:49:42.885224850Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:42.886050 env[1189]: time="2025-09-13T00:49:42.885994685Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\" returns image reference \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\"" Sep 13 00:49:42.886927 env[1189]: time="2025-09-13T00:49:42.886891767Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 13 00:49:43.440426 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1772207140.mount: Deactivated successfully. Sep 13 00:49:45.051536 env[1189]: time="2025-09-13T00:49:45.051460309Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.12.0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.055193 env[1189]: time="2025-09-13T00:49:45.055123003Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.058817 env[1189]: time="2025-09-13T00:49:45.058750775Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.12.0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.061676 env[1189]: time="2025-09-13T00:49:45.061595426Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.063020 env[1189]: time="2025-09-13T00:49:45.062964705Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 13 00:49:45.063855 env[1189]: time="2025-09-13T00:49:45.063819630Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 13 00:49:45.526933 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount465054515.mount: Deactivated successfully. Sep 13 00:49:45.532129 env[1189]: time="2025-09-13T00:49:45.532032314Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.534801 env[1189]: time="2025-09-13T00:49:45.534726553Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.537000 env[1189]: time="2025-09-13T00:49:45.536926281Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.538521 env[1189]: time="2025-09-13T00:49:45.538470941Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:45.539186 env[1189]: time="2025-09-13T00:49:45.539139074Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 13 00:49:45.539897 env[1189]: time="2025-09-13T00:49:45.539859198Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 13 00:49:45.971251 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2742821287.mount: Deactivated successfully. Sep 13 00:49:48.862003 env[1189]: time="2025-09-13T00:49:48.861949821Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.21-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:48.866163 env[1189]: time="2025-09-13T00:49:48.866095375Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:48.868982 env[1189]: time="2025-09-13T00:49:48.868920558Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.21-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:48.874198 env[1189]: time="2025-09-13T00:49:48.874144466Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:48.876371 env[1189]: time="2025-09-13T00:49:48.875585147Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 13 00:49:52.686901 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 13 00:49:52.687168 systemd[1]: Stopped kubelet.service. Sep 13 00:49:52.692035 systemd[1]: Starting kubelet.service... Sep 13 00:49:53.094837 systemd[1]: Started kubelet.service. Sep 13 00:49:53.170675 kubelet[1478]: E0913 00:49:53.170077 1478 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:49:53.173291 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:49:53.173440 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:49:54.610451 systemd[1]: Stopped kubelet.service. Sep 13 00:49:54.614387 systemd[1]: Starting kubelet.service... Sep 13 00:49:54.654381 systemd[1]: Reloading. Sep 13 00:49:54.797243 /usr/lib/systemd/system-generators/torcx-generator[1509]: time="2025-09-13T00:49:54Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:49:54.797276 /usr/lib/systemd/system-generators/torcx-generator[1509]: time="2025-09-13T00:49:54Z" level=info msg="torcx already run" Sep 13 00:49:54.900452 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:49:54.900475 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:49:54.925579 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:49:55.046851 systemd[1]: Started kubelet.service. Sep 13 00:49:55.050466 systemd[1]: Stopping kubelet.service... Sep 13 00:49:55.051226 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:49:55.051490 systemd[1]: Stopped kubelet.service. Sep 13 00:49:55.054288 systemd[1]: Starting kubelet.service... Sep 13 00:49:55.187326 systemd[1]: Started kubelet.service. Sep 13 00:49:55.244925 kubelet[1563]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:49:55.245451 kubelet[1563]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 13 00:49:55.245525 kubelet[1563]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:49:55.245724 kubelet[1563]: I0913 00:49:55.245680 1563 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:49:56.089676 kubelet[1563]: I0913 00:49:56.089512 1563 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 13 00:49:56.089676 kubelet[1563]: I0913 00:49:56.089582 1563 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:49:56.090030 kubelet[1563]: I0913 00:49:56.089993 1563 server.go:956] "Client rotation is on, will bootstrap in background" Sep 13 00:49:56.135207 kubelet[1563]: E0913 00:49:56.135146 1563 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://24.144.95.58:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 13 00:49:56.135465 kubelet[1563]: I0913 00:49:56.135441 1563 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:49:56.150218 kubelet[1563]: E0913 00:49:56.150109 1563 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:49:56.150218 kubelet[1563]: I0913 00:49:56.150194 1563 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:49:56.155956 kubelet[1563]: I0913 00:49:56.155902 1563 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:49:56.156338 kubelet[1563]: I0913 00:49:56.156280 1563 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:49:56.156661 kubelet[1563]: I0913 00:49:56.156338 1563 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510.3.8-n-1fe98121a2","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 00:49:56.156661 kubelet[1563]: I0913 00:49:56.156662 1563 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:49:56.156892 kubelet[1563]: I0913 00:49:56.156681 1563 container_manager_linux.go:303] "Creating device plugin manager" Sep 13 00:49:56.156892 kubelet[1563]: I0913 00:49:56.156864 1563 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:49:56.162025 kubelet[1563]: I0913 00:49:56.161961 1563 kubelet.go:480] "Attempting to sync node with API server" Sep 13 00:49:56.162199 kubelet[1563]: I0913 00:49:56.162049 1563 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:49:56.162199 kubelet[1563]: I0913 00:49:56.162098 1563 kubelet.go:386] "Adding apiserver pod source" Sep 13 00:49:56.162199 kubelet[1563]: I0913 00:49:56.162116 1563 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:49:56.182523 kubelet[1563]: E0913 00:49:56.182460 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://24.144.95.58:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.8-n-1fe98121a2&limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 13 00:49:56.183933 kubelet[1563]: E0913 00:49:56.183890 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://24.144.95.58:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 13 00:49:56.184262 kubelet[1563]: I0913 00:49:56.184239 1563 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:49:56.185379 kubelet[1563]: I0913 00:49:56.185350 1563 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 13 00:49:56.186413 kubelet[1563]: W0913 00:49:56.186387 1563 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 13 00:49:56.192356 kubelet[1563]: I0913 00:49:56.192324 1563 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 13 00:49:56.192644 kubelet[1563]: I0913 00:49:56.192598 1563 server.go:1289] "Started kubelet" Sep 13 00:49:56.205302 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Sep 13 00:49:56.205698 kubelet[1563]: I0913 00:49:56.205641 1563 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:49:56.206690 kubelet[1563]: I0913 00:49:56.205819 1563 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:49:56.208006 kubelet[1563]: I0913 00:49:56.207976 1563 server.go:317] "Adding debug handlers to kubelet server" Sep 13 00:49:56.215533 kubelet[1563]: I0913 00:49:56.215449 1563 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:49:56.215912 kubelet[1563]: I0913 00:49:56.215886 1563 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:49:56.216005 kubelet[1563]: I0913 00:49:56.215985 1563 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:49:56.217666 kubelet[1563]: E0913 00:49:56.204126 1563 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://24.144.95.58:6443/api/v1/namespaces/default/events\": dial tcp 24.144.95.58:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510.3.8-n-1fe98121a2.1864b1351906177c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510.3.8-n-1fe98121a2,UID:ci-3510.3.8-n-1fe98121a2,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510.3.8-n-1fe98121a2,},FirstTimestamp:2025-09-13 00:49:56.192548732 +0000 UTC m=+0.997708738,LastTimestamp:2025-09-13 00:49:56.192548732 +0000 UTC m=+0.997708738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510.3.8-n-1fe98121a2,}" Sep 13 00:49:56.217916 kubelet[1563]: I0913 00:49:56.217898 1563 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 13 00:49:56.218247 kubelet[1563]: E0913 00:49:56.218205 1563 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" Sep 13 00:49:56.220556 kubelet[1563]: I0913 00:49:56.220534 1563 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 13 00:49:56.220791 kubelet[1563]: I0913 00:49:56.220774 1563 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:49:56.221870 kubelet[1563]: I0913 00:49:56.221843 1563 factory.go:223] Registration of the systemd container factory successfully Sep 13 00:49:56.222091 kubelet[1563]: I0913 00:49:56.222070 1563 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:49:56.222541 kubelet[1563]: E0913 00:49:56.222511 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://24.144.95.58:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 13 00:49:56.223314 kubelet[1563]: E0913 00:49:56.223282 1563 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.95.58:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.8-n-1fe98121a2?timeout=10s\": dial tcp 24.144.95.58:6443: connect: connection refused" interval="200ms" Sep 13 00:49:56.234737 kubelet[1563]: E0913 00:49:56.234691 1563 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:49:56.236290 kubelet[1563]: I0913 00:49:56.236262 1563 factory.go:223] Registration of the containerd container factory successfully Sep 13 00:49:56.258197 kubelet[1563]: I0913 00:49:56.258117 1563 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 13 00:49:56.260362 kubelet[1563]: I0913 00:49:56.260308 1563 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 13 00:49:56.260362 kubelet[1563]: I0913 00:49:56.260349 1563 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 13 00:49:56.260556 kubelet[1563]: I0913 00:49:56.260384 1563 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 13 00:49:56.260556 kubelet[1563]: I0913 00:49:56.260396 1563 kubelet.go:2436] "Starting kubelet main sync loop" Sep 13 00:49:56.260556 kubelet[1563]: E0913 00:49:56.260468 1563 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:49:56.275880 kubelet[1563]: I0913 00:49:56.275843 1563 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 13 00:49:56.276227 kubelet[1563]: I0913 00:49:56.276175 1563 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 13 00:49:56.276391 kubelet[1563]: I0913 00:49:56.276375 1563 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:49:56.276590 kubelet[1563]: E0913 00:49:56.275891 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://24.144.95.58:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 13 00:49:56.280737 kubelet[1563]: I0913 00:49:56.280683 1563 policy_none.go:49] "None policy: Start" Sep 13 00:49:56.280967 kubelet[1563]: I0913 00:49:56.280947 1563 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 13 00:49:56.281119 kubelet[1563]: I0913 00:49:56.281103 1563 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:49:56.288923 systemd[1]: Created slice kubepods.slice. Sep 13 00:49:56.295957 systemd[1]: Created slice kubepods-burstable.slice. Sep 13 00:49:56.300061 systemd[1]: Created slice kubepods-besteffort.slice. Sep 13 00:49:56.307514 kubelet[1563]: E0913 00:49:56.307456 1563 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 13 00:49:56.308309 kubelet[1563]: I0913 00:49:56.308280 1563 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:49:56.308448 kubelet[1563]: I0913 00:49:56.308310 1563 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:49:56.310317 kubelet[1563]: I0913 00:49:56.310287 1563 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:49:56.311092 kubelet[1563]: E0913 00:49:56.311055 1563 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 13 00:49:56.311213 kubelet[1563]: E0913 00:49:56.311111 1563 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510.3.8-n-1fe98121a2\" not found" Sep 13 00:49:56.393380 systemd[1]: Created slice kubepods-burstable-pode3027084f5adea3d50ba8ec4e66c4d76.slice. Sep 13 00:49:56.409972 kubelet[1563]: E0913 00:49:56.409900 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.411236 kubelet[1563]: I0913 00:49:56.411176 1563 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.414084 systemd[1]: Created slice kubepods-burstable-pod7887b819beb8d2c4a1ef785b9058d9f0.slice. Sep 13 00:49:56.416628 kubelet[1563]: E0913 00:49:56.415959 1563 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://24.144.95.58:6443/api/v1/nodes\": dial tcp 24.144.95.58:6443: connect: connection refused" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.416628 kubelet[1563]: E0913 00:49:56.416814 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.420018 systemd[1]: Created slice kubepods-burstable-podb53712a4a68cdd6313314f5800aa4bf1.slice. Sep 13 00:49:56.421262 kubelet[1563]: I0913 00:49:56.421227 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421407 kubelet[1563]: I0913 00:49:56.421292 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-ca-certs\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421407 kubelet[1563]: I0913 00:49:56.421316 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421407 kubelet[1563]: I0913 00:49:56.421375 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421636 kubelet[1563]: I0913 00:49:56.421406 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421636 kubelet[1563]: I0913 00:49:56.421452 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b53712a4a68cdd6313314f5800aa4bf1-kubeconfig\") pod \"kube-scheduler-ci-3510.3.8-n-1fe98121a2\" (UID: \"b53712a4a68cdd6313314f5800aa4bf1\") " pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421636 kubelet[1563]: I0913 00:49:56.421469 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-k8s-certs\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421636 kubelet[1563]: I0913 00:49:56.421486 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-ca-certs\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.421636 kubelet[1563]: I0913 00:49:56.421523 1563 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.423287 kubelet[1563]: E0913 00:49:56.423249 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.424032 kubelet[1563]: E0913 00:49:56.423996 1563 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.95.58:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.8-n-1fe98121a2?timeout=10s\": dial tcp 24.144.95.58:6443: connect: connection refused" interval="400ms" Sep 13 00:49:56.592680 kubelet[1563]: E0913 00:49:56.592519 1563 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://24.144.95.58:6443/api/v1/namespaces/default/events\": dial tcp 24.144.95.58:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510.3.8-n-1fe98121a2.1864b1351906177c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510.3.8-n-1fe98121a2,UID:ci-3510.3.8-n-1fe98121a2,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510.3.8-n-1fe98121a2,},FirstTimestamp:2025-09-13 00:49:56.192548732 +0000 UTC m=+0.997708738,LastTimestamp:2025-09-13 00:49:56.192548732 +0000 UTC m=+0.997708738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510.3.8-n-1fe98121a2,}" Sep 13 00:49:56.618934 kubelet[1563]: I0913 00:49:56.618873 1563 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.619465 kubelet[1563]: E0913 00:49:56.619421 1563 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://24.144.95.58:6443/api/v1/nodes\": dial tcp 24.144.95.58:6443: connect: connection refused" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:56.714479 kubelet[1563]: E0913 00:49:56.713216 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:56.715999 env[1189]: time="2025-09-13T00:49:56.715903080Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.8-n-1fe98121a2,Uid:e3027084f5adea3d50ba8ec4e66c4d76,Namespace:kube-system,Attempt:0,}" Sep 13 00:49:56.718045 kubelet[1563]: E0913 00:49:56.718011 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:56.719765 env[1189]: time="2025-09-13T00:49:56.719723197Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.8-n-1fe98121a2,Uid:7887b819beb8d2c4a1ef785b9058d9f0,Namespace:kube-system,Attempt:0,}" Sep 13 00:49:56.724489 kubelet[1563]: E0913 00:49:56.724457 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:56.725045 env[1189]: time="2025-09-13T00:49:56.725009262Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.8-n-1fe98121a2,Uid:b53712a4a68cdd6313314f5800aa4bf1,Namespace:kube-system,Attempt:0,}" Sep 13 00:49:56.824702 kubelet[1563]: E0913 00:49:56.824660 1563 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.95.58:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.8-n-1fe98121a2?timeout=10s\": dial tcp 24.144.95.58:6443: connect: connection refused" interval="800ms" Sep 13 00:49:57.021674 kubelet[1563]: I0913 00:49:57.020950 1563 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:57.022488 kubelet[1563]: E0913 00:49:57.022447 1563 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://24.144.95.58:6443/api/v1/nodes\": dial tcp 24.144.95.58:6443: connect: connection refused" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:57.157153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount223547901.mount: Deactivated successfully. Sep 13 00:49:57.163870 env[1189]: time="2025-09-13T00:49:57.163802814Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.166901 env[1189]: time="2025-09-13T00:49:57.166853655Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.168378 env[1189]: time="2025-09-13T00:49:57.168309117Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.170423 env[1189]: time="2025-09-13T00:49:57.170370477Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.172396 env[1189]: time="2025-09-13T00:49:57.172356527Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.173414 env[1189]: time="2025-09-13T00:49:57.173374056Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.176284 env[1189]: time="2025-09-13T00:49:57.176231866Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.177206 env[1189]: time="2025-09-13T00:49:57.177172042Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.178509 env[1189]: time="2025-09-13T00:49:57.178474939Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.181514 env[1189]: time="2025-09-13T00:49:57.181442446Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.182362 env[1189]: time="2025-09-13T00:49:57.182334886Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.196895 env[1189]: time="2025-09-13T00:49:57.196819947Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:49:57.224913 env[1189]: time="2025-09-13T00:49:57.216973269Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:49:57.224913 env[1189]: time="2025-09-13T00:49:57.217035053Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:49:57.224913 env[1189]: time="2025-09-13T00:49:57.217138396Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:49:57.224913 env[1189]: time="2025-09-13T00:49:57.217822234Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/30ff3b13e5a3cb9f1bb034aa49632b3b1c19f3816cd57f36d8c3bc9610866081 pid=1611 runtime=io.containerd.runc.v2 Sep 13 00:49:57.229395 env[1189]: time="2025-09-13T00:49:57.229306760Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:49:57.229706 env[1189]: time="2025-09-13T00:49:57.229578196Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:49:57.229706 env[1189]: time="2025-09-13T00:49:57.229636418Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:49:57.230029 env[1189]: time="2025-09-13T00:49:57.229974931Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/edd78ed5af3d5a3d266f8fbefc0065a2d97fe0fe8c40342a4768964918514a29 pid=1622 runtime=io.containerd.runc.v2 Sep 13 00:49:57.251467 systemd[1]: Started cri-containerd-edd78ed5af3d5a3d266f8fbefc0065a2d97fe0fe8c40342a4768964918514a29.scope. Sep 13 00:49:57.263452 systemd[1]: Started cri-containerd-30ff3b13e5a3cb9f1bb034aa49632b3b1c19f3816cd57f36d8c3bc9610866081.scope. Sep 13 00:49:57.298199 env[1189]: time="2025-09-13T00:49:57.297850471Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:49:57.298199 env[1189]: time="2025-09-13T00:49:57.297914541Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:49:57.298199 env[1189]: time="2025-09-13T00:49:57.297926797Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:49:57.298463 env[1189]: time="2025-09-13T00:49:57.298221718Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/8c1c8c4d4739cabe51264f6da9fc098b9cdddf305e050f077cb8a0e9bb5c42a7 pid=1673 runtime=io.containerd.runc.v2 Sep 13 00:49:57.317913 systemd[1]: Started cri-containerd-8c1c8c4d4739cabe51264f6da9fc098b9cdddf305e050f077cb8a0e9bb5c42a7.scope. Sep 13 00:49:57.359421 env[1189]: time="2025-09-13T00:49:57.359354317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.8-n-1fe98121a2,Uid:7887b819beb8d2c4a1ef785b9058d9f0,Namespace:kube-system,Attempt:0,} returns sandbox id \"edd78ed5af3d5a3d266f8fbefc0065a2d97fe0fe8c40342a4768964918514a29\"" Sep 13 00:49:57.360991 kubelet[1563]: E0913 00:49:57.360958 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:57.370481 env[1189]: time="2025-09-13T00:49:57.370432237Z" level=info msg="CreateContainer within sandbox \"edd78ed5af3d5a3d266f8fbefc0065a2d97fe0fe8c40342a4768964918514a29\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 13 00:49:57.385851 env[1189]: time="2025-09-13T00:49:57.385803188Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.8-n-1fe98121a2,Uid:b53712a4a68cdd6313314f5800aa4bf1,Namespace:kube-system,Attempt:0,} returns sandbox id \"30ff3b13e5a3cb9f1bb034aa49632b3b1c19f3816cd57f36d8c3bc9610866081\"" Sep 13 00:49:57.386686 kubelet[1563]: E0913 00:49:57.386651 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:57.392113 env[1189]: time="2025-09-13T00:49:57.392052333Z" level=info msg="CreateContainer within sandbox \"30ff3b13e5a3cb9f1bb034aa49632b3b1c19f3816cd57f36d8c3bc9610866081\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 13 00:49:57.395454 env[1189]: time="2025-09-13T00:49:57.395391690Z" level=info msg="CreateContainer within sandbox \"edd78ed5af3d5a3d266f8fbefc0065a2d97fe0fe8c40342a4768964918514a29\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"736dd647baf887a162326206acc78095b867d12a463188ad84e021f7dadbd744\"" Sep 13 00:49:57.396294 env[1189]: time="2025-09-13T00:49:57.396259007Z" level=info msg="StartContainer for \"736dd647baf887a162326206acc78095b867d12a463188ad84e021f7dadbd744\"" Sep 13 00:49:57.408963 env[1189]: time="2025-09-13T00:49:57.408885782Z" level=info msg="CreateContainer within sandbox \"30ff3b13e5a3cb9f1bb034aa49632b3b1c19f3816cd57f36d8c3bc9610866081\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"b3bb4eb5d375d948980fca2cd63bbdd8be1fa5fff6d92bc1cb85be725a3c8d3f\"" Sep 13 00:49:57.409536 env[1189]: time="2025-09-13T00:49:57.409503804Z" level=info msg="StartContainer for \"b3bb4eb5d375d948980fca2cd63bbdd8be1fa5fff6d92bc1cb85be725a3c8d3f\"" Sep 13 00:49:57.415152 env[1189]: time="2025-09-13T00:49:57.415099105Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.8-n-1fe98121a2,Uid:e3027084f5adea3d50ba8ec4e66c4d76,Namespace:kube-system,Attempt:0,} returns sandbox id \"8c1c8c4d4739cabe51264f6da9fc098b9cdddf305e050f077cb8a0e9bb5c42a7\"" Sep 13 00:49:57.415942 kubelet[1563]: E0913 00:49:57.415910 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:57.420578 env[1189]: time="2025-09-13T00:49:57.420534045Z" level=info msg="CreateContainer within sandbox \"8c1c8c4d4739cabe51264f6da9fc098b9cdddf305e050f077cb8a0e9bb5c42a7\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 13 00:49:57.437982 systemd[1]: Started cri-containerd-736dd647baf887a162326206acc78095b867d12a463188ad84e021f7dadbd744.scope. Sep 13 00:49:57.440842 env[1189]: time="2025-09-13T00:49:57.436753215Z" level=info msg="CreateContainer within sandbox \"8c1c8c4d4739cabe51264f6da9fc098b9cdddf305e050f077cb8a0e9bb5c42a7\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0e50e623416653fa8bfb4220689d4cc1d732fb3c1ad47691a218275f183d3052\"" Sep 13 00:49:57.444518 env[1189]: time="2025-09-13T00:49:57.444473944Z" level=info msg="StartContainer for \"0e50e623416653fa8bfb4220689d4cc1d732fb3c1ad47691a218275f183d3052\"" Sep 13 00:49:57.451322 systemd[1]: Started cri-containerd-b3bb4eb5d375d948980fca2cd63bbdd8be1fa5fff6d92bc1cb85be725a3c8d3f.scope. Sep 13 00:49:57.481696 systemd[1]: Started cri-containerd-0e50e623416653fa8bfb4220689d4cc1d732fb3c1ad47691a218275f183d3052.scope. Sep 13 00:49:57.500888 kubelet[1563]: E0913 00:49:57.500813 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://24.144.95.58:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 13 00:49:57.535174 env[1189]: time="2025-09-13T00:49:57.535117607Z" level=info msg="StartContainer for \"736dd647baf887a162326206acc78095b867d12a463188ad84e021f7dadbd744\" returns successfully" Sep 13 00:49:57.540755 kubelet[1563]: E0913 00:49:57.540709 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://24.144.95.58:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 13 00:49:57.559359 env[1189]: time="2025-09-13T00:49:57.559219650Z" level=info msg="StartContainer for \"0e50e623416653fa8bfb4220689d4cc1d732fb3c1ad47691a218275f183d3052\" returns successfully" Sep 13 00:49:57.584431 env[1189]: time="2025-09-13T00:49:57.584355900Z" level=info msg="StartContainer for \"b3bb4eb5d375d948980fca2cd63bbdd8be1fa5fff6d92bc1cb85be725a3c8d3f\" returns successfully" Sep 13 00:49:57.611995 kubelet[1563]: E0913 00:49:57.611944 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://24.144.95.58:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.8-n-1fe98121a2&limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 13 00:49:57.613505 kubelet[1563]: E0913 00:49:57.613456 1563 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://24.144.95.58:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 13 00:49:57.626690 kubelet[1563]: E0913 00:49:57.626637 1563 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.144.95.58:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.8-n-1fe98121a2?timeout=10s\": dial tcp 24.144.95.58:6443: connect: connection refused" interval="1.6s" Sep 13 00:49:57.824544 kubelet[1563]: I0913 00:49:57.824416 1563 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:57.825141 kubelet[1563]: E0913 00:49:57.824838 1563 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://24.144.95.58:6443/api/v1/nodes\": dial tcp 24.144.95.58:6443: connect: connection refused" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:58.177632 kubelet[1563]: E0913 00:49:58.177450 1563 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://24.144.95.58:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 24.144.95.58:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 13 00:49:58.280502 kubelet[1563]: E0913 00:49:58.280454 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:58.280760 kubelet[1563]: E0913 00:49:58.280725 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:58.282939 kubelet[1563]: E0913 00:49:58.282896 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:58.283149 kubelet[1563]: E0913 00:49:58.283039 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:58.286998 kubelet[1563]: E0913 00:49:58.286959 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:58.287246 kubelet[1563]: E0913 00:49:58.287095 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:59.287629 kubelet[1563]: E0913 00:49:59.287564 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:59.288085 kubelet[1563]: E0913 00:49:59.287769 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:59.288579 kubelet[1563]: E0913 00:49:59.288516 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:49:59.288749 kubelet[1563]: E0913 00:49:59.288722 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:49:59.427332 kubelet[1563]: I0913 00:49:59.426721 1563 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.290147 kubelet[1563]: E0913 00:50:00.290109 1563 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.291193 kubelet[1563]: E0913 00:50:00.291161 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:00.572308 kubelet[1563]: E0913 00:50:00.572161 1563 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-3510.3.8-n-1fe98121a2\" not found" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.616748 kubelet[1563]: I0913 00:50:00.616683 1563 kubelet_node_status.go:78] "Successfully registered node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.617032 kubelet[1563]: E0913 00:50:00.617005 1563 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-3510.3.8-n-1fe98121a2\": node \"ci-3510.3.8-n-1fe98121a2\" not found" Sep 13 00:50:00.720115 kubelet[1563]: I0913 00:50:00.720061 1563 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.728848 kubelet[1563]: E0913 00:50:00.728796 1563 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.729209 kubelet[1563]: I0913 00:50:00.729183 1563 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.733597 kubelet[1563]: E0913 00:50:00.733544 1563 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.733908 kubelet[1563]: I0913 00:50:00.733883 1563 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:00.736743 kubelet[1563]: E0913 00:50:00.736695 1563 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-3510.3.8-n-1fe98121a2\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:01.186381 kubelet[1563]: I0913 00:50:01.185821 1563 apiserver.go:52] "Watching apiserver" Sep 13 00:50:01.221394 kubelet[1563]: I0913 00:50:01.221343 1563 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 13 00:50:04.620748 kubelet[1563]: I0913 00:50:04.620581 1563 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:04.652309 kubelet[1563]: I0913 00:50:04.652252 1563 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:04.652789 kubelet[1563]: E0913 00:50:04.652756 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:04.736474 systemd[1]: Reloading. Sep 13 00:50:04.976737 /usr/lib/systemd/system-generators/torcx-generator[1861]: time="2025-09-13T00:50:04Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:50:04.976806 /usr/lib/systemd/system-generators/torcx-generator[1861]: time="2025-09-13T00:50:04Z" level=info msg="torcx already run" Sep 13 00:50:05.141772 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:50:05.142063 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:50:05.183027 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:50:05.320293 kubelet[1563]: E0913 00:50:05.320246 1563 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:05.407786 kubelet[1563]: I0913 00:50:05.407701 1563 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:50:05.412942 systemd[1]: Stopping kubelet.service... Sep 13 00:50:05.438114 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:50:05.438412 systemd[1]: Stopped kubelet.service. Sep 13 00:50:05.438501 systemd[1]: kubelet.service: Consumed 1.557s CPU time. Sep 13 00:50:05.443383 systemd[1]: Starting kubelet.service... Sep 13 00:50:06.751075 systemd[1]: Started kubelet.service. Sep 13 00:50:06.857175 kubelet[1913]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:50:06.857921 kubelet[1913]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 13 00:50:06.858063 kubelet[1913]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:50:06.860470 kubelet[1913]: I0913 00:50:06.860358 1913 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:50:06.874661 kubelet[1913]: I0913 00:50:06.874541 1913 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 13 00:50:06.875022 kubelet[1913]: I0913 00:50:06.874996 1913 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:50:06.875674 kubelet[1913]: I0913 00:50:06.875591 1913 server.go:956] "Client rotation is on, will bootstrap in background" Sep 13 00:50:06.878554 kubelet[1913]: I0913 00:50:06.878493 1913 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 13 00:50:06.898767 kubelet[1913]: I0913 00:50:06.898358 1913 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:50:06.916286 kubelet[1913]: E0913 00:50:06.916226 1913 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:50:06.916774 kubelet[1913]: I0913 00:50:06.916743 1913 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:50:06.917372 sudo[1927]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 13 00:50:06.918839 sudo[1927]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Sep 13 00:50:06.923803 kubelet[1913]: I0913 00:50:06.923759 1913 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:50:06.925461 kubelet[1913]: I0913 00:50:06.924375 1913 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:50:06.925461 kubelet[1913]: I0913 00:50:06.924423 1913 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510.3.8-n-1fe98121a2","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 13 00:50:06.925461 kubelet[1913]: I0913 00:50:06.924643 1913 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:50:06.925461 kubelet[1913]: I0913 00:50:06.924659 1913 container_manager_linux.go:303] "Creating device plugin manager" Sep 13 00:50:06.925461 kubelet[1913]: I0913 00:50:06.924724 1913 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:50:06.925939 kubelet[1913]: I0913 00:50:06.924997 1913 kubelet.go:480] "Attempting to sync node with API server" Sep 13 00:50:06.925939 kubelet[1913]: I0913 00:50:06.925014 1913 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:50:06.925939 kubelet[1913]: I0913 00:50:06.925037 1913 kubelet.go:386] "Adding apiserver pod source" Sep 13 00:50:06.925939 kubelet[1913]: I0913 00:50:06.925053 1913 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:50:06.932033 kubelet[1913]: I0913 00:50:06.931690 1913 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:50:06.934305 kubelet[1913]: I0913 00:50:06.933936 1913 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 13 00:50:06.961221 kubelet[1913]: I0913 00:50:06.961150 1913 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 13 00:50:06.961427 kubelet[1913]: I0913 00:50:06.961266 1913 server.go:1289] "Started kubelet" Sep 13 00:50:06.972172 kubelet[1913]: I0913 00:50:06.971460 1913 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:50:06.974226 kubelet[1913]: I0913 00:50:06.974158 1913 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:50:06.976724 kubelet[1913]: I0913 00:50:06.976689 1913 server.go:317] "Adding debug handlers to kubelet server" Sep 13 00:50:06.987515 kubelet[1913]: I0913 00:50:06.985217 1913 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:50:06.987515 kubelet[1913]: I0913 00:50:06.985919 1913 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:50:06.988312 kubelet[1913]: I0913 00:50:06.988286 1913 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 13 00:50:06.997348 kubelet[1913]: I0913 00:50:06.997275 1913 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 13 00:50:07.002456 kubelet[1913]: I0913 00:50:07.002416 1913 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:50:07.006066 kubelet[1913]: I0913 00:50:07.003165 1913 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:50:07.015366 kubelet[1913]: I0913 00:50:07.015295 1913 factory.go:223] Registration of the systemd container factory successfully Sep 13 00:50:07.020828 kubelet[1913]: I0913 00:50:07.017434 1913 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:50:07.032322 kubelet[1913]: E0913 00:50:07.032285 1913 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:50:07.054992 kubelet[1913]: I0913 00:50:07.054937 1913 factory.go:223] Registration of the containerd container factory successfully Sep 13 00:50:07.055400 kubelet[1913]: I0913 00:50:07.055353 1913 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 13 00:50:07.057435 kubelet[1913]: I0913 00:50:07.057388 1913 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 13 00:50:07.057698 kubelet[1913]: I0913 00:50:07.057678 1913 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 13 00:50:07.057848 kubelet[1913]: I0913 00:50:07.057829 1913 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 13 00:50:07.057944 kubelet[1913]: I0913 00:50:07.057929 1913 kubelet.go:2436] "Starting kubelet main sync loop" Sep 13 00:50:07.058127 kubelet[1913]: E0913 00:50:07.058098 1913 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:50:07.158888 kubelet[1913]: E0913 00:50:07.158846 1913 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 00:50:07.175496 kubelet[1913]: I0913 00:50:07.175456 1913 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 13 00:50:07.175815 kubelet[1913]: I0913 00:50:07.175787 1913 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 13 00:50:07.175959 kubelet[1913]: I0913 00:50:07.175941 1913 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:50:07.176296 kubelet[1913]: I0913 00:50:07.176267 1913 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 13 00:50:07.176478 kubelet[1913]: I0913 00:50:07.176420 1913 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 13 00:50:07.176584 kubelet[1913]: I0913 00:50:07.176565 1913 policy_none.go:49] "None policy: Start" Sep 13 00:50:07.176725 kubelet[1913]: I0913 00:50:07.176708 1913 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 13 00:50:07.176834 kubelet[1913]: I0913 00:50:07.176821 1913 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:50:07.177105 kubelet[1913]: I0913 00:50:07.177079 1913 state_mem.go:75] "Updated machine memory state" Sep 13 00:50:07.185678 kubelet[1913]: E0913 00:50:07.185644 1913 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 13 00:50:07.190896 kubelet[1913]: I0913 00:50:07.190869 1913 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:50:07.191143 kubelet[1913]: I0913 00:50:07.191082 1913 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:50:07.191662 kubelet[1913]: I0913 00:50:07.191644 1913 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:50:07.203332 kubelet[1913]: E0913 00:50:07.203299 1913 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 13 00:50:07.302185 kubelet[1913]: I0913 00:50:07.302049 1913 kubelet_node_status.go:75] "Attempting to register node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.316273 kubelet[1913]: I0913 00:50:07.316221 1913 kubelet_node_status.go:124] "Node was previously registered" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.316483 kubelet[1913]: I0913 00:50:07.316387 1913 kubelet_node_status.go:78] "Successfully registered node" node="ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.360715 kubelet[1913]: I0913 00:50:07.360659 1913 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.361433 kubelet[1913]: I0913 00:50:07.361403 1913 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.362226 kubelet[1913]: I0913 00:50:07.362193 1913 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.373843 kubelet[1913]: I0913 00:50:07.373796 1913 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:07.390663 kubelet[1913]: I0913 00:50:07.390562 1913 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:07.390663 kubelet[1913]: E0913 00:50:07.390670 1913 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-3510.3.8-n-1fe98121a2\" already exists" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.391094 kubelet[1913]: I0913 00:50:07.391021 1913 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:07.426315 kubelet[1913]: I0913 00:50:07.426253 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-ca-certs\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426315 kubelet[1913]: I0913 00:50:07.426307 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426642 kubelet[1913]: I0913 00:50:07.426357 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426642 kubelet[1913]: I0913 00:50:07.426381 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-ca-certs\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426642 kubelet[1913]: I0913 00:50:07.426399 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-k8s-certs\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426642 kubelet[1913]: I0913 00:50:07.426414 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.426642 kubelet[1913]: I0913 00:50:07.426432 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7887b819beb8d2c4a1ef785b9058d9f0-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" (UID: \"7887b819beb8d2c4a1ef785b9058d9f0\") " pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.427051 kubelet[1913]: I0913 00:50:07.426450 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b53712a4a68cdd6313314f5800aa4bf1-kubeconfig\") pod \"kube-scheduler-ci-3510.3.8-n-1fe98121a2\" (UID: \"b53712a4a68cdd6313314f5800aa4bf1\") " pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.427051 kubelet[1913]: I0913 00:50:07.426476 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e3027084f5adea3d50ba8ec4e66c4d76-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" (UID: \"e3027084f5adea3d50ba8ec4e66c4d76\") " pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:07.674908 kubelet[1913]: E0913 00:50:07.674750 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:07.692001 kubelet[1913]: E0913 00:50:07.691949 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:07.692923 kubelet[1913]: E0913 00:50:07.692519 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:07.869125 sudo[1927]: pam_unix(sudo:session): session closed for user root Sep 13 00:50:07.941289 kubelet[1913]: I0913 00:50:07.941104 1913 apiserver.go:52] "Watching apiserver" Sep 13 00:50:07.998621 kubelet[1913]: I0913 00:50:07.998557 1913 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 13 00:50:08.048024 kubelet[1913]: I0913 00:50:08.047815 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" podStartSLOduration=1.04778718 podStartE2EDuration="1.04778718s" podCreationTimestamp="2025-09-13 00:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:08.02614807 +0000 UTC m=+1.256681417" watchObservedRunningTime="2025-09-13 00:50:08.04778718 +0000 UTC m=+1.278320525" Sep 13 00:50:08.048626 kubelet[1913]: I0913 00:50:08.048535 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" podStartSLOduration=1.048516977 podStartE2EDuration="1.048516977s" podCreationTimestamp="2025-09-13 00:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:08.048495664 +0000 UTC m=+1.279029025" watchObservedRunningTime="2025-09-13 00:50:08.048516977 +0000 UTC m=+1.279050321" Sep 13 00:50:08.120281 kubelet[1913]: E0913 00:50:08.120224 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:08.122049 kubelet[1913]: I0913 00:50:08.121996 1913 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:08.122351 kubelet[1913]: I0913 00:50:08.122329 1913 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:08.156829 kubelet[1913]: I0913 00:50:08.156772 1913 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:08.157322 kubelet[1913]: E0913 00:50:08.157293 1913 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-3510.3.8-n-1fe98121a2\" already exists" pod="kube-system/kube-controller-manager-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:08.157839 kubelet[1913]: E0913 00:50:08.157813 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:08.158595 kubelet[1913]: I0913 00:50:08.158461 1913 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Sep 13 00:50:08.158994 kubelet[1913]: E0913 00:50:08.158946 1913 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-3510.3.8-n-1fe98121a2\" already exists" pod="kube-system/kube-apiserver-ci-3510.3.8-n-1fe98121a2" Sep 13 00:50:08.159165 kubelet[1913]: E0913 00:50:08.159142 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:08.178293 kubelet[1913]: I0913 00:50:08.178195 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510.3.8-n-1fe98121a2" podStartSLOduration=4.178146387 podStartE2EDuration="4.178146387s" podCreationTimestamp="2025-09-13 00:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:08.124147303 +0000 UTC m=+1.354680646" watchObservedRunningTime="2025-09-13 00:50:08.178146387 +0000 UTC m=+1.408679730" Sep 13 00:50:08.976351 kubelet[1913]: I0913 00:50:08.976303 1913 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 13 00:50:08.977371 env[1189]: time="2025-09-13T00:50:08.977317364Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 13 00:50:08.978091 kubelet[1913]: I0913 00:50:08.978052 1913 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 13 00:50:09.123668 kubelet[1913]: E0913 00:50:09.123511 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:09.125987 kubelet[1913]: E0913 00:50:09.124978 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:09.125987 kubelet[1913]: E0913 00:50:09.125653 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:09.838409 systemd[1]: Created slice kubepods-besteffort-pod79104a44_a181_44dc_bda1_8fc333133fed.slice. Sep 13 00:50:09.876162 systemd[1]: Created slice kubepods-burstable-pod7cab6998_f239_454e_9ce4_f878e05c862b.slice. Sep 13 00:50:09.943704 kubelet[1913]: I0913 00:50:09.943642 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/79104a44-a181-44dc-bda1-8fc333133fed-kube-proxy\") pod \"kube-proxy-xjf5w\" (UID: \"79104a44-a181-44dc-bda1-8fc333133fed\") " pod="kube-system/kube-proxy-xjf5w" Sep 13 00:50:09.943704 kubelet[1913]: I0913 00:50:09.943696 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7nzb\" (UniqueName: \"kubernetes.io/projected/79104a44-a181-44dc-bda1-8fc333133fed-kube-api-access-n7nzb\") pod \"kube-proxy-xjf5w\" (UID: \"79104a44-a181-44dc-bda1-8fc333133fed\") " pod="kube-system/kube-proxy-xjf5w" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943766 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cni-path\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943797 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-xtables-lock\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943815 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-hubble-tls\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943841 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/79104a44-a181-44dc-bda1-8fc333133fed-lib-modules\") pod \"kube-proxy-xjf5w\" (UID: \"79104a44-a181-44dc-bda1-8fc333133fed\") " pod="kube-system/kube-proxy-xjf5w" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943863 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-hostproc\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.943935 kubelet[1913]: I0913 00:50:09.943878 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-cgroup\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944122 kubelet[1913]: I0913 00:50:09.943894 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-etc-cni-netd\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944122 kubelet[1913]: I0913 00:50:09.943916 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-lib-modules\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944122 kubelet[1913]: I0913 00:50:09.943939 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtzvf\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-kube-api-access-gtzvf\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944122 kubelet[1913]: I0913 00:50:09.943961 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/79104a44-a181-44dc-bda1-8fc333133fed-xtables-lock\") pod \"kube-proxy-xjf5w\" (UID: \"79104a44-a181-44dc-bda1-8fc333133fed\") " pod="kube-system/kube-proxy-xjf5w" Sep 13 00:50:09.944122 kubelet[1913]: I0913 00:50:09.943986 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7cab6998-f239-454e-9ce4-f878e05c862b-clustermesh-secrets\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944395 kubelet[1913]: I0913 00:50:09.944001 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-net\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944395 kubelet[1913]: I0913 00:50:09.944018 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-run\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944395 kubelet[1913]: I0913 00:50:09.944042 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-bpf-maps\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944395 kubelet[1913]: I0913 00:50:09.944065 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-config-path\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.944395 kubelet[1913]: I0913 00:50:09.944095 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-kernel\") pod \"cilium-gxsll\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " pod="kube-system/cilium-gxsll" Sep 13 00:50:09.957058 kubelet[1913]: E0913 00:50:09.956996 1913 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-gtzvf lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-gxsll" podUID="7cab6998-f239-454e-9ce4-f878e05c862b" Sep 13 00:50:10.002489 systemd[1]: Created slice kubepods-besteffort-pod245b2d08_f123_4a56_8b4d_0d4b7b5373d9.slice. Sep 13 00:50:10.045240 kubelet[1913]: I0913 00:50:10.045190 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-j9ltv\" (UID: \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\") " pod="kube-system/cilium-operator-6c4d7847fc-j9ltv" Sep 13 00:50:10.045240 kubelet[1913]: I0913 00:50:10.045237 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrjx\" (UniqueName: \"kubernetes.io/projected/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-kube-api-access-ksrjx\") pod \"cilium-operator-6c4d7847fc-j9ltv\" (UID: \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\") " pod="kube-system/cilium-operator-6c4d7847fc-j9ltv" Sep 13 00:50:10.045846 kubelet[1913]: I0913 00:50:10.045793 1913 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 13 00:50:10.246791 kubelet[1913]: I0913 00:50:10.246729 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-xtables-lock\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.247123 kubelet[1913]: I0913 00:50:10.247075 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-lib-modules\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.247369 kubelet[1913]: I0913 00:50:10.247321 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-kernel\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.247560 kubelet[1913]: I0913 00:50:10.247528 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7cab6998-f239-454e-9ce4-f878e05c862b-clustermesh-secrets\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.247843 kubelet[1913]: I0913 00:50:10.247804 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-bpf-maps\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248011 kubelet[1913]: I0913 00:50:10.247979 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-hubble-tls\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248137 kubelet[1913]: I0913 00:50:10.248116 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-etc-cni-netd\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248277 kubelet[1913]: I0913 00:50:10.248257 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtzvf\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-kube-api-access-gtzvf\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248402 kubelet[1913]: I0913 00:50:10.248387 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-hostproc\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248555 kubelet[1913]: I0913 00:50:10.248533 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-cgroup\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248730 kubelet[1913]: I0913 00:50:10.248707 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-run\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248858 kubelet[1913]: I0913 00:50:10.248840 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-config-path\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.248971 kubelet[1913]: I0913 00:50:10.248956 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-net\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.249073 kubelet[1913]: I0913 00:50:10.249058 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cni-path\") pod \"7cab6998-f239-454e-9ce4-f878e05c862b\" (UID: \"7cab6998-f239-454e-9ce4-f878e05c862b\") " Sep 13 00:50:10.250003 kubelet[1913]: I0913 00:50:10.246876 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250098 kubelet[1913]: I0913 00:50:10.247148 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250098 kubelet[1913]: I0913 00:50:10.247387 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250098 kubelet[1913]: I0913 00:50:10.249237 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cni-path" (OuterVolumeSpecName: "cni-path") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250098 kubelet[1913]: I0913 00:50:10.249257 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250286 kubelet[1913]: I0913 00:50:10.250091 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250286 kubelet[1913]: I0913 00:50:10.250120 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250286 kubelet[1913]: I0913 00:50:10.250140 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-hostproc" (OuterVolumeSpecName: "hostproc") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.250286 kubelet[1913]: I0913 00:50:10.250156 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.251849 kubelet[1913]: I0913 00:50:10.251789 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:50:10.254473 kubelet[1913]: I0913 00:50:10.254390 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 13 00:50:10.259691 kubelet[1913]: I0913 00:50:10.259639 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cab6998-f239-454e-9ce4-f878e05c862b-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 13 00:50:10.262087 kubelet[1913]: I0913 00:50:10.262031 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:50:10.267286 kubelet[1913]: I0913 00:50:10.267235 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-kube-api-access-gtzvf" (OuterVolumeSpecName: "kube-api-access-gtzvf") pod "7cab6998-f239-454e-9ce4-f878e05c862b" (UID: "7cab6998-f239-454e-9ce4-f878e05c862b"). InnerVolumeSpecName "kube-api-access-gtzvf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:50:10.306815 kubelet[1913]: E0913 00:50:10.306759 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:10.308805 env[1189]: time="2025-09-13T00:50:10.308277141Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-j9ltv,Uid:245b2d08-f123-4a56-8b4d-0d4b7b5373d9,Namespace:kube-system,Attempt:0,}" Sep 13 00:50:10.335426 env[1189]: time="2025-09-13T00:50:10.335267076Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:50:10.335426 env[1189]: time="2025-09-13T00:50:10.335343575Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:50:10.335426 env[1189]: time="2025-09-13T00:50:10.335360999Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:50:10.336310 env[1189]: time="2025-09-13T00:50:10.336142245Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471 pid=1986 runtime=io.containerd.runc.v2 Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350230 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-net\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350273 1913 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cni-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350284 1913 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-xtables-lock\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350301 1913 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-lib-modules\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350322 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-host-proc-sys-kernel\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350337 1913 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/7cab6998-f239-454e-9ce4-f878e05c862b-clustermesh-secrets\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350351 1913 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-bpf-maps\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.350786 kubelet[1913]: I0913 00:50:10.350365 1913 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-hubble-tls\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350403 1913 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-etc-cni-netd\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350419 1913 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-gtzvf\" (UniqueName: \"kubernetes.io/projected/7cab6998-f239-454e-9ce4-f878e05c862b-kube-api-access-gtzvf\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350432 1913 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-hostproc\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350451 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-cgroup\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350721 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-run\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.351257 kubelet[1913]: I0913 00:50:10.350740 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/7cab6998-f239-454e-9ce4-f878e05c862b-cilium-config-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:50:10.356247 systemd[1]: Started cri-containerd-3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471.scope. Sep 13 00:50:10.421114 env[1189]: time="2025-09-13T00:50:10.421054885Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-j9ltv,Uid:245b2d08-f123-4a56-8b4d-0d4b7b5373d9,Namespace:kube-system,Attempt:0,} returns sandbox id \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\"" Sep 13 00:50:10.422684 kubelet[1913]: E0913 00:50:10.422284 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:10.425989 env[1189]: time="2025-09-13T00:50:10.425930288Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 13 00:50:10.445660 kubelet[1913]: E0913 00:50:10.445165 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:10.446725 env[1189]: time="2025-09-13T00:50:10.446670855Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xjf5w,Uid:79104a44-a181-44dc-bda1-8fc333133fed,Namespace:kube-system,Attempt:0,}" Sep 13 00:50:10.472287 env[1189]: time="2025-09-13T00:50:10.471937901Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:50:10.472287 env[1189]: time="2025-09-13T00:50:10.472013103Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:50:10.472287 env[1189]: time="2025-09-13T00:50:10.472026672Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:50:10.472742 env[1189]: time="2025-09-13T00:50:10.472297666Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b6cb20a5d5c0a35d8af535dff7cdda626b7f6d2f41890c87a27cb4323b205259 pid=2028 runtime=io.containerd.runc.v2 Sep 13 00:50:10.487264 systemd[1]: Started cri-containerd-b6cb20a5d5c0a35d8af535dff7cdda626b7f6d2f41890c87a27cb4323b205259.scope. Sep 13 00:50:10.532810 env[1189]: time="2025-09-13T00:50:10.531035241Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xjf5w,Uid:79104a44-a181-44dc-bda1-8fc333133fed,Namespace:kube-system,Attempt:0,} returns sandbox id \"b6cb20a5d5c0a35d8af535dff7cdda626b7f6d2f41890c87a27cb4323b205259\"" Sep 13 00:50:10.536009 kubelet[1913]: E0913 00:50:10.534298 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:10.543461 env[1189]: time="2025-09-13T00:50:10.543403475Z" level=info msg="CreateContainer within sandbox \"b6cb20a5d5c0a35d8af535dff7cdda626b7f6d2f41890c87a27cb4323b205259\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 13 00:50:10.565789 env[1189]: time="2025-09-13T00:50:10.565703854Z" level=info msg="CreateContainer within sandbox \"b6cb20a5d5c0a35d8af535dff7cdda626b7f6d2f41890c87a27cb4323b205259\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"a282315fdddf61be38388c99850517204a56a317a596b5a0fec0a3ce41e36432\"" Sep 13 00:50:10.569326 env[1189]: time="2025-09-13T00:50:10.568473887Z" level=info msg="StartContainer for \"a282315fdddf61be38388c99850517204a56a317a596b5a0fec0a3ce41e36432\"" Sep 13 00:50:10.597591 systemd[1]: Started cri-containerd-a282315fdddf61be38388c99850517204a56a317a596b5a0fec0a3ce41e36432.scope. Sep 13 00:50:10.649434 env[1189]: time="2025-09-13T00:50:10.649257607Z" level=info msg="StartContainer for \"a282315fdddf61be38388c99850517204a56a317a596b5a0fec0a3ce41e36432\" returns successfully" Sep 13 00:50:11.068833 systemd[1]: Removed slice kubepods-burstable-pod7cab6998_f239_454e_9ce4_f878e05c862b.slice. Sep 13 00:50:11.080854 systemd[1]: var-lib-kubelet-pods-7cab6998\x2df239\x2d454e\x2d9ce4\x2df878e05c862b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dgtzvf.mount: Deactivated successfully. Sep 13 00:50:11.081003 systemd[1]: var-lib-kubelet-pods-7cab6998\x2df239\x2d454e\x2d9ce4\x2df878e05c862b-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:50:11.081095 systemd[1]: var-lib-kubelet-pods-7cab6998\x2df239\x2d454e\x2d9ce4\x2df878e05c862b-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:50:11.128520 kubelet[1913]: E0913 00:50:11.128479 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:11.206403 kubelet[1913]: I0913 00:50:11.206299 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-xjf5w" podStartSLOduration=2.206237953 podStartE2EDuration="2.206237953s" podCreationTimestamp="2025-09-13 00:50:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:11.168211289 +0000 UTC m=+4.398744632" watchObservedRunningTime="2025-09-13 00:50:11.206237953 +0000 UTC m=+4.436771296" Sep 13 00:50:11.252711 systemd[1]: Created slice kubepods-burstable-pod274059f7_91b4_475c_b55e_e7ba251bd6a0.slice. Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357149 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-cgroup\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357215 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-config-path\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357247 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-hubble-tls\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357280 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-bpf-maps\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357306 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cni-path\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357371 kubelet[1913]: I0913 00:50:11.357328 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-lib-modules\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357356 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-xtables-lock\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357380 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-hostproc\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357406 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/274059f7-91b4-475c-b55e-e7ba251bd6a0-clustermesh-secrets\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357435 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-kernel\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357463 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-run\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.357873 kubelet[1913]: I0913 00:50:11.357489 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-etc-cni-netd\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.358147 kubelet[1913]: I0913 00:50:11.357523 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mk7n\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-kube-api-access-8mk7n\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.358147 kubelet[1913]: I0913 00:50:11.357553 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-net\") pod \"cilium-xchls\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " pod="kube-system/cilium-xchls" Sep 13 00:50:11.401640 kubelet[1913]: E0913 00:50:11.401557 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:11.557895 kubelet[1913]: E0913 00:50:11.557832 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:11.559230 env[1189]: time="2025-09-13T00:50:11.558708928Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xchls,Uid:274059f7-91b4-475c-b55e-e7ba251bd6a0,Namespace:kube-system,Attempt:0,}" Sep 13 00:50:11.586351 env[1189]: time="2025-09-13T00:50:11.586228395Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:50:11.586786 env[1189]: time="2025-09-13T00:50:11.586725369Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:50:11.587021 env[1189]: time="2025-09-13T00:50:11.586980452Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:50:11.587536 env[1189]: time="2025-09-13T00:50:11.587479172Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d pid=2240 runtime=io.containerd.runc.v2 Sep 13 00:50:11.603404 kubelet[1913]: E0913 00:50:11.601644 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:11.617625 systemd[1]: Started cri-containerd-9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d.scope. Sep 13 00:50:11.697158 env[1189]: time="2025-09-13T00:50:11.697088437Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xchls,Uid:274059f7-91b4-475c-b55e-e7ba251bd6a0,Namespace:kube-system,Attempt:0,} returns sandbox id \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\"" Sep 13 00:50:11.699301 kubelet[1913]: E0913 00:50:11.698686 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:12.178651 kubelet[1913]: E0913 00:50:12.176525 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:12.178651 kubelet[1913]: E0913 00:50:12.178097 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:13.064510 kubelet[1913]: I0913 00:50:13.064423 1913 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cab6998-f239-454e-9ce4-f878e05c862b" path="/var/lib/kubelet/pods/7cab6998-f239-454e-9ce4-f878e05c862b/volumes" Sep 13 00:50:13.178866 kubelet[1913]: E0913 00:50:13.178814 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:13.298850 kubelet[1913]: E0913 00:50:13.297444 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:13.302676 env[1189]: time="2025-09-13T00:50:13.301631494Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:13.312374 env[1189]: time="2025-09-13T00:50:13.311442290Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:13.315395 env[1189]: time="2025-09-13T00:50:13.315031897Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:13.317262 env[1189]: time="2025-09-13T00:50:13.317186161Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 13 00:50:13.332529 env[1189]: time="2025-09-13T00:50:13.332437357Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 13 00:50:13.338154 env[1189]: time="2025-09-13T00:50:13.338101843Z" level=info msg="CreateContainer within sandbox \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 13 00:50:13.362543 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount92638582.mount: Deactivated successfully. Sep 13 00:50:13.377855 env[1189]: time="2025-09-13T00:50:13.377337885Z" level=info msg="CreateContainer within sandbox \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\"" Sep 13 00:50:13.379088 env[1189]: time="2025-09-13T00:50:13.379031311Z" level=info msg="StartContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\"" Sep 13 00:50:13.415144 systemd[1]: Started cri-containerd-3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5.scope. Sep 13 00:50:13.475336 env[1189]: time="2025-09-13T00:50:13.475279261Z" level=info msg="StartContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" returns successfully" Sep 13 00:50:14.185652 kubelet[1913]: E0913 00:50:14.185588 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:14.186551 kubelet[1913]: E0913 00:50:14.186507 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:14.391391 update_engine[1183]: I0913 00:50:14.390686 1183 update_attempter.cc:509] Updating boot flags... Sep 13 00:50:14.427650 kubelet[1913]: I0913 00:50:14.427534 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-j9ltv" podStartSLOduration=2.528675956 podStartE2EDuration="5.427502598s" podCreationTimestamp="2025-09-13 00:50:09 +0000 UTC" firstStartedPulling="2025-09-13 00:50:10.425169353 +0000 UTC m=+3.655702684" lastFinishedPulling="2025-09-13 00:50:13.323995979 +0000 UTC m=+6.554529326" observedRunningTime="2025-09-13 00:50:14.336631039 +0000 UTC m=+7.567164376" watchObservedRunningTime="2025-09-13 00:50:14.427502598 +0000 UTC m=+7.658035952" Sep 13 00:50:15.203642 kubelet[1913]: E0913 00:50:15.203554 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:20.218726 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount236572847.mount: Deactivated successfully. Sep 13 00:50:25.277797 env[1189]: time="2025-09-13T00:50:25.277689899Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:25.280423 env[1189]: time="2025-09-13T00:50:25.280360616Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:25.283421 env[1189]: time="2025-09-13T00:50:25.283360570Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:50:25.284764 env[1189]: time="2025-09-13T00:50:25.284709238Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 13 00:50:25.297191 env[1189]: time="2025-09-13T00:50:25.297119025Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:50:25.317151 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2931384799.mount: Deactivated successfully. Sep 13 00:50:25.330907 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3005848033.mount: Deactivated successfully. Sep 13 00:50:25.336795 env[1189]: time="2025-09-13T00:50:25.336649951Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\"" Sep 13 00:50:25.339306 env[1189]: time="2025-09-13T00:50:25.338122868Z" level=info msg="StartContainer for \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\"" Sep 13 00:50:25.377932 systemd[1]: Started cri-containerd-46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33.scope. Sep 13 00:50:25.434303 env[1189]: time="2025-09-13T00:50:25.434187713Z" level=info msg="StartContainer for \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\" returns successfully" Sep 13 00:50:25.465730 systemd[1]: cri-containerd-46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33.scope: Deactivated successfully. Sep 13 00:50:25.505632 env[1189]: time="2025-09-13T00:50:25.505557556Z" level=info msg="shim disconnected" id=46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33 Sep 13 00:50:25.505936 env[1189]: time="2025-09-13T00:50:25.505697692Z" level=warning msg="cleaning up after shim disconnected" id=46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33 namespace=k8s.io Sep 13 00:50:25.505936 env[1189]: time="2025-09-13T00:50:25.505713086Z" level=info msg="cleaning up dead shim" Sep 13 00:50:25.516700 env[1189]: time="2025-09-13T00:50:25.516634238Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2377 runtime=io.containerd.runc.v2\n" Sep 13 00:50:26.228316 kubelet[1913]: E0913 00:50:26.228254 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:26.240439 env[1189]: time="2025-09-13T00:50:26.240011868Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:50:26.261011 env[1189]: time="2025-09-13T00:50:26.260909874Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\"" Sep 13 00:50:26.262670 env[1189]: time="2025-09-13T00:50:26.261875235Z" level=info msg="StartContainer for \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\"" Sep 13 00:50:26.288218 systemd[1]: Started cri-containerd-c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0.scope. Sep 13 00:50:26.313765 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33-rootfs.mount: Deactivated successfully. Sep 13 00:50:26.358900 env[1189]: time="2025-09-13T00:50:26.358818978Z" level=info msg="StartContainer for \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\" returns successfully" Sep 13 00:50:26.376543 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:50:26.378740 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:50:26.382771 systemd[1]: Stopping systemd-sysctl.service... Sep 13 00:50:26.386229 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:50:26.400854 systemd[1]: cri-containerd-c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0.scope: Deactivated successfully. Sep 13 00:50:26.421054 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0-rootfs.mount: Deactivated successfully. Sep 13 00:50:26.422579 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:50:26.434889 env[1189]: time="2025-09-13T00:50:26.434796557Z" level=info msg="shim disconnected" id=c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0 Sep 13 00:50:26.434889 env[1189]: time="2025-09-13T00:50:26.434867923Z" level=warning msg="cleaning up after shim disconnected" id=c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0 namespace=k8s.io Sep 13 00:50:26.434889 env[1189]: time="2025-09-13T00:50:26.434883397Z" level=info msg="cleaning up dead shim" Sep 13 00:50:26.449952 env[1189]: time="2025-09-13T00:50:26.449856212Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:26Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2442 runtime=io.containerd.runc.v2\n" Sep 13 00:50:27.241849 kubelet[1913]: E0913 00:50:27.241676 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:27.255552 env[1189]: time="2025-09-13T00:50:27.255455460Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:50:27.311888 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1024775341.mount: Deactivated successfully. Sep 13 00:50:27.313660 env[1189]: time="2025-09-13T00:50:27.313571069Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\"" Sep 13 00:50:27.317671 env[1189]: time="2025-09-13T00:50:27.317590272Z" level=info msg="StartContainer for \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\"" Sep 13 00:50:27.356828 systemd[1]: run-containerd-runc-k8s.io-a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8-runc.3YtBIg.mount: Deactivated successfully. Sep 13 00:50:27.367045 systemd[1]: Started cri-containerd-a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8.scope. Sep 13 00:50:27.414054 env[1189]: time="2025-09-13T00:50:27.413938888Z" level=info msg="StartContainer for \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\" returns successfully" Sep 13 00:50:27.426285 systemd[1]: cri-containerd-a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8.scope: Deactivated successfully. Sep 13 00:50:27.482454 env[1189]: time="2025-09-13T00:50:27.482290813Z" level=info msg="shim disconnected" id=a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8 Sep 13 00:50:27.482454 env[1189]: time="2025-09-13T00:50:27.482426751Z" level=warning msg="cleaning up after shim disconnected" id=a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8 namespace=k8s.io Sep 13 00:50:27.482454 env[1189]: time="2025-09-13T00:50:27.482448322Z" level=info msg="cleaning up dead shim" Sep 13 00:50:27.497538 env[1189]: time="2025-09-13T00:50:27.497352488Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:27Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2501 runtime=io.containerd.runc.v2\n" Sep 13 00:50:28.246515 kubelet[1913]: E0913 00:50:28.246047 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:28.258479 env[1189]: time="2025-09-13T00:50:28.258406465Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:50:28.287787 env[1189]: time="2025-09-13T00:50:28.287684297Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\"" Sep 13 00:50:28.289750 env[1189]: time="2025-09-13T00:50:28.289685883Z" level=info msg="StartContainer for \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\"" Sep 13 00:50:28.312222 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8-rootfs.mount: Deactivated successfully. Sep 13 00:50:28.324594 systemd[1]: run-containerd-runc-k8s.io-bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73-runc.sOxPoQ.mount: Deactivated successfully. Sep 13 00:50:28.333416 systemd[1]: Started cri-containerd-bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73.scope. Sep 13 00:50:28.385911 systemd[1]: cri-containerd-bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73.scope: Deactivated successfully. Sep 13 00:50:28.388545 env[1189]: time="2025-09-13T00:50:28.387972932Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod274059f7_91b4_475c_b55e_e7ba251bd6a0.slice/cri-containerd-bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73.scope/memory.events\": no such file or directory" Sep 13 00:50:28.391848 env[1189]: time="2025-09-13T00:50:28.391773342Z" level=info msg="StartContainer for \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\" returns successfully" Sep 13 00:50:28.424167 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73-rootfs.mount: Deactivated successfully. Sep 13 00:50:28.432298 env[1189]: time="2025-09-13T00:50:28.432162762Z" level=info msg="shim disconnected" id=bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73 Sep 13 00:50:28.432298 env[1189]: time="2025-09-13T00:50:28.432213550Z" level=warning msg="cleaning up after shim disconnected" id=bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73 namespace=k8s.io Sep 13 00:50:28.432298 env[1189]: time="2025-09-13T00:50:28.432228339Z" level=info msg="cleaning up dead shim" Sep 13 00:50:28.449932 env[1189]: time="2025-09-13T00:50:28.449829085Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2556 runtime=io.containerd.runc.v2\n" Sep 13 00:50:29.253063 kubelet[1913]: E0913 00:50:29.253009 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:29.259841 env[1189]: time="2025-09-13T00:50:29.259773104Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:50:29.313188 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount310893493.mount: Deactivated successfully. Sep 13 00:50:29.317347 env[1189]: time="2025-09-13T00:50:29.317219808Z" level=info msg="CreateContainer within sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\"" Sep 13 00:50:29.321637 env[1189]: time="2025-09-13T00:50:29.321541343Z" level=info msg="StartContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\"" Sep 13 00:50:29.377381 systemd[1]: Started cri-containerd-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207.scope. Sep 13 00:50:29.389700 systemd[1]: run-containerd-runc-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-runc.vnhyT2.mount: Deactivated successfully. Sep 13 00:50:29.452376 env[1189]: time="2025-09-13T00:50:29.452293119Z" level=info msg="StartContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" returns successfully" Sep 13 00:50:29.707903 kubelet[1913]: I0913 00:50:29.706775 1913 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 13 00:50:29.777451 systemd[1]: Created slice kubepods-burstable-pod4b5a04a8_278c_4bb3_8773_24425470a846.slice. Sep 13 00:50:29.781991 systemd[1]: Created slice kubepods-burstable-pode5092b5e_c784_4082_a3a2_4e4f6d9dd472.slice. Sep 13 00:50:29.856192 kubelet[1913]: I0913 00:50:29.856113 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b5a04a8-278c-4bb3-8773-24425470a846-config-volume\") pod \"coredns-674b8bbfcf-8nhzp\" (UID: \"4b5a04a8-278c-4bb3-8773-24425470a846\") " pod="kube-system/coredns-674b8bbfcf-8nhzp" Sep 13 00:50:29.856662 kubelet[1913]: I0913 00:50:29.856635 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbd62\" (UniqueName: \"kubernetes.io/projected/4b5a04a8-278c-4bb3-8773-24425470a846-kube-api-access-rbd62\") pod \"coredns-674b8bbfcf-8nhzp\" (UID: \"4b5a04a8-278c-4bb3-8773-24425470a846\") " pod="kube-system/coredns-674b8bbfcf-8nhzp" Sep 13 00:50:29.957277 kubelet[1913]: I0913 00:50:29.957194 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xc5s\" (UniqueName: \"kubernetes.io/projected/e5092b5e-c784-4082-a3a2-4e4f6d9dd472-kube-api-access-7xc5s\") pod \"coredns-674b8bbfcf-hb686\" (UID: \"e5092b5e-c784-4082-a3a2-4e4f6d9dd472\") " pod="kube-system/coredns-674b8bbfcf-hb686" Sep 13 00:50:29.957583 kubelet[1913]: I0913 00:50:29.957562 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5092b5e-c784-4082-a3a2-4e4f6d9dd472-config-volume\") pod \"coredns-674b8bbfcf-hb686\" (UID: \"e5092b5e-c784-4082-a3a2-4e4f6d9dd472\") " pod="kube-system/coredns-674b8bbfcf-hb686" Sep 13 00:50:30.081677 kubelet[1913]: E0913 00:50:30.081586 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:30.083110 env[1189]: time="2025-09-13T00:50:30.083039622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-8nhzp,Uid:4b5a04a8-278c-4bb3-8773-24425470a846,Namespace:kube-system,Attempt:0,}" Sep 13 00:50:30.087271 kubelet[1913]: E0913 00:50:30.087220 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:30.088554 env[1189]: time="2025-09-13T00:50:30.088415021Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hb686,Uid:e5092b5e-c784-4082-a3a2-4e4f6d9dd472,Namespace:kube-system,Attempt:0,}" Sep 13 00:50:30.265177 kubelet[1913]: E0913 00:50:30.265118 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:30.297133 kubelet[1913]: I0913 00:50:30.297001 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-xchls" podStartSLOduration=5.711667621 podStartE2EDuration="19.296947303s" podCreationTimestamp="2025-09-13 00:50:11 +0000 UTC" firstStartedPulling="2025-09-13 00:50:11.701842194 +0000 UTC m=+4.932375522" lastFinishedPulling="2025-09-13 00:50:25.287121868 +0000 UTC m=+18.517655204" observedRunningTime="2025-09-13 00:50:30.294940259 +0000 UTC m=+23.525473605" watchObservedRunningTime="2025-09-13 00:50:30.296947303 +0000 UTC m=+23.527480662" Sep 13 00:50:31.267977 kubelet[1913]: E0913 00:50:31.267916 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:32.089471 systemd-networkd[1014]: cilium_host: Link UP Sep 13 00:50:32.094879 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Sep 13 00:50:32.096153 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Sep 13 00:50:32.095190 systemd-networkd[1014]: cilium_net: Link UP Sep 13 00:50:32.095470 systemd-networkd[1014]: cilium_net: Gained carrier Sep 13 00:50:32.095667 systemd-networkd[1014]: cilium_host: Gained carrier Sep 13 00:50:32.270451 kubelet[1913]: E0913 00:50:32.270398 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:32.279951 systemd-networkd[1014]: cilium_net: Gained IPv6LL Sep 13 00:50:32.294478 systemd-networkd[1014]: cilium_vxlan: Link UP Sep 13 00:50:32.294493 systemd-networkd[1014]: cilium_vxlan: Gained carrier Sep 13 00:50:32.731683 kernel: NET: Registered PF_ALG protocol family Sep 13 00:50:32.911933 systemd-networkd[1014]: cilium_host: Gained IPv6LL Sep 13 00:50:33.807914 systemd-networkd[1014]: cilium_vxlan: Gained IPv6LL Sep 13 00:50:33.985070 systemd-networkd[1014]: lxc_health: Link UP Sep 13 00:50:33.991907 systemd-networkd[1014]: lxc_health: Gained carrier Sep 13 00:50:33.992722 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:50:34.180553 systemd-networkd[1014]: lxc308e92484da0: Link UP Sep 13 00:50:34.185666 kernel: eth0: renamed from tmpf3cf6 Sep 13 00:50:34.192651 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc308e92484da0: link becomes ready Sep 13 00:50:34.192276 systemd-networkd[1014]: lxc308e92484da0: Gained carrier Sep 13 00:50:34.228563 systemd-networkd[1014]: lxc4709914992b7: Link UP Sep 13 00:50:34.237905 kernel: eth0: renamed from tmp8554a Sep 13 00:50:34.246660 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc4709914992b7: link becomes ready Sep 13 00:50:34.247682 systemd-networkd[1014]: lxc4709914992b7: Gained carrier Sep 13 00:50:34.309141 systemd[1]: run-containerd-runc-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-runc.VuyXoa.mount: Deactivated successfully. Sep 13 00:50:35.472070 systemd-networkd[1014]: lxc_health: Gained IPv6LL Sep 13 00:50:35.561493 kubelet[1913]: E0913 00:50:35.561438 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:35.984016 systemd-networkd[1014]: lxc308e92484da0: Gained IPv6LL Sep 13 00:50:36.303987 systemd-networkd[1014]: lxc4709914992b7: Gained IPv6LL Sep 13 00:50:36.583313 systemd[1]: run-containerd-runc-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-runc.yWfFeC.mount: Deactivated successfully. Sep 13 00:50:38.174835 kubelet[1913]: I0913 00:50:38.174691 1913 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 13 00:50:38.175535 kubelet[1913]: E0913 00:50:38.175331 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:38.293197 kubelet[1913]: E0913 00:50:38.293156 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:38.787698 systemd[1]: run-containerd-runc-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-runc.e5dAjz.mount: Deactivated successfully. Sep 13 00:50:40.866146 env[1189]: time="2025-09-13T00:50:40.866006024Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:50:40.866971 env[1189]: time="2025-09-13T00:50:40.866175644Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:50:40.866971 env[1189]: time="2025-09-13T00:50:40.866328004Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:50:40.866971 env[1189]: time="2025-09-13T00:50:40.866770716Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1 pid=3191 runtime=io.containerd.runc.v2 Sep 13 00:50:40.913227 systemd[1]: run-containerd-runc-k8s.io-f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1-runc.YrKtDG.mount: Deactivated successfully. Sep 13 00:50:40.932387 systemd[1]: Started cri-containerd-f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1.scope. Sep 13 00:50:40.955154 env[1189]: time="2025-09-13T00:50:40.947522592Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:50:40.956586 env[1189]: time="2025-09-13T00:50:40.955778295Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:50:40.956586 env[1189]: time="2025-09-13T00:50:40.955836426Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:50:40.956924 env[1189]: time="2025-09-13T00:50:40.956727262Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad pid=3214 runtime=io.containerd.runc.v2 Sep 13 00:50:41.001323 systemd[1]: Started cri-containerd-8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad.scope. Sep 13 00:50:41.174297 env[1189]: time="2025-09-13T00:50:41.170795481Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hb686,Uid:e5092b5e-c784-4082-a3a2-4e4f6d9dd472,Namespace:kube-system,Attempt:0,} returns sandbox id \"8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad\"" Sep 13 00:50:41.175281 kubelet[1913]: E0913 00:50:41.175207 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:41.196461 env[1189]: time="2025-09-13T00:50:41.196386657Z" level=info msg="CreateContainer within sandbox \"8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:50:41.203875 env[1189]: time="2025-09-13T00:50:41.203815638Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-8nhzp,Uid:4b5a04a8-278c-4bb3-8773-24425470a846,Namespace:kube-system,Attempt:0,} returns sandbox id \"f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1\"" Sep 13 00:50:41.205113 kubelet[1913]: E0913 00:50:41.205058 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:41.215454 env[1189]: time="2025-09-13T00:50:41.215138311Z" level=info msg="CreateContainer within sandbox \"f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:50:41.240025 env[1189]: time="2025-09-13T00:50:41.239952503Z" level=info msg="CreateContainer within sandbox \"f3cf680d8521a35192501dbec57e28cad468908311af4e1f43af993ebf80c7d1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"26aceba771e9e08f2260ef43b2e0493d8ba92e07bfceb63ab142e6cf9c211591\"" Sep 13 00:50:41.240861 env[1189]: time="2025-09-13T00:50:41.240817471Z" level=info msg="StartContainer for \"26aceba771e9e08f2260ef43b2e0493d8ba92e07bfceb63ab142e6cf9c211591\"" Sep 13 00:50:41.248896 env[1189]: time="2025-09-13T00:50:41.248805366Z" level=info msg="CreateContainer within sandbox \"8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b34247b684016e5a1a547925a7571866cb11dc2f04e826d77a9ef43269e30820\"" Sep 13 00:50:41.250015 env[1189]: time="2025-09-13T00:50:41.249963594Z" level=info msg="StartContainer for \"b34247b684016e5a1a547925a7571866cb11dc2f04e826d77a9ef43269e30820\"" Sep 13 00:50:41.291092 systemd[1]: Started cri-containerd-b34247b684016e5a1a547925a7571866cb11dc2f04e826d77a9ef43269e30820.scope. Sep 13 00:50:41.308987 systemd[1]: Started cri-containerd-26aceba771e9e08f2260ef43b2e0493d8ba92e07bfceb63ab142e6cf9c211591.scope. Sep 13 00:50:41.428393 env[1189]: time="2025-09-13T00:50:41.428202507Z" level=info msg="StartContainer for \"26aceba771e9e08f2260ef43b2e0493d8ba92e07bfceb63ab142e6cf9c211591\" returns successfully" Sep 13 00:50:41.437955 env[1189]: time="2025-09-13T00:50:41.437866698Z" level=info msg="StartContainer for \"b34247b684016e5a1a547925a7571866cb11dc2f04e826d77a9ef43269e30820\" returns successfully" Sep 13 00:50:41.875790 systemd[1]: run-containerd-runc-k8s.io-8554a65dbe1c985750658a3393ea29dbea8f978c90d531793e8119bc434064ad-runc.nu4gYX.mount: Deactivated successfully. Sep 13 00:50:42.330428 kubelet[1913]: E0913 00:50:42.330376 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:42.333987 kubelet[1913]: E0913 00:50:42.332529 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:42.361727 kubelet[1913]: I0913 00:50:42.361636 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hb686" podStartSLOduration=33.361598531 podStartE2EDuration="33.361598531s" podCreationTimestamp="2025-09-13 00:50:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:42.358001304 +0000 UTC m=+35.588534643" watchObservedRunningTime="2025-09-13 00:50:42.361598531 +0000 UTC m=+35.592131870" Sep 13 00:50:42.433863 kubelet[1913]: I0913 00:50:42.433757 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-8nhzp" podStartSLOduration=33.433710588 podStartE2EDuration="33.433710588s" podCreationTimestamp="2025-09-13 00:50:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:50:42.39235299 +0000 UTC m=+35.622886330" watchObservedRunningTime="2025-09-13 00:50:42.433710588 +0000 UTC m=+35.664243944" Sep 13 00:50:42.782013 sudo[1297]: pam_unix(sudo:session): session closed for user root Sep 13 00:50:42.796676 sshd[1294]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:42.815460 systemd[1]: sshd@4-24.144.95.58:22-147.75.109.163:47244.service: Deactivated successfully. Sep 13 00:50:42.816529 systemd[1]: session-5.scope: Deactivated successfully. Sep 13 00:50:42.816734 systemd[1]: session-5.scope: Consumed 9.566s CPU time. Sep 13 00:50:42.819283 systemd-logind[1182]: Session 5 logged out. Waiting for processes to exit. Sep 13 00:50:42.821308 systemd-logind[1182]: Removed session 5. Sep 13 00:50:43.335756 kubelet[1913]: E0913 00:50:43.335657 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:43.337576 kubelet[1913]: E0913 00:50:43.337529 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:44.338404 kubelet[1913]: E0913 00:50:44.338347 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:50:44.340392 kubelet[1913]: E0913 00:50:44.339709 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:16.775663 systemd[1]: Started sshd@5-24.144.95.58:22-185.156.73.233:28720.service. Sep 13 00:51:18.329257 sshd[3385]: Invalid user prueba from 185.156.73.233 port 28720 Sep 13 00:51:18.520855 sshd[3385]: pam_faillock(sshd:auth): User unknown Sep 13 00:51:18.522021 sshd[3385]: pam_unix(sshd:auth): check pass; user unknown Sep 13 00:51:18.522240 sshd[3385]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=185.156.73.233 Sep 13 00:51:18.523196 sshd[3385]: pam_faillock(sshd:auth): User unknown Sep 13 00:51:19.716273 systemd[1]: Started sshd@6-24.144.95.58:22-147.75.109.163:45456.service. Sep 13 00:51:19.774415 sshd[3388]: Accepted publickey for core from 147.75.109.163 port 45456 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:19.777093 sshd[3388]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:19.784170 systemd-logind[1182]: New session 6 of user core. Sep 13 00:51:19.785006 systemd[1]: Started session-6.scope. Sep 13 00:51:20.007443 sshd[3388]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:20.012694 systemd[1]: sshd@6-24.144.95.58:22-147.75.109.163:45456.service: Deactivated successfully. Sep 13 00:51:20.014136 systemd[1]: session-6.scope: Deactivated successfully. Sep 13 00:51:20.016420 systemd-logind[1182]: Session 6 logged out. Waiting for processes to exit. Sep 13 00:51:20.017934 systemd-logind[1182]: Removed session 6. Sep 13 00:51:20.765590 sshd[3385]: Failed password for invalid user prueba from 185.156.73.233 port 28720 ssh2 Sep 13 00:51:21.060389 kubelet[1913]: E0913 00:51:21.059825 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:21.534130 sshd[3385]: Connection closed by invalid user prueba 185.156.73.233 port 28720 [preauth] Sep 13 00:51:21.536058 systemd[1]: sshd@5-24.144.95.58:22-185.156.73.233:28720.service: Deactivated successfully. Sep 13 00:51:25.014056 systemd[1]: Started sshd@7-24.144.95.58:22-147.75.109.163:38664.service. Sep 13 00:51:25.073514 sshd[3402]: Accepted publickey for core from 147.75.109.163 port 38664 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:25.076700 sshd[3402]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:25.084431 systemd-logind[1182]: New session 7 of user core. Sep 13 00:51:25.085333 systemd[1]: Started session-7.scope. Sep 13 00:51:25.259758 sshd[3402]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:25.264586 systemd-logind[1182]: Session 7 logged out. Waiting for processes to exit. Sep 13 00:51:25.265084 systemd[1]: sshd@7-24.144.95.58:22-147.75.109.163:38664.service: Deactivated successfully. Sep 13 00:51:25.266265 systemd[1]: session-7.scope: Deactivated successfully. Sep 13 00:51:25.268597 systemd-logind[1182]: Removed session 7. Sep 13 00:51:30.269052 systemd[1]: Started sshd@8-24.144.95.58:22-147.75.109.163:35976.service. Sep 13 00:51:30.336707 sshd[3415]: Accepted publickey for core from 147.75.109.163 port 35976 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:30.340523 sshd[3415]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:30.348068 systemd[1]: Started session-8.scope. Sep 13 00:51:30.349372 systemd-logind[1182]: New session 8 of user core. Sep 13 00:51:30.526572 sshd[3415]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:30.531473 systemd[1]: sshd@8-24.144.95.58:22-147.75.109.163:35976.service: Deactivated successfully. Sep 13 00:51:30.532923 systemd[1]: session-8.scope: Deactivated successfully. Sep 13 00:51:30.533821 systemd-logind[1182]: Session 8 logged out. Waiting for processes to exit. Sep 13 00:51:30.535374 systemd-logind[1182]: Removed session 8. Sep 13 00:51:31.059403 kubelet[1913]: E0913 00:51:31.059351 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:35.535930 systemd[1]: Started sshd@9-24.144.95.58:22-147.75.109.163:35986.service. Sep 13 00:51:35.601057 sshd[3428]: Accepted publickey for core from 147.75.109.163 port 35986 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:35.603240 sshd[3428]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:35.611025 systemd[1]: Started session-9.scope. Sep 13 00:51:35.611863 systemd-logind[1182]: New session 9 of user core. Sep 13 00:51:35.772341 sshd[3428]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:35.777629 systemd[1]: sshd@9-24.144.95.58:22-147.75.109.163:35986.service: Deactivated successfully. Sep 13 00:51:35.779271 systemd[1]: session-9.scope: Deactivated successfully. Sep 13 00:51:35.780532 systemd-logind[1182]: Session 9 logged out. Waiting for processes to exit. Sep 13 00:51:35.783588 systemd[1]: Started sshd@10-24.144.95.58:22-147.75.109.163:35998.service. Sep 13 00:51:35.786353 systemd-logind[1182]: Removed session 9. Sep 13 00:51:35.847758 sshd[3440]: Accepted publickey for core from 147.75.109.163 port 35998 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:35.850564 sshd[3440]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:35.857780 systemd-logind[1182]: New session 10 of user core. Sep 13 00:51:35.859497 systemd[1]: Started session-10.scope. Sep 13 00:51:36.113434 sshd[3440]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:36.116857 systemd[1]: Started sshd@11-24.144.95.58:22-147.75.109.163:36006.service. Sep 13 00:51:36.127086 systemd[1]: sshd@10-24.144.95.58:22-147.75.109.163:35998.service: Deactivated successfully. Sep 13 00:51:36.130925 systemd[1]: session-10.scope: Deactivated successfully. Sep 13 00:51:36.134284 systemd-logind[1182]: Session 10 logged out. Waiting for processes to exit. Sep 13 00:51:36.137266 systemd-logind[1182]: Removed session 10. Sep 13 00:51:36.195436 sshd[3449]: Accepted publickey for core from 147.75.109.163 port 36006 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:36.198857 sshd[3449]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:36.207331 systemd-logind[1182]: New session 11 of user core. Sep 13 00:51:36.208299 systemd[1]: Started session-11.scope. Sep 13 00:51:36.441559 sshd[3449]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:36.449156 systemd[1]: sshd@11-24.144.95.58:22-147.75.109.163:36006.service: Deactivated successfully. Sep 13 00:51:36.450288 systemd[1]: session-11.scope: Deactivated successfully. Sep 13 00:51:36.451259 systemd-logind[1182]: Session 11 logged out. Waiting for processes to exit. Sep 13 00:51:36.452904 systemd-logind[1182]: Removed session 11. Sep 13 00:51:40.059448 kubelet[1913]: E0913 00:51:40.059390 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:41.451466 systemd[1]: Started sshd@12-24.144.95.58:22-147.75.109.163:48270.service. Sep 13 00:51:41.505424 sshd[3464]: Accepted publickey for core from 147.75.109.163 port 48270 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:41.508575 sshd[3464]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:41.516758 systemd[1]: Started session-12.scope. Sep 13 00:51:41.517732 systemd-logind[1182]: New session 12 of user core. Sep 13 00:51:41.695835 sshd[3464]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:41.700409 systemd[1]: sshd@12-24.144.95.58:22-147.75.109.163:48270.service: Deactivated successfully. Sep 13 00:51:41.701451 systemd[1]: session-12.scope: Deactivated successfully. Sep 13 00:51:41.703402 systemd-logind[1182]: Session 12 logged out. Waiting for processes to exit. Sep 13 00:51:41.704870 systemd-logind[1182]: Removed session 12. Sep 13 00:51:44.059712 kubelet[1913]: E0913 00:51:44.059655 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:45.059294 kubelet[1913]: E0913 00:51:45.059234 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:46.705246 systemd[1]: Started sshd@13-24.144.95.58:22-147.75.109.163:48276.service. Sep 13 00:51:46.774131 sshd[3476]: Accepted publickey for core from 147.75.109.163 port 48276 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:46.777345 sshd[3476]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:46.784802 systemd[1]: Started session-13.scope. Sep 13 00:51:46.785661 systemd-logind[1182]: New session 13 of user core. Sep 13 00:51:46.959859 sshd[3476]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:46.965305 systemd[1]: sshd@13-24.144.95.58:22-147.75.109.163:48276.service: Deactivated successfully. Sep 13 00:51:46.966305 systemd[1]: session-13.scope: Deactivated successfully. Sep 13 00:51:46.967213 systemd-logind[1182]: Session 13 logged out. Waiting for processes to exit. Sep 13 00:51:46.968376 systemd-logind[1182]: Removed session 13. Sep 13 00:51:48.059090 kubelet[1913]: E0913 00:51:48.059023 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:51.966660 systemd[1]: Started sshd@14-24.144.95.58:22-147.75.109.163:44010.service. Sep 13 00:51:52.024247 sshd[3488]: Accepted publickey for core from 147.75.109.163 port 44010 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:52.026937 sshd[3488]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:52.033947 systemd[1]: Started session-14.scope. Sep 13 00:51:52.034782 systemd-logind[1182]: New session 14 of user core. Sep 13 00:51:52.193484 sshd[3488]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:52.200366 systemd[1]: sshd@14-24.144.95.58:22-147.75.109.163:44010.service: Deactivated successfully. Sep 13 00:51:52.201918 systemd[1]: session-14.scope: Deactivated successfully. Sep 13 00:51:52.203325 systemd-logind[1182]: Session 14 logged out. Waiting for processes to exit. Sep 13 00:51:52.206542 systemd[1]: Started sshd@15-24.144.95.58:22-147.75.109.163:44016.service. Sep 13 00:51:52.209101 systemd-logind[1182]: Removed session 14. Sep 13 00:51:52.263766 sshd[3500]: Accepted publickey for core from 147.75.109.163 port 44016 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:52.266011 sshd[3500]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:52.272470 systemd[1]: Started session-15.scope. Sep 13 00:51:52.273401 systemd-logind[1182]: New session 15 of user core. Sep 13 00:51:52.638824 sshd[3500]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:52.645907 systemd[1]: Started sshd@16-24.144.95.58:22-147.75.109.163:44020.service. Sep 13 00:51:52.647904 systemd[1]: sshd@15-24.144.95.58:22-147.75.109.163:44016.service: Deactivated successfully. Sep 13 00:51:52.649485 systemd[1]: session-15.scope: Deactivated successfully. Sep 13 00:51:52.651167 systemd-logind[1182]: Session 15 logged out. Waiting for processes to exit. Sep 13 00:51:52.652880 systemd-logind[1182]: Removed session 15. Sep 13 00:51:52.717333 sshd[3509]: Accepted publickey for core from 147.75.109.163 port 44020 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:52.719845 sshd[3509]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:52.726070 systemd-logind[1182]: New session 16 of user core. Sep 13 00:51:52.726987 systemd[1]: Started session-16.scope. Sep 13 00:51:53.062679 kubelet[1913]: E0913 00:51:53.062640 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:51:53.602485 sshd[3509]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:53.610742 systemd[1]: sshd@16-24.144.95.58:22-147.75.109.163:44020.service: Deactivated successfully. Sep 13 00:51:53.613234 systemd[1]: session-16.scope: Deactivated successfully. Sep 13 00:51:53.615057 systemd-logind[1182]: Session 16 logged out. Waiting for processes to exit. Sep 13 00:51:53.619128 systemd[1]: Started sshd@17-24.144.95.58:22-147.75.109.163:44030.service. Sep 13 00:51:53.624943 systemd-logind[1182]: Removed session 16. Sep 13 00:51:53.682396 sshd[3525]: Accepted publickey for core from 147.75.109.163 port 44030 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:53.684863 sshd[3525]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:53.692746 systemd[1]: Started session-17.scope. Sep 13 00:51:53.693286 systemd-logind[1182]: New session 17 of user core. Sep 13 00:51:54.037869 sshd[3525]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:54.044650 systemd[1]: Started sshd@18-24.144.95.58:22-147.75.109.163:44036.service. Sep 13 00:51:54.046356 systemd[1]: sshd@17-24.144.95.58:22-147.75.109.163:44030.service: Deactivated successfully. Sep 13 00:51:54.048018 systemd[1]: session-17.scope: Deactivated successfully. Sep 13 00:51:54.051150 systemd-logind[1182]: Session 17 logged out. Waiting for processes to exit. Sep 13 00:51:54.054354 systemd-logind[1182]: Removed session 17. Sep 13 00:51:54.124348 sshd[3536]: Accepted publickey for core from 147.75.109.163 port 44036 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:54.127441 sshd[3536]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:54.134368 systemd[1]: Started session-18.scope. Sep 13 00:51:54.135021 systemd-logind[1182]: New session 18 of user core. Sep 13 00:51:54.298247 sshd[3536]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:54.302316 systemd[1]: sshd@18-24.144.95.58:22-147.75.109.163:44036.service: Deactivated successfully. Sep 13 00:51:54.303301 systemd[1]: session-18.scope: Deactivated successfully. Sep 13 00:51:54.304087 systemd-logind[1182]: Session 18 logged out. Waiting for processes to exit. Sep 13 00:51:54.305001 systemd-logind[1182]: Removed session 18. Sep 13 00:51:59.306900 systemd[1]: Started sshd@19-24.144.95.58:22-147.75.109.163:44044.service. Sep 13 00:51:59.365092 sshd[3548]: Accepted publickey for core from 147.75.109.163 port 44044 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:51:59.367785 sshd[3548]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:59.376297 systemd-logind[1182]: New session 19 of user core. Sep 13 00:51:59.376544 systemd[1]: Started session-19.scope. Sep 13 00:51:59.534485 sshd[3548]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:59.538056 systemd[1]: sshd@19-24.144.95.58:22-147.75.109.163:44044.service: Deactivated successfully. Sep 13 00:51:59.539084 systemd[1]: session-19.scope: Deactivated successfully. Sep 13 00:51:59.540201 systemd-logind[1182]: Session 19 logged out. Waiting for processes to exit. Sep 13 00:51:59.541507 systemd-logind[1182]: Removed session 19. Sep 13 00:52:04.543593 systemd[1]: Started sshd@20-24.144.95.58:22-147.75.109.163:33804.service. Sep 13 00:52:04.607647 sshd[3562]: Accepted publickey for core from 147.75.109.163 port 33804 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:04.610133 sshd[3562]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:04.622075 systemd-logind[1182]: New session 20 of user core. Sep 13 00:52:04.622862 systemd[1]: Started session-20.scope. Sep 13 00:52:04.802396 sshd[3562]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:04.806904 systemd-logind[1182]: Session 20 logged out. Waiting for processes to exit. Sep 13 00:52:04.807341 systemd[1]: sshd@20-24.144.95.58:22-147.75.109.163:33804.service: Deactivated successfully. Sep 13 00:52:04.808311 systemd[1]: session-20.scope: Deactivated successfully. Sep 13 00:52:04.809749 systemd-logind[1182]: Removed session 20. Sep 13 00:52:05.065055 kubelet[1913]: E0913 00:52:05.060771 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:09.809561 systemd[1]: Started sshd@21-24.144.95.58:22-147.75.109.163:33820.service. Sep 13 00:52:09.862752 sshd[3576]: Accepted publickey for core from 147.75.109.163 port 33820 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:09.863957 sshd[3576]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:09.870561 systemd-logind[1182]: New session 21 of user core. Sep 13 00:52:09.871572 systemd[1]: Started session-21.scope. Sep 13 00:52:10.032548 sshd[3576]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:10.036581 systemd-logind[1182]: Session 21 logged out. Waiting for processes to exit. Sep 13 00:52:10.036944 systemd[1]: sshd@21-24.144.95.58:22-147.75.109.163:33820.service: Deactivated successfully. Sep 13 00:52:10.038011 systemd[1]: session-21.scope: Deactivated successfully. Sep 13 00:52:10.039340 systemd-logind[1182]: Removed session 21. Sep 13 00:52:15.041428 systemd[1]: Started sshd@22-24.144.95.58:22-147.75.109.163:35656.service. Sep 13 00:52:15.103977 sshd[3590]: Accepted publickey for core from 147.75.109.163 port 35656 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:15.106345 sshd[3590]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:15.114259 systemd[1]: Started session-22.scope. Sep 13 00:52:15.115090 systemd-logind[1182]: New session 22 of user core. Sep 13 00:52:15.282992 sshd[3590]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:15.287685 systemd-logind[1182]: Session 22 logged out. Waiting for processes to exit. Sep 13 00:52:15.288670 systemd[1]: sshd@22-24.144.95.58:22-147.75.109.163:35656.service: Deactivated successfully. Sep 13 00:52:15.289631 systemd[1]: session-22.scope: Deactivated successfully. Sep 13 00:52:15.290894 systemd-logind[1182]: Removed session 22. Sep 13 00:52:20.289885 systemd[1]: Started sshd@23-24.144.95.58:22-147.75.109.163:38938.service. Sep 13 00:52:20.347227 sshd[3602]: Accepted publickey for core from 147.75.109.163 port 38938 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:20.349894 sshd[3602]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:20.356678 systemd[1]: Started session-23.scope. Sep 13 00:52:20.357477 systemd-logind[1182]: New session 23 of user core. Sep 13 00:52:20.511961 sshd[3602]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:20.520893 systemd[1]: Started sshd@24-24.144.95.58:22-147.75.109.163:38942.service. Sep 13 00:52:20.525330 systemd[1]: sshd@23-24.144.95.58:22-147.75.109.163:38938.service: Deactivated successfully. Sep 13 00:52:20.526597 systemd[1]: session-23.scope: Deactivated successfully. Sep 13 00:52:20.528414 systemd-logind[1182]: Session 23 logged out. Waiting for processes to exit. Sep 13 00:52:20.530084 systemd-logind[1182]: Removed session 23. Sep 13 00:52:20.592980 sshd[3613]: Accepted publickey for core from 147.75.109.163 port 38942 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:20.596009 sshd[3613]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:20.604308 systemd[1]: Started session-24.scope. Sep 13 00:52:20.604698 systemd-logind[1182]: New session 24 of user core. Sep 13 00:52:22.795738 env[1189]: time="2025-09-13T00:52:22.794795386Z" level=info msg="StopContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" with timeout 30 (s)" Sep 13 00:52:22.803462 env[1189]: time="2025-09-13T00:52:22.796508426Z" level=info msg="Stop container \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" with signal terminated" Sep 13 00:52:22.810448 systemd[1]: run-containerd-runc-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-runc.eRko0G.mount: Deactivated successfully. Sep 13 00:52:22.827642 systemd[1]: cri-containerd-3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5.scope: Deactivated successfully. Sep 13 00:52:22.864389 env[1189]: time="2025-09-13T00:52:22.864294621Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:52:22.874054 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5-rootfs.mount: Deactivated successfully. Sep 13 00:52:22.879206 env[1189]: time="2025-09-13T00:52:22.879143120Z" level=info msg="StopContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" with timeout 2 (s)" Sep 13 00:52:22.880830 env[1189]: time="2025-09-13T00:52:22.880774383Z" level=info msg="Stop container \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" with signal terminated" Sep 13 00:52:22.889595 env[1189]: time="2025-09-13T00:52:22.889309225Z" level=info msg="shim disconnected" id=3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5 Sep 13 00:52:22.890365 env[1189]: time="2025-09-13T00:52:22.890284579Z" level=warning msg="cleaning up after shim disconnected" id=3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5 namespace=k8s.io Sep 13 00:52:22.890751 env[1189]: time="2025-09-13T00:52:22.890716780Z" level=info msg="cleaning up dead shim" Sep 13 00:52:22.905049 systemd-networkd[1014]: lxc_health: Link DOWN Sep 13 00:52:22.905061 systemd-networkd[1014]: lxc_health: Lost carrier Sep 13 00:52:22.919350 env[1189]: time="2025-09-13T00:52:22.919295516Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3664 runtime=io.containerd.runc.v2\n" Sep 13 00:52:22.949522 systemd[1]: cri-containerd-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207.scope: Deactivated successfully. Sep 13 00:52:22.949874 systemd[1]: cri-containerd-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207.scope: Consumed 11.447s CPU time. Sep 13 00:52:22.959234 env[1189]: time="2025-09-13T00:52:22.959168655Z" level=info msg="StopContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" returns successfully" Sep 13 00:52:22.962332 env[1189]: time="2025-09-13T00:52:22.962274066Z" level=info msg="StopPodSandbox for \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\"" Sep 13 00:52:22.962505 env[1189]: time="2025-09-13T00:52:22.962377224Z" level=info msg="Container to stop \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:22.965646 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471-shm.mount: Deactivated successfully. Sep 13 00:52:22.979237 systemd[1]: cri-containerd-3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471.scope: Deactivated successfully. Sep 13 00:52:23.012841 env[1189]: time="2025-09-13T00:52:23.012778989Z" level=info msg="shim disconnected" id=06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207 Sep 13 00:52:23.013278 env[1189]: time="2025-09-13T00:52:23.013233232Z" level=warning msg="cleaning up after shim disconnected" id=06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207 namespace=k8s.io Sep 13 00:52:23.013399 env[1189]: time="2025-09-13T00:52:23.013379471Z" level=info msg="cleaning up dead shim" Sep 13 00:52:23.028544 env[1189]: time="2025-09-13T00:52:23.028431371Z" level=info msg="shim disconnected" id=3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471 Sep 13 00:52:23.029225 env[1189]: time="2025-09-13T00:52:23.028911221Z" level=warning msg="cleaning up after shim disconnected" id=3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471 namespace=k8s.io Sep 13 00:52:23.029225 env[1189]: time="2025-09-13T00:52:23.028939845Z" level=info msg="cleaning up dead shim" Sep 13 00:52:23.050915 env[1189]: time="2025-09-13T00:52:23.050771274Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:23Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3714 runtime=io.containerd.runc.v2\n" Sep 13 00:52:23.054535 env[1189]: time="2025-09-13T00:52:23.054454604Z" level=info msg="StopContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" returns successfully" Sep 13 00:52:23.055435 env[1189]: time="2025-09-13T00:52:23.055382422Z" level=info msg="StopPodSandbox for \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\"" Sep 13 00:52:23.055926 env[1189]: time="2025-09-13T00:52:23.055895968Z" level=info msg="Container to stop \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:23.056040 env[1189]: time="2025-09-13T00:52:23.056021333Z" level=info msg="Container to stop \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:23.056145 env[1189]: time="2025-09-13T00:52:23.056125895Z" level=info msg="Container to stop \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:23.056299 env[1189]: time="2025-09-13T00:52:23.056280444Z" level=info msg="Container to stop \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:23.056396 env[1189]: time="2025-09-13T00:52:23.056376920Z" level=info msg="Container to stop \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:23.063229 env[1189]: time="2025-09-13T00:52:23.063179031Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:23Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3718 runtime=io.containerd.runc.v2\n" Sep 13 00:52:23.063577 env[1189]: time="2025-09-13T00:52:23.063538301Z" level=info msg="TearDown network for sandbox \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\" successfully" Sep 13 00:52:23.065813 env[1189]: time="2025-09-13T00:52:23.063576481Z" level=info msg="StopPodSandbox for \"3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471\" returns successfully" Sep 13 00:52:23.089927 systemd[1]: cri-containerd-9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d.scope: Deactivated successfully. Sep 13 00:52:23.134992 env[1189]: time="2025-09-13T00:52:23.134918331Z" level=info msg="shim disconnected" id=9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d Sep 13 00:52:23.135498 env[1189]: time="2025-09-13T00:52:23.135407315Z" level=warning msg="cleaning up after shim disconnected" id=9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d namespace=k8s.io Sep 13 00:52:23.135704 env[1189]: time="2025-09-13T00:52:23.135681494Z" level=info msg="cleaning up dead shim" Sep 13 00:52:23.151379 env[1189]: time="2025-09-13T00:52:23.151305984Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:23Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3757 runtime=io.containerd.runc.v2\n" Sep 13 00:52:23.152092 env[1189]: time="2025-09-13T00:52:23.152043341Z" level=info msg="TearDown network for sandbox \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" successfully" Sep 13 00:52:23.152092 env[1189]: time="2025-09-13T00:52:23.152088623Z" level=info msg="StopPodSandbox for \"9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d\" returns successfully" Sep 13 00:52:23.175714 kubelet[1913]: I0913 00:52:23.175654 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksrjx\" (UniqueName: \"kubernetes.io/projected/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-kube-api-access-ksrjx\") pod \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\" (UID: \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\") " Sep 13 00:52:23.176301 kubelet[1913]: I0913 00:52:23.176263 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-cilium-config-path\") pod \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\" (UID: \"245b2d08-f123-4a56-8b4d-0d4b7b5373d9\") " Sep 13 00:52:23.198926 kubelet[1913]: I0913 00:52:23.195310 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "245b2d08-f123-4a56-8b4d-0d4b7b5373d9" (UID: "245b2d08-f123-4a56-8b4d-0d4b7b5373d9"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 13 00:52:23.199687 kubelet[1913]: I0913 00:52:23.199502 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-kube-api-access-ksrjx" (OuterVolumeSpecName: "kube-api-access-ksrjx") pod "245b2d08-f123-4a56-8b4d-0d4b7b5373d9" (UID: "245b2d08-f123-4a56-8b4d-0d4b7b5373d9"). InnerVolumeSpecName "kube-api-access-ksrjx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:52:23.277425 kubelet[1913]: I0913 00:52:23.277366 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.277728 kubelet[1913]: I0913 00:52:23.277246 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-etc-cni-netd\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278456 kubelet[1913]: I0913 00:52:23.277870 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-config-path\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278686 kubelet[1913]: I0913 00:52:23.278660 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-hostproc\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278760 kubelet[1913]: I0913 00:52:23.278697 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-net\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278760 kubelet[1913]: I0913 00:52:23.278730 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-hubble-tls\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278832 kubelet[1913]: I0913 00:52:23.278759 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-xtables-lock\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278832 kubelet[1913]: I0913 00:52:23.278796 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-cgroup\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278892 kubelet[1913]: I0913 00:52:23.278833 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/274059f7-91b4-475c-b55e-e7ba251bd6a0-clustermesh-secrets\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.278930 kubelet[1913]: I0913 00:52:23.278886 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cni-path\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279069 kubelet[1913]: I0913 00:52:23.278922 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-run\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279069 kubelet[1913]: I0913 00:52:23.278965 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-bpf-maps\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279069 kubelet[1913]: I0913 00:52:23.278998 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-lib-modules\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279069 kubelet[1913]: I0913 00:52:23.279027 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-kernel\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279069 kubelet[1913]: I0913 00:52:23.279059 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mk7n\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-kube-api-access-8mk7n\") pod \"274059f7-91b4-475c-b55e-e7ba251bd6a0\" (UID: \"274059f7-91b4-475c-b55e-e7ba251bd6a0\") " Sep 13 00:52:23.279289 kubelet[1913]: I0913 00:52:23.279197 1913 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ksrjx\" (UniqueName: \"kubernetes.io/projected/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-kube-api-access-ksrjx\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.279289 kubelet[1913]: I0913 00:52:23.279228 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/245b2d08-f123-4a56-8b4d-0d4b7b5373d9-cilium-config-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.279289 kubelet[1913]: I0913 00:52:23.279254 1913 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-etc-cni-netd\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.280696 kubelet[1913]: I0913 00:52:23.280643 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-hostproc" (OuterVolumeSpecName: "hostproc") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.280810 kubelet[1913]: I0913 00:52:23.280701 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.283002 kubelet[1913]: I0913 00:52:23.282928 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 13 00:52:23.284725 kubelet[1913]: I0913 00:52:23.283249 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cni-path" (OuterVolumeSpecName: "cni-path") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.284994 kubelet[1913]: I0913 00:52:23.283298 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.285129 kubelet[1913]: I0913 00:52:23.283317 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.285243 kubelet[1913]: I0913 00:52:23.284472 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.287017 kubelet[1913]: I0913 00:52:23.284514 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.287017 kubelet[1913]: I0913 00:52:23.284535 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.287017 kubelet[1913]: I0913 00:52:23.284554 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:23.287017 kubelet[1913]: I0913 00:52:23.285263 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-kube-api-access-8mk7n" (OuterVolumeSpecName: "kube-api-access-8mk7n") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "kube-api-access-8mk7n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:52:23.289528 kubelet[1913]: I0913 00:52:23.289457 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:52:23.292427 kubelet[1913]: I0913 00:52:23.292376 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274059f7-91b4-475c-b55e-e7ba251bd6a0-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "274059f7-91b4-475c-b55e-e7ba251bd6a0" (UID: "274059f7-91b4-475c-b55e-e7ba251bd6a0"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 13 00:52:23.380406 kubelet[1913]: I0913 00:52:23.380258 1913 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-lib-modules\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.380758 kubelet[1913]: I0913 00:52:23.380720 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-kernel\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.380886 kubelet[1913]: I0913 00:52:23.380865 1913 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8mk7n\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-kube-api-access-8mk7n\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381085 kubelet[1913]: I0913 00:52:23.381066 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-config-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381216 kubelet[1913]: I0913 00:52:23.381201 1913 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-hostproc\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381334 kubelet[1913]: I0913 00:52:23.381317 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-host-proc-sys-net\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381437 kubelet[1913]: I0913 00:52:23.381423 1913 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/274059f7-91b4-475c-b55e-e7ba251bd6a0-hubble-tls\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381536 kubelet[1913]: I0913 00:52:23.381522 1913 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-xtables-lock\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381659 kubelet[1913]: I0913 00:52:23.381637 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-cgroup\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381808 kubelet[1913]: I0913 00:52:23.381791 1913 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/274059f7-91b4-475c-b55e-e7ba251bd6a0-clustermesh-secrets\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.381925 kubelet[1913]: I0913 00:52:23.381907 1913 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cni-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.382024 kubelet[1913]: I0913 00:52:23.382011 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-cilium-run\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.382121 kubelet[1913]: I0913 00:52:23.382108 1913 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/274059f7-91b4-475c-b55e-e7ba251bd6a0-bpf-maps\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:23.606247 kubelet[1913]: I0913 00:52:23.606185 1913 scope.go:117] "RemoveContainer" containerID="3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5" Sep 13 00:52:23.607779 systemd[1]: Removed slice kubepods-besteffort-pod245b2d08_f123_4a56_8b4d_0d4b7b5373d9.slice. Sep 13 00:52:23.612564 env[1189]: time="2025-09-13T00:52:23.611221564Z" level=info msg="RemoveContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\"" Sep 13 00:52:23.619653 env[1189]: time="2025-09-13T00:52:23.619492216Z" level=info msg="RemoveContainer for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" returns successfully" Sep 13 00:52:23.621783 kubelet[1913]: I0913 00:52:23.620125 1913 scope.go:117] "RemoveContainer" containerID="3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5" Sep 13 00:52:23.632968 env[1189]: time="2025-09-13T00:52:23.632314333Z" level=error msg="ContainerStatus for \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\": not found" Sep 13 00:52:23.637162 kubelet[1913]: E0913 00:52:23.637071 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\": not found" containerID="3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5" Sep 13 00:52:23.637727 systemd[1]: Removed slice kubepods-burstable-pod274059f7_91b4_475c_b55e_e7ba251bd6a0.slice. Sep 13 00:52:23.637887 systemd[1]: kubepods-burstable-pod274059f7_91b4_475c_b55e_e7ba251bd6a0.slice: Consumed 11.619s CPU time. Sep 13 00:52:23.640685 kubelet[1913]: I0913 00:52:23.638377 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5"} err="failed to get container status \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\": rpc error: code = NotFound desc = an error occurred when try to find container \"3f1a9acb8675940434316e25eb60b851cb1a5b9bcd6ea5dd8e9773bca2447ec5\": not found" Sep 13 00:52:23.640967 kubelet[1913]: I0913 00:52:23.640927 1913 scope.go:117] "RemoveContainer" containerID="06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207" Sep 13 00:52:23.645166 env[1189]: time="2025-09-13T00:52:23.644737055Z" level=info msg="RemoveContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\"" Sep 13 00:52:23.653558 env[1189]: time="2025-09-13T00:52:23.653489766Z" level=info msg="RemoveContainer for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" returns successfully" Sep 13 00:52:23.654180 kubelet[1913]: I0913 00:52:23.654140 1913 scope.go:117] "RemoveContainer" containerID="bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73" Sep 13 00:52:23.656093 env[1189]: time="2025-09-13T00:52:23.656033849Z" level=info msg="RemoveContainer for \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\"" Sep 13 00:52:23.659523 env[1189]: time="2025-09-13T00:52:23.659462801Z" level=info msg="RemoveContainer for \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\" returns successfully" Sep 13 00:52:23.660920 kubelet[1913]: I0913 00:52:23.660879 1913 scope.go:117] "RemoveContainer" containerID="a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8" Sep 13 00:52:23.663161 env[1189]: time="2025-09-13T00:52:23.662805833Z" level=info msg="RemoveContainer for \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\"" Sep 13 00:52:23.666897 env[1189]: time="2025-09-13T00:52:23.666844056Z" level=info msg="RemoveContainer for \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\" returns successfully" Sep 13 00:52:23.668485 kubelet[1913]: I0913 00:52:23.668439 1913 scope.go:117] "RemoveContainer" containerID="c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0" Sep 13 00:52:23.676339 env[1189]: time="2025-09-13T00:52:23.675849429Z" level=info msg="RemoveContainer for \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\"" Sep 13 00:52:23.679409 env[1189]: time="2025-09-13T00:52:23.679349902Z" level=info msg="RemoveContainer for \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\" returns successfully" Sep 13 00:52:23.679949 kubelet[1913]: I0913 00:52:23.679905 1913 scope.go:117] "RemoveContainer" containerID="46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33" Sep 13 00:52:23.681806 env[1189]: time="2025-09-13T00:52:23.681742330Z" level=info msg="RemoveContainer for \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\"" Sep 13 00:52:23.689030 env[1189]: time="2025-09-13T00:52:23.688956824Z" level=info msg="RemoveContainer for \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\" returns successfully" Sep 13 00:52:23.689658 kubelet[1913]: I0913 00:52:23.689588 1913 scope.go:117] "RemoveContainer" containerID="06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207" Sep 13 00:52:23.690068 env[1189]: time="2025-09-13T00:52:23.689955260Z" level=error msg="ContainerStatus for \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\": not found" Sep 13 00:52:23.690456 kubelet[1913]: E0913 00:52:23.690426 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\": not found" containerID="06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207" Sep 13 00:52:23.690586 kubelet[1913]: I0913 00:52:23.690495 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207"} err="failed to get container status \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\": rpc error: code = NotFound desc = an error occurred when try to find container \"06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207\": not found" Sep 13 00:52:23.690709 kubelet[1913]: I0913 00:52:23.690592 1913 scope.go:117] "RemoveContainer" containerID="bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73" Sep 13 00:52:23.691126 env[1189]: time="2025-09-13T00:52:23.691057503Z" level=error msg="ContainerStatus for \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\": not found" Sep 13 00:52:23.691517 kubelet[1913]: E0913 00:52:23.691472 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\": not found" containerID="bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73" Sep 13 00:52:23.691578 kubelet[1913]: I0913 00:52:23.691522 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73"} err="failed to get container status \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\": rpc error: code = NotFound desc = an error occurred when try to find container \"bc4e5b5eaa3d23421a9436bea5880c6ef8700b837ec9cb02144dae2779660d73\": not found" Sep 13 00:52:23.691578 kubelet[1913]: I0913 00:52:23.691542 1913 scope.go:117] "RemoveContainer" containerID="a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8" Sep 13 00:52:23.691910 env[1189]: time="2025-09-13T00:52:23.691850579Z" level=error msg="ContainerStatus for \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\": not found" Sep 13 00:52:23.692284 kubelet[1913]: E0913 00:52:23.692248 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\": not found" containerID="a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8" Sep 13 00:52:23.692439 kubelet[1913]: I0913 00:52:23.692412 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8"} err="failed to get container status \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\": rpc error: code = NotFound desc = an error occurred when try to find container \"a7aa7614f1b51722ceb95e6945b5a3c1cfecb1001831eb3d3b07658b057db4b8\": not found" Sep 13 00:52:23.692525 kubelet[1913]: I0913 00:52:23.692510 1913 scope.go:117] "RemoveContainer" containerID="c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0" Sep 13 00:52:23.692987 env[1189]: time="2025-09-13T00:52:23.692912161Z" level=error msg="ContainerStatus for \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\": not found" Sep 13 00:52:23.693218 kubelet[1913]: E0913 00:52:23.693194 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\": not found" containerID="c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0" Sep 13 00:52:23.693309 kubelet[1913]: I0913 00:52:23.693238 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0"} err="failed to get container status \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\": rpc error: code = NotFound desc = an error occurred when try to find container \"c0e807a4c6bd207998310909025e1fa6f31a593c8c94e9c62941db83bbf405b0\": not found" Sep 13 00:52:23.693309 kubelet[1913]: I0913 00:52:23.693259 1913 scope.go:117] "RemoveContainer" containerID="46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33" Sep 13 00:52:23.693707 env[1189]: time="2025-09-13T00:52:23.693593874Z" level=error msg="ContainerStatus for \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\": not found" Sep 13 00:52:23.693985 kubelet[1913]: E0913 00:52:23.693943 1913 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\": not found" containerID="46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33" Sep 13 00:52:23.694071 kubelet[1913]: I0913 00:52:23.694008 1913 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33"} err="failed to get container status \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\": rpc error: code = NotFound desc = an error occurred when try to find container \"46f8bf4d3fcc721da15518a9bd8bf9988874601548d3d6b0ed0fafe03002ef33\": not found" Sep 13 00:52:23.792031 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-06626e5c6223565df9965e945f5008c50a92764edde4d089239b66e2db003207-rootfs.mount: Deactivated successfully. Sep 13 00:52:23.792203 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d-rootfs.mount: Deactivated successfully. Sep 13 00:52:23.792269 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9ffc437dfb983dee89640614603233dc435b73926acb1da6bff134a661964e8d-shm.mount: Deactivated successfully. Sep 13 00:52:23.792339 systemd[1]: var-lib-kubelet-pods-274059f7\x2d91b4\x2d475c\x2db55e\x2de7ba251bd6a0-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:52:23.792407 systemd[1]: var-lib-kubelet-pods-274059f7\x2d91b4\x2d475c\x2db55e\x2de7ba251bd6a0-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8mk7n.mount: Deactivated successfully. Sep 13 00:52:23.792469 systemd[1]: var-lib-kubelet-pods-274059f7\x2d91b4\x2d475c\x2db55e\x2de7ba251bd6a0-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:52:23.792528 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d68fdc45b11e60daa97ca165f8e55f0e88343604474c798059db3154b889471-rootfs.mount: Deactivated successfully. Sep 13 00:52:23.792586 systemd[1]: var-lib-kubelet-pods-245b2d08\x2df123\x2d4a56\x2d8b4d\x2d0d4b7b5373d9-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dksrjx.mount: Deactivated successfully. Sep 13 00:52:24.059048 kubelet[1913]: E0913 00:52:24.058999 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:24.695209 sshd[3613]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:24.699442 systemd-logind[1182]: Session 24 logged out. Waiting for processes to exit. Sep 13 00:52:24.701230 systemd[1]: Started sshd@25-24.144.95.58:22-147.75.109.163:38954.service. Sep 13 00:52:24.702239 systemd[1]: sshd@24-24.144.95.58:22-147.75.109.163:38942.service: Deactivated successfully. Sep 13 00:52:24.703741 systemd[1]: session-24.scope: Deactivated successfully. Sep 13 00:52:24.704280 systemd[1]: session-24.scope: Consumed 1.364s CPU time. Sep 13 00:52:24.706038 systemd-logind[1182]: Removed session 24. Sep 13 00:52:24.769138 sshd[3776]: Accepted publickey for core from 147.75.109.163 port 38954 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:24.771795 sshd[3776]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:24.778348 systemd[1]: Started session-25.scope. Sep 13 00:52:24.779215 systemd-logind[1182]: New session 25 of user core. Sep 13 00:52:25.064197 kubelet[1913]: I0913 00:52:25.064128 1913 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245b2d08-f123-4a56-8b4d-0d4b7b5373d9" path="/var/lib/kubelet/pods/245b2d08-f123-4a56-8b4d-0d4b7b5373d9/volumes" Sep 13 00:52:25.065399 kubelet[1913]: I0913 00:52:25.065359 1913 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274059f7-91b4-475c-b55e-e7ba251bd6a0" path="/var/lib/kubelet/pods/274059f7-91b4-475c-b55e-e7ba251bd6a0/volumes" Sep 13 00:52:25.537967 sshd[3776]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:25.544390 systemd[1]: sshd@25-24.144.95.58:22-147.75.109.163:38954.service: Deactivated successfully. Sep 13 00:52:25.546115 systemd[1]: session-25.scope: Deactivated successfully. Sep 13 00:52:25.547782 systemd-logind[1182]: Session 25 logged out. Waiting for processes to exit. Sep 13 00:52:25.553584 systemd[1]: Started sshd@26-24.144.95.58:22-147.75.109.163:38962.service. Sep 13 00:52:25.558863 systemd-logind[1182]: Removed session 25. Sep 13 00:52:25.609627 systemd[1]: Created slice kubepods-burstable-pod9124305d_947c_4de4_8e7e_b33cd33d7ee3.slice. Sep 13 00:52:25.633977 sshd[3788]: Accepted publickey for core from 147.75.109.163 port 38962 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:25.636934 sshd[3788]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:25.649221 systemd-logind[1182]: New session 26 of user core. Sep 13 00:52:25.650109 systemd[1]: Started session-26.scope. Sep 13 00:52:25.711621 kubelet[1913]: I0913 00:52:25.711529 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-config-path\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711621 kubelet[1913]: I0913 00:52:25.711617 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hostproc\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711655 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-cgroup\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711680 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-net\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711711 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-etc-cni-netd\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711741 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-lib-modules\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711767 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-clustermesh-secrets\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.711830 kubelet[1913]: I0913 00:52:25.711799 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-kernel\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711826 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2psfj\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-kube-api-access-2psfj\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711856 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-bpf-maps\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711881 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cni-path\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711910 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-xtables-lock\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711935 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hubble-tls\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712060 kubelet[1913]: I0913 00:52:25.711965 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-run\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.712273 kubelet[1913]: I0913 00:52:25.711994 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-ipsec-secrets\") pod \"cilium-b9gpm\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " pod="kube-system/cilium-b9gpm" Sep 13 00:52:25.906110 sshd[3788]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:25.912434 systemd[1]: sshd@26-24.144.95.58:22-147.75.109.163:38962.service: Deactivated successfully. Sep 13 00:52:25.913699 systemd[1]: session-26.scope: Deactivated successfully. Sep 13 00:52:25.915266 systemd-logind[1182]: Session 26 logged out. Waiting for processes to exit. Sep 13 00:52:25.916538 kubelet[1913]: E0913 00:52:25.916494 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:25.921420 env[1189]: time="2025-09-13T00:52:25.921154485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b9gpm,Uid:9124305d-947c-4de4-8e7e-b33cd33d7ee3,Namespace:kube-system,Attempt:0,}" Sep 13 00:52:25.924483 systemd[1]: Started sshd@27-24.144.95.58:22-147.75.109.163:38964.service. Sep 13 00:52:25.930760 systemd-logind[1182]: Removed session 26. Sep 13 00:52:25.973475 env[1189]: time="2025-09-13T00:52:25.973334031Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:52:25.973475 env[1189]: time="2025-09-13T00:52:25.973418293Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:52:25.973757 env[1189]: time="2025-09-13T00:52:25.973438501Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:52:25.975136 env[1189]: time="2025-09-13T00:52:25.973884631Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001 pid=3813 runtime=io.containerd.runc.v2 Sep 13 00:52:26.009316 systemd[1]: Started cri-containerd-efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001.scope. Sep 13 00:52:26.012482 sshd[3805]: Accepted publickey for core from 147.75.109.163 port 38964 ssh2: RSA SHA256:Z+gHFjZa6FbNcZ3OMDgtPyMdExX9gV+gkyGg/y2DokA Sep 13 00:52:26.015081 sshd[3805]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:52:26.031685 systemd[1]: Started session-27.scope. Sep 13 00:52:26.032532 systemd-logind[1182]: New session 27 of user core. Sep 13 00:52:26.092959 env[1189]: time="2025-09-13T00:52:26.092905261Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b9gpm,Uid:9124305d-947c-4de4-8e7e-b33cd33d7ee3,Namespace:kube-system,Attempt:0,} returns sandbox id \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\"" Sep 13 00:52:26.094208 kubelet[1913]: E0913 00:52:26.094160 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:26.106197 env[1189]: time="2025-09-13T00:52:26.106120795Z" level=info msg="CreateContainer within sandbox \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:52:26.144000 env[1189]: time="2025-09-13T00:52:26.143868976Z" level=info msg="CreateContainer within sandbox \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\"" Sep 13 00:52:26.148395 env[1189]: time="2025-09-13T00:52:26.148273767Z" level=info msg="StartContainer for \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\"" Sep 13 00:52:26.179891 systemd[1]: Started cri-containerd-11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef.scope. Sep 13 00:52:26.212386 systemd[1]: cri-containerd-11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef.scope: Deactivated successfully. Sep 13 00:52:26.239879 env[1189]: time="2025-09-13T00:52:26.237852584Z" level=info msg="shim disconnected" id=11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef Sep 13 00:52:26.239879 env[1189]: time="2025-09-13T00:52:26.237930081Z" level=warning msg="cleaning up after shim disconnected" id=11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef namespace=k8s.io Sep 13 00:52:26.239879 env[1189]: time="2025-09-13T00:52:26.237945932Z" level=info msg="cleaning up dead shim" Sep 13 00:52:26.258957 env[1189]: time="2025-09-13T00:52:26.258810985Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:26Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3883 runtime=io.containerd.runc.v2\ntime=\"2025-09-13T00:52:26Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Sep 13 00:52:26.259729 env[1189]: time="2025-09-13T00:52:26.259566528Z" level=error msg="copy shim log" error="read /proc/self/fd/40: file already closed" Sep 13 00:52:26.260132 env[1189]: time="2025-09-13T00:52:26.260091932Z" level=error msg="Failed to pipe stdout of container \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\"" error="reading from a closed fifo" Sep 13 00:52:26.266651 env[1189]: time="2025-09-13T00:52:26.263841245Z" level=error msg="Failed to pipe stderr of container \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\"" error="reading from a closed fifo" Sep 13 00:52:26.275584 env[1189]: time="2025-09-13T00:52:26.274373175Z" level=error msg="StartContainer for \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Sep 13 00:52:26.276998 kubelet[1913]: E0913 00:52:26.276743 1913 log.go:32] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef" Sep 13 00:52:26.288119 kubelet[1913]: E0913 00:52:26.288045 1913 kuberuntime_manager.go:1358] "Unhandled Error" err=< Sep 13 00:52:26.288119 kubelet[1913]: init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Sep 13 00:52:26.288119 kubelet[1913]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Sep 13 00:52:26.288119 kubelet[1913]: rm /hostbin/cilium-mount Sep 13 00:52:26.288464 kubelet[1913]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2psfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:&AppArmorProfile{Type:Unconfined,LocalhostProfile:nil,},},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-b9gpm_kube-system(9124305d-947c-4de4-8e7e-b33cd33d7ee3): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Sep 13 00:52:26.288464 kubelet[1913]: > logger="UnhandledError" Sep 13 00:52:26.291207 kubelet[1913]: E0913 00:52:26.291105 1913 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-b9gpm" podUID="9124305d-947c-4de4-8e7e-b33cd33d7ee3" Sep 13 00:52:26.629568 env[1189]: time="2025-09-13T00:52:26.629501775Z" level=info msg="StopPodSandbox for \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\"" Sep 13 00:52:26.629841 env[1189]: time="2025-09-13T00:52:26.629593763Z" level=info msg="Container to stop \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:52:26.643475 systemd[1]: cri-containerd-efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001.scope: Deactivated successfully. Sep 13 00:52:26.688812 env[1189]: time="2025-09-13T00:52:26.688752372Z" level=info msg="shim disconnected" id=efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001 Sep 13 00:52:26.689097 env[1189]: time="2025-09-13T00:52:26.689073176Z" level=warning msg="cleaning up after shim disconnected" id=efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001 namespace=k8s.io Sep 13 00:52:26.689172 env[1189]: time="2025-09-13T00:52:26.689155574Z" level=info msg="cleaning up dead shim" Sep 13 00:52:26.700584 env[1189]: time="2025-09-13T00:52:26.700518724Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:26Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3916 runtime=io.containerd.runc.v2\n" Sep 13 00:52:26.701270 env[1189]: time="2025-09-13T00:52:26.701229021Z" level=info msg="TearDown network for sandbox \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\" successfully" Sep 13 00:52:26.701671 env[1189]: time="2025-09-13T00:52:26.701633239Z" level=info msg="StopPodSandbox for \"efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001\" returns successfully" Sep 13 00:52:26.822993 kubelet[1913]: I0913 00:52:26.822924 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-ipsec-secrets\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.822993 kubelet[1913]: I0913 00:52:26.822981 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-net\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.822993 kubelet[1913]: I0913 00:52:26.823008 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-kernel\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823032 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-xtables-lock\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823058 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-lib-modules\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823079 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cni-path\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823104 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-bpf-maps\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823134 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-config-path\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823157 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-etc-cni-netd\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823182 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-run\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823208 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hubble-tls\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823266 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-clustermesh-secrets\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823282 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2psfj\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-kube-api-access-2psfj\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823298 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hostproc\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823375 kubelet[1913]: I0913 00:52:26.823312 1913 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-cgroup\") pod \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\" (UID: \"9124305d-947c-4de4-8e7e-b33cd33d7ee3\") " Sep 13 00:52:26.823769 kubelet[1913]: I0913 00:52:26.823414 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824249 kubelet[1913]: I0913 00:52:26.824210 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824359 kubelet[1913]: I0913 00:52:26.824270 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824546 kubelet[1913]: I0913 00:52:26.824495 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824722 kubelet[1913]: I0913 00:52:26.824697 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824839 kubelet[1913]: I0913 00:52:26.824822 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.824935 kubelet[1913]: I0913 00:52:26.824915 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.825066 kubelet[1913]: I0913 00:52:26.825043 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cni-path" (OuterVolumeSpecName: "cni-path") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.825183 kubelet[1913]: I0913 00:52:26.825161 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.827436 kubelet[1913]: I0913 00:52:26.827387 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hostproc" (OuterVolumeSpecName: "hostproc") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Sep 13 00:52:26.829799 kubelet[1913]: I0913 00:52:26.829762 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 13 00:52:26.831545 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-efa03ca43670418715b4c663f99382484517ec59e935482699ed0bc6dee69001-shm.mount: Deactivated successfully. Sep 13 00:52:26.831733 systemd[1]: var-lib-kubelet-pods-9124305d\x2d947c\x2d4de4\x2d8e7e\x2db33cd33d7ee3-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Sep 13 00:52:26.834456 kubelet[1913]: I0913 00:52:26.834394 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 13 00:52:26.837762 systemd[1]: var-lib-kubelet-pods-9124305d\x2d947c\x2d4de4\x2d8e7e\x2db33cd33d7ee3-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:52:26.842481 systemd[1]: var-lib-kubelet-pods-9124305d\x2d947c\x2d4de4\x2d8e7e\x2db33cd33d7ee3-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2psfj.mount: Deactivated successfully. Sep 13 00:52:26.843774 kubelet[1913]: I0913 00:52:26.843723 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-kube-api-access-2psfj" (OuterVolumeSpecName: "kube-api-access-2psfj") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "kube-api-access-2psfj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:52:26.845026 kubelet[1913]: I0913 00:52:26.844979 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 13 00:52:26.849477 systemd[1]: var-lib-kubelet-pods-9124305d\x2d947c\x2d4de4\x2d8e7e\x2db33cd33d7ee3-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:52:26.851474 kubelet[1913]: I0913 00:52:26.851421 1913 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "9124305d-947c-4de4-8e7e-b33cd33d7ee3" (UID: "9124305d-947c-4de4-8e7e-b33cd33d7ee3"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.924089 1913 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2psfj\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-kube-api-access-2psfj\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925801 1913 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hostproc\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925841 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-cgroup\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925854 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-ipsec-secrets\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925865 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-net\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925884 1913 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-host-proc-sys-kernel\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925896 1913 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-xtables-lock\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925906 1913 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-lib-modules\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925917 1913 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cni-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925934 1913 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-bpf-maps\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925951 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-config-path\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925961 1913 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-etc-cni-netd\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925971 1913 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9124305d-947c-4de4-8e7e-b33cd33d7ee3-cilium-run\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925980 1913 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9124305d-947c-4de4-8e7e-b33cd33d7ee3-hubble-tls\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:26.926287 kubelet[1913]: I0913 00:52:26.925989 1913 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9124305d-947c-4de4-8e7e-b33cd33d7ee3-clustermesh-secrets\") on node \"ci-3510.3.8-n-1fe98121a2\" DevicePath \"\"" Sep 13 00:52:27.067271 systemd[1]: Removed slice kubepods-burstable-pod9124305d_947c_4de4_8e7e_b33cd33d7ee3.slice. Sep 13 00:52:27.245105 kubelet[1913]: E0913 00:52:27.245034 1913 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 00:52:27.633303 kubelet[1913]: I0913 00:52:27.633183 1913 scope.go:117] "RemoveContainer" containerID="11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef" Sep 13 00:52:27.639542 env[1189]: time="2025-09-13T00:52:27.639468792Z" level=info msg="RemoveContainer for \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\"" Sep 13 00:52:27.644876 env[1189]: time="2025-09-13T00:52:27.644713653Z" level=info msg="RemoveContainer for \"11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef\" returns successfully" Sep 13 00:52:27.710383 systemd[1]: Created slice kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice. Sep 13 00:52:27.831060 kubelet[1913]: I0913 00:52:27.830990 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/e361e0ee-62a6-4868-ad51-0fe54062ff4b-cilium-ipsec-secrets\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.831399 kubelet[1913]: I0913 00:52:27.831346 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-lib-modules\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.831675 kubelet[1913]: I0913 00:52:27.831647 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl9cb\" (UniqueName: \"kubernetes.io/projected/e361e0ee-62a6-4868-ad51-0fe54062ff4b-kube-api-access-fl9cb\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832032 kubelet[1913]: I0913 00:52:27.831985 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-bpf-maps\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832235 kubelet[1913]: I0913 00:52:27.832211 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-cni-path\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832415 kubelet[1913]: I0913 00:52:27.832375 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e361e0ee-62a6-4868-ad51-0fe54062ff4b-clustermesh-secrets\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832576 kubelet[1913]: I0913 00:52:27.832542 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-etc-cni-netd\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832790 kubelet[1913]: I0913 00:52:27.832750 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-cilium-cgroup\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.832953 kubelet[1913]: I0913 00:52:27.832917 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e361e0ee-62a6-4868-ad51-0fe54062ff4b-cilium-config-path\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833129 kubelet[1913]: I0913 00:52:27.833095 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-hostproc\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833269 kubelet[1913]: I0913 00:52:27.833248 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-host-proc-sys-net\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833419 kubelet[1913]: I0913 00:52:27.833397 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e361e0ee-62a6-4868-ad51-0fe54062ff4b-hubble-tls\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833579 kubelet[1913]: I0913 00:52:27.833557 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-host-proc-sys-kernel\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833846 kubelet[1913]: I0913 00:52:27.833809 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-cilium-run\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:27.833998 kubelet[1913]: I0913 00:52:27.833978 1913 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e361e0ee-62a6-4868-ad51-0fe54062ff4b-xtables-lock\") pod \"cilium-lq9jl\" (UID: \"e361e0ee-62a6-4868-ad51-0fe54062ff4b\") " pod="kube-system/cilium-lq9jl" Sep 13 00:52:28.021942 kubelet[1913]: E0913 00:52:28.021896 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:28.023039 env[1189]: time="2025-09-13T00:52:28.022970926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lq9jl,Uid:e361e0ee-62a6-4868-ad51-0fe54062ff4b,Namespace:kube-system,Attempt:0,}" Sep 13 00:52:28.043270 env[1189]: time="2025-09-13T00:52:28.043140101Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:52:28.043270 env[1189]: time="2025-09-13T00:52:28.043190493Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:52:28.043719 env[1189]: time="2025-09-13T00:52:28.043203852Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:52:28.043719 env[1189]: time="2025-09-13T00:52:28.043331695Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769 pid=3946 runtime=io.containerd.runc.v2 Sep 13 00:52:28.061363 systemd[1]: Started cri-containerd-3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769.scope. Sep 13 00:52:28.110310 env[1189]: time="2025-09-13T00:52:28.110230585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lq9jl,Uid:e361e0ee-62a6-4868-ad51-0fe54062ff4b,Namespace:kube-system,Attempt:0,} returns sandbox id \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\"" Sep 13 00:52:28.112021 kubelet[1913]: E0913 00:52:28.111962 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:28.124331 env[1189]: time="2025-09-13T00:52:28.123230194Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:52:28.141028 env[1189]: time="2025-09-13T00:52:28.140955411Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261\"" Sep 13 00:52:28.144562 env[1189]: time="2025-09-13T00:52:28.144069768Z" level=info msg="StartContainer for \"64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261\"" Sep 13 00:52:28.171668 systemd[1]: Started cri-containerd-64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261.scope. Sep 13 00:52:28.229987 env[1189]: time="2025-09-13T00:52:28.229914359Z" level=info msg="StartContainer for \"64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261\" returns successfully" Sep 13 00:52:28.247594 systemd[1]: cri-containerd-64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261.scope: Deactivated successfully. Sep 13 00:52:28.290958 env[1189]: time="2025-09-13T00:52:28.290780658Z" level=info msg="shim disconnected" id=64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261 Sep 13 00:52:28.291270 env[1189]: time="2025-09-13T00:52:28.291233857Z" level=warning msg="cleaning up after shim disconnected" id=64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261 namespace=k8s.io Sep 13 00:52:28.291371 env[1189]: time="2025-09-13T00:52:28.291353161Z" level=info msg="cleaning up dead shim" Sep 13 00:52:28.305442 env[1189]: time="2025-09-13T00:52:28.305373783Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4029 runtime=io.containerd.runc.v2\n" Sep 13 00:52:28.640163 kubelet[1913]: E0913 00:52:28.639977 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:28.648739 env[1189]: time="2025-09-13T00:52:28.648671001Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:52:28.666448 env[1189]: time="2025-09-13T00:52:28.666345623Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98\"" Sep 13 00:52:28.667827 env[1189]: time="2025-09-13T00:52:28.667752854Z" level=info msg="StartContainer for \"5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98\"" Sep 13 00:52:28.702191 systemd[1]: Started cri-containerd-5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98.scope. Sep 13 00:52:28.756167 env[1189]: time="2025-09-13T00:52:28.756092362Z" level=info msg="StartContainer for \"5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98\" returns successfully" Sep 13 00:52:28.764420 systemd[1]: cri-containerd-5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98.scope: Deactivated successfully. Sep 13 00:52:28.800189 env[1189]: time="2025-09-13T00:52:28.800114745Z" level=info msg="shim disconnected" id=5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98 Sep 13 00:52:28.800189 env[1189]: time="2025-09-13T00:52:28.800185553Z" level=warning msg="cleaning up after shim disconnected" id=5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98 namespace=k8s.io Sep 13 00:52:28.800189 env[1189]: time="2025-09-13T00:52:28.800203478Z" level=info msg="cleaning up dead shim" Sep 13 00:52:28.816968 env[1189]: time="2025-09-13T00:52:28.816456878Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4092 runtime=io.containerd.runc.v2\n" Sep 13 00:52:29.062326 kubelet[1913]: I0913 00:52:29.062270 1913 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9124305d-947c-4de4-8e7e-b33cd33d7ee3" path="/var/lib/kubelet/pods/9124305d-947c-4de4-8e7e-b33cd33d7ee3/volumes" Sep 13 00:52:29.347660 kubelet[1913]: W0913 00:52:29.347397 1913 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9124305d_947c_4de4_8e7e_b33cd33d7ee3.slice/cri-containerd-11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef.scope WatchSource:0}: container "11b1d4831a93ae386288ed528ba3f3f9c659090f3c379cd0f8b56d4796ef63ef" in namespace "k8s.io": not found Sep 13 00:52:29.655456 kubelet[1913]: E0913 00:52:29.655310 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:29.672471 env[1189]: time="2025-09-13T00:52:29.671686613Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:52:29.700363 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3662969404.mount: Deactivated successfully. Sep 13 00:52:29.714458 env[1189]: time="2025-09-13T00:52:29.714336427Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944\"" Sep 13 00:52:29.715950 env[1189]: time="2025-09-13T00:52:29.715890037Z" level=info msg="StartContainer for \"b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944\"" Sep 13 00:52:29.759349 systemd[1]: Started cri-containerd-b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944.scope. Sep 13 00:52:29.813454 env[1189]: time="2025-09-13T00:52:29.813382919Z" level=info msg="StartContainer for \"b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944\" returns successfully" Sep 13 00:52:29.819368 systemd[1]: cri-containerd-b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944.scope: Deactivated successfully. Sep 13 00:52:29.853289 env[1189]: time="2025-09-13T00:52:29.853201497Z" level=info msg="shim disconnected" id=b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944 Sep 13 00:52:29.853289 env[1189]: time="2025-09-13T00:52:29.853290354Z" level=warning msg="cleaning up after shim disconnected" id=b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944 namespace=k8s.io Sep 13 00:52:29.853577 env[1189]: time="2025-09-13T00:52:29.853309075Z" level=info msg="cleaning up dead shim" Sep 13 00:52:29.866956 env[1189]: time="2025-09-13T00:52:29.866881755Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:29Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4149 runtime=io.containerd.runc.v2\n" Sep 13 00:52:29.943151 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944-rootfs.mount: Deactivated successfully. Sep 13 00:52:30.034023 kubelet[1913]: I0913 00:52:30.033932 1913 setters.go:618] "Node became not ready" node="ci-3510.3.8-n-1fe98121a2" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-13T00:52:30Z","lastTransitionTime":"2025-09-13T00:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 13 00:52:30.660924 kubelet[1913]: E0913 00:52:30.660848 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:30.669268 env[1189]: time="2025-09-13T00:52:30.669197706Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:52:30.692011 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2949059832.mount: Deactivated successfully. Sep 13 00:52:30.697926 env[1189]: time="2025-09-13T00:52:30.695513230Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe\"" Sep 13 00:52:30.698886 env[1189]: time="2025-09-13T00:52:30.698837967Z" level=info msg="StartContainer for \"0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe\"" Sep 13 00:52:30.702204 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2944286661.mount: Deactivated successfully. Sep 13 00:52:30.740455 systemd[1]: Started cri-containerd-0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe.scope. Sep 13 00:52:30.789212 systemd[1]: cri-containerd-0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe.scope: Deactivated successfully. Sep 13 00:52:30.792365 env[1189]: time="2025-09-13T00:52:30.791668354Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice/cri-containerd-0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe.scope/memory.events\": no such file or directory" Sep 13 00:52:30.796007 env[1189]: time="2025-09-13T00:52:30.795920480Z" level=info msg="StartContainer for \"0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe\" returns successfully" Sep 13 00:52:30.830160 env[1189]: time="2025-09-13T00:52:30.830029950Z" level=info msg="shim disconnected" id=0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe Sep 13 00:52:30.830815 env[1189]: time="2025-09-13T00:52:30.830754323Z" level=warning msg="cleaning up after shim disconnected" id=0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe namespace=k8s.io Sep 13 00:52:30.830815 env[1189]: time="2025-09-13T00:52:30.830804748Z" level=info msg="cleaning up dead shim" Sep 13 00:52:30.843472 env[1189]: time="2025-09-13T00:52:30.843405515Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:52:30Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4205 runtime=io.containerd.runc.v2\n" Sep 13 00:52:30.943172 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe-rootfs.mount: Deactivated successfully. Sep 13 00:52:31.688176 kubelet[1913]: E0913 00:52:31.688124 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:31.695366 env[1189]: time="2025-09-13T00:52:31.695088462Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:52:31.722189 env[1189]: time="2025-09-13T00:52:31.722124284Z" level=info msg="CreateContainer within sandbox \"3b4c08260182fdb66567f3576871d5919248413ae9c6c98d33ebeb87776d4769\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1\"" Sep 13 00:52:31.723879 env[1189]: time="2025-09-13T00:52:31.723824999Z" level=info msg="StartContainer for \"cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1\"" Sep 13 00:52:31.753131 systemd[1]: Started cri-containerd-cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1.scope. Sep 13 00:52:31.801691 env[1189]: time="2025-09-13T00:52:31.801579515Z" level=info msg="StartContainer for \"cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1\" returns successfully" Sep 13 00:52:32.470542 kubelet[1913]: W0913 00:52:32.470446 1913 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice/cri-containerd-64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261.scope WatchSource:0}: task 64dc39d0f7dcfe79e1b12e39b1ccd3329308c2bde63f189a6d70fac30f535261 not found Sep 13 00:52:32.497650 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 13 00:52:32.694332 kubelet[1913]: E0913 00:52:32.694285 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:32.725912 kubelet[1913]: I0913 00:52:32.725771 1913 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-lq9jl" podStartSLOduration=5.72574966 podStartE2EDuration="5.72574966s" podCreationTimestamp="2025-09-13 00:52:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:52:32.723949171 +0000 UTC m=+145.954482511" watchObservedRunningTime="2025-09-13 00:52:32.72574966 +0000 UTC m=+145.956283000" Sep 13 00:52:34.023290 kubelet[1913]: E0913 00:52:34.023223 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:34.557028 systemd[1]: run-containerd-runc-k8s.io-cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1-runc.WEP4PD.mount: Deactivated successfully. Sep 13 00:52:35.582181 kubelet[1913]: W0913 00:52:35.582103 1913 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice/cri-containerd-5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98.scope WatchSource:0}: task 5416508dc5c775a0abd029075e5c6e7f70a2841d58d1db732b561daa4048bb98 not found Sep 13 00:52:35.877188 systemd-networkd[1014]: lxc_health: Link UP Sep 13 00:52:35.885279 systemd-networkd[1014]: lxc_health: Gained carrier Sep 13 00:52:35.886077 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:52:36.026359 kubelet[1913]: E0913 00:52:36.026291 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:36.705416 kubelet[1913]: E0913 00:52:36.705376 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:36.803914 systemd[1]: run-containerd-runc-k8s.io-cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1-runc.sBgFr6.mount: Deactivated successfully. Sep 13 00:52:36.943839 systemd-networkd[1014]: lxc_health: Gained IPv6LL Sep 13 00:52:37.707185 kubelet[1913]: E0913 00:52:37.707135 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Sep 13 00:52:38.692022 kubelet[1913]: W0913 00:52:38.691919 1913 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice/cri-containerd-b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944.scope WatchSource:0}: task b3959e450bf55f0acb661d289c8305fbd5245a15fe0326a3b74ba1f123318944 not found Sep 13 00:52:39.046850 systemd[1]: run-containerd-runc-k8s.io-cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1-runc.CR5CzH.mount: Deactivated successfully. Sep 13 00:52:41.234856 systemd[1]: run-containerd-runc-k8s.io-cd3660122606b298dd752079d0bc7e72c27dcab0888419b4a4c91bab767e5aa1-runc.lKvbS0.mount: Deactivated successfully. Sep 13 00:52:41.378163 sshd[3805]: pam_unix(sshd:session): session closed for user core Sep 13 00:52:41.383355 systemd-logind[1182]: Session 27 logged out. Waiting for processes to exit. Sep 13 00:52:41.384652 systemd[1]: sshd@27-24.144.95.58:22-147.75.109.163:38964.service: Deactivated successfully. Sep 13 00:52:41.385423 systemd[1]: session-27.scope: Deactivated successfully. Sep 13 00:52:41.387230 systemd-logind[1182]: Removed session 27. Sep 13 00:52:41.800990 kubelet[1913]: W0913 00:52:41.800907 1913 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode361e0ee_62a6_4868_ad51_0fe54062ff4b.slice/cri-containerd-0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe.scope WatchSource:0}: task 0c1298fcc42a587e62698a7d40b9d2f1743d5a2b21101ab60f4a189b2eb499fe not found Sep 13 00:52:42.059829 kubelet[1913]: E0913 00:52:42.059660 1913 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2"