Aug 13 01:04:19.059072 kernel: Linux version 5.15.189-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Tue Aug 12 23:01:50 -00 2025 Aug 13 01:04:19.059099 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8f8aacd9fbcdd713563d390e899e90bedf5577e4b1b261b4e57687d87edd6b57 Aug 13 01:04:19.059107 kernel: BIOS-provided physical RAM map: Aug 13 01:04:19.059113 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Aug 13 01:04:19.059118 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Aug 13 01:04:19.059124 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Aug 13 01:04:19.059130 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Aug 13 01:04:19.059136 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Aug 13 01:04:19.059146 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Aug 13 01:04:19.059151 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Aug 13 01:04:19.059157 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Aug 13 01:04:19.059163 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Aug 13 01:04:19.059168 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Aug 13 01:04:19.059174 kernel: NX (Execute Disable) protection: active Aug 13 01:04:19.059184 kernel: SMBIOS 2.8 present. Aug 13 01:04:19.059191 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Aug 13 01:04:19.059197 kernel: Hypervisor detected: KVM Aug 13 01:04:19.059203 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Aug 13 01:04:19.059211 kernel: kvm-clock: cpu 0, msr 5a19e001, primary cpu clock Aug 13 01:04:19.059217 kernel: kvm-clock: using sched offset of 3931890174 cycles Aug 13 01:04:19.059224 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Aug 13 01:04:19.059230 kernel: tsc: Detected 2794.750 MHz processor Aug 13 01:04:19.059249 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 01:04:19.059260 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 01:04:19.059266 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Aug 13 01:04:19.059273 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 01:04:19.059279 kernel: Using GB pages for direct mapping Aug 13 01:04:19.059285 kernel: ACPI: Early table checksum verification disabled Aug 13 01:04:19.059291 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Aug 13 01:04:19.059297 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059303 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059310 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059320 kernel: ACPI: FACS 0x000000009CFE0000 000040 Aug 13 01:04:19.059326 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059332 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059338 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059344 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Aug 13 01:04:19.059350 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Aug 13 01:04:19.059356 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Aug 13 01:04:19.059363 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Aug 13 01:04:19.059378 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Aug 13 01:04:19.059385 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Aug 13 01:04:19.059391 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Aug 13 01:04:19.059398 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Aug 13 01:04:19.059404 kernel: No NUMA configuration found Aug 13 01:04:19.059411 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Aug 13 01:04:19.059421 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Aug 13 01:04:19.059428 kernel: Zone ranges: Aug 13 01:04:19.059434 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 01:04:19.059441 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Aug 13 01:04:19.059447 kernel: Normal empty Aug 13 01:04:19.059454 kernel: Movable zone start for each node Aug 13 01:04:19.059460 kernel: Early memory node ranges Aug 13 01:04:19.059467 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Aug 13 01:04:19.059473 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Aug 13 01:04:19.059480 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Aug 13 01:04:19.059492 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 01:04:19.059499 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Aug 13 01:04:19.059506 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Aug 13 01:04:19.059512 kernel: ACPI: PM-Timer IO Port: 0x608 Aug 13 01:04:19.059519 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Aug 13 01:04:19.059525 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Aug 13 01:04:19.059532 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Aug 13 01:04:19.059538 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Aug 13 01:04:19.059545 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 01:04:19.059557 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Aug 13 01:04:19.059564 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Aug 13 01:04:19.059570 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 01:04:19.059577 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Aug 13 01:04:19.059584 kernel: TSC deadline timer available Aug 13 01:04:19.059590 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Aug 13 01:04:19.059597 kernel: kvm-guest: KVM setup pv remote TLB flush Aug 13 01:04:19.059603 kernel: kvm-guest: setup PV sched yield Aug 13 01:04:19.059610 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Aug 13 01:04:19.059620 kernel: Booting paravirtualized kernel on KVM Aug 13 01:04:19.059627 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 01:04:19.059633 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:4 nr_node_ids:1 Aug 13 01:04:19.059640 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u524288 Aug 13 01:04:19.059646 kernel: pcpu-alloc: s188696 r8192 d32488 u524288 alloc=1*2097152 Aug 13 01:04:19.059653 kernel: pcpu-alloc: [0] 0 1 2 3 Aug 13 01:04:19.059659 kernel: kvm-guest: setup async PF for cpu 0 Aug 13 01:04:19.059666 kernel: kvm-guest: stealtime: cpu 0, msr 9a41c0c0 Aug 13 01:04:19.059672 kernel: kvm-guest: PV spinlocks enabled Aug 13 01:04:19.059682 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Aug 13 01:04:19.059689 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Aug 13 01:04:19.059695 kernel: Policy zone: DMA32 Aug 13 01:04:19.059703 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8f8aacd9fbcdd713563d390e899e90bedf5577e4b1b261b4e57687d87edd6b57 Aug 13 01:04:19.059710 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 01:04:19.059717 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 01:04:19.059723 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 01:04:19.059730 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 01:04:19.059740 kernel: Memory: 2436696K/2571752K available (12295K kernel code, 2276K rwdata, 13732K rodata, 47488K init, 4092K bss, 134796K reserved, 0K cma-reserved) Aug 13 01:04:19.059747 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Aug 13 01:04:19.059754 kernel: ftrace: allocating 34608 entries in 136 pages Aug 13 01:04:19.059760 kernel: ftrace: allocated 136 pages with 2 groups Aug 13 01:04:19.059767 kernel: rcu: Hierarchical RCU implementation. Aug 13 01:04:19.059774 kernel: rcu: RCU event tracing is enabled. Aug 13 01:04:19.059780 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Aug 13 01:04:19.059787 kernel: Rude variant of Tasks RCU enabled. Aug 13 01:04:19.059794 kernel: Tracing variant of Tasks RCU enabled. Aug 13 01:04:19.059810 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 01:04:19.059817 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Aug 13 01:04:19.059824 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Aug 13 01:04:19.059830 kernel: random: crng init done Aug 13 01:04:19.059837 kernel: Console: colour VGA+ 80x25 Aug 13 01:04:19.059843 kernel: printk: console [ttyS0] enabled Aug 13 01:04:19.059850 kernel: ACPI: Core revision 20210730 Aug 13 01:04:19.059857 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Aug 13 01:04:19.059864 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 01:04:19.059875 kernel: x2apic enabled Aug 13 01:04:19.059881 kernel: Switched APIC routing to physical x2apic. Aug 13 01:04:19.059890 kernel: kvm-guest: setup PV IPIs Aug 13 01:04:19.059896 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Aug 13 01:04:19.059903 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Aug 13 01:04:19.059912 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Aug 13 01:04:19.059919 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Aug 13 01:04:19.059925 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Aug 13 01:04:19.059932 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Aug 13 01:04:19.059947 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 01:04:19.059954 kernel: Spectre V2 : Mitigation: Retpolines Aug 13 01:04:19.059961 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Aug 13 01:04:19.059969 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Aug 13 01:04:19.059976 kernel: RETBleed: Mitigation: untrained return thunk Aug 13 01:04:19.059983 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Aug 13 01:04:19.059990 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Aug 13 01:04:19.059997 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 01:04:19.060004 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 01:04:19.060012 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 01:04:19.060019 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 01:04:19.060026 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Aug 13 01:04:19.060033 kernel: Freeing SMP alternatives memory: 32K Aug 13 01:04:19.060040 kernel: pid_max: default: 32768 minimum: 301 Aug 13 01:04:19.060047 kernel: LSM: Security Framework initializing Aug 13 01:04:19.060054 kernel: SELinux: Initializing. Aug 13 01:04:19.060062 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 01:04:19.060069 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 01:04:19.060076 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Aug 13 01:04:19.060083 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Aug 13 01:04:19.060090 kernel: ... version: 0 Aug 13 01:04:19.060097 kernel: ... bit width: 48 Aug 13 01:04:19.060103 kernel: ... generic registers: 6 Aug 13 01:04:19.060110 kernel: ... value mask: 0000ffffffffffff Aug 13 01:04:19.060117 kernel: ... max period: 00007fffffffffff Aug 13 01:04:19.060125 kernel: ... fixed-purpose events: 0 Aug 13 01:04:19.060132 kernel: ... event mask: 000000000000003f Aug 13 01:04:19.060139 kernel: signal: max sigframe size: 1776 Aug 13 01:04:19.060146 kernel: rcu: Hierarchical SRCU implementation. Aug 13 01:04:19.060152 kernel: smp: Bringing up secondary CPUs ... Aug 13 01:04:19.060159 kernel: x86: Booting SMP configuration: Aug 13 01:04:19.060166 kernel: .... node #0, CPUs: #1 Aug 13 01:04:19.060173 kernel: kvm-clock: cpu 1, msr 5a19e041, secondary cpu clock Aug 13 01:04:19.060180 kernel: kvm-guest: setup async PF for cpu 1 Aug 13 01:04:19.060188 kernel: kvm-guest: stealtime: cpu 1, msr 9a49c0c0 Aug 13 01:04:19.060195 kernel: #2 Aug 13 01:04:19.060202 kernel: kvm-clock: cpu 2, msr 5a19e081, secondary cpu clock Aug 13 01:04:19.060208 kernel: kvm-guest: setup async PF for cpu 2 Aug 13 01:04:19.060215 kernel: kvm-guest: stealtime: cpu 2, msr 9a51c0c0 Aug 13 01:04:19.060222 kernel: #3 Aug 13 01:04:19.060232 kernel: kvm-clock: cpu 3, msr 5a19e0c1, secondary cpu clock Aug 13 01:04:19.060250 kernel: kvm-guest: setup async PF for cpu 3 Aug 13 01:04:19.060257 kernel: kvm-guest: stealtime: cpu 3, msr 9a59c0c0 Aug 13 01:04:19.060264 kernel: smp: Brought up 1 node, 4 CPUs Aug 13 01:04:19.060273 kernel: smpboot: Max logical packages: 1 Aug 13 01:04:19.060280 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Aug 13 01:04:19.060287 kernel: devtmpfs: initialized Aug 13 01:04:19.060294 kernel: x86/mm: Memory block size: 128MB Aug 13 01:04:19.060301 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 01:04:19.060308 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Aug 13 01:04:19.060315 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 01:04:19.060321 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 01:04:19.060328 kernel: audit: initializing netlink subsys (disabled) Aug 13 01:04:19.060337 kernel: audit: type=2000 audit(1755047059.030:1): state=initialized audit_enabled=0 res=1 Aug 13 01:04:19.060343 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 01:04:19.060350 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 01:04:19.060358 kernel: cpuidle: using governor menu Aug 13 01:04:19.060364 kernel: ACPI: bus type PCI registered Aug 13 01:04:19.060371 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 01:04:19.060378 kernel: dca service started, version 1.12.1 Aug 13 01:04:19.060385 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Aug 13 01:04:19.060392 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved in E820 Aug 13 01:04:19.060401 kernel: PCI: Using configuration type 1 for base access Aug 13 01:04:19.060407 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 01:04:19.060415 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 01:04:19.060421 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 01:04:19.060428 kernel: ACPI: Added _OSI(Module Device) Aug 13 01:04:19.060435 kernel: ACPI: Added _OSI(Processor Device) Aug 13 01:04:19.060442 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 01:04:19.060449 kernel: ACPI: Added _OSI(Linux-Dell-Video) Aug 13 01:04:19.060456 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Aug 13 01:04:19.060464 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Aug 13 01:04:19.060471 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Aug 13 01:04:19.060478 kernel: ACPI: Interpreter enabled Aug 13 01:04:19.060485 kernel: ACPI: PM: (supports S0 S3 S5) Aug 13 01:04:19.060492 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 01:04:19.060499 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 01:04:19.060506 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Aug 13 01:04:19.060512 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Aug 13 01:04:19.060698 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 01:04:19.060783 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Aug 13 01:04:19.060868 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Aug 13 01:04:19.060878 kernel: PCI host bridge to bus 0000:00 Aug 13 01:04:19.060972 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Aug 13 01:04:19.061042 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Aug 13 01:04:19.061109 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Aug 13 01:04:19.061179 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Aug 13 01:04:19.061260 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Aug 13 01:04:19.061329 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Aug 13 01:04:19.061394 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Aug 13 01:04:19.061507 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Aug 13 01:04:19.061601 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Aug 13 01:04:19.061680 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Aug 13 01:04:19.061758 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Aug 13 01:04:19.061841 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Aug 13 01:04:19.061915 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Aug 13 01:04:19.062012 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Aug 13 01:04:19.062093 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Aug 13 01:04:19.062174 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Aug 13 01:04:19.062265 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Aug 13 01:04:19.062377 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Aug 13 01:04:19.062456 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Aug 13 01:04:19.062530 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Aug 13 01:04:19.062604 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Aug 13 01:04:19.062692 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Aug 13 01:04:19.062771 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Aug 13 01:04:19.062859 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Aug 13 01:04:19.062934 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Aug 13 01:04:19.063008 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Aug 13 01:04:19.063093 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Aug 13 01:04:19.063169 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Aug 13 01:04:19.063272 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Aug 13 01:04:19.063358 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Aug 13 01:04:19.063444 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Aug 13 01:04:19.063532 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Aug 13 01:04:19.063615 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Aug 13 01:04:19.063625 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Aug 13 01:04:19.063633 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Aug 13 01:04:19.063640 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Aug 13 01:04:19.063647 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Aug 13 01:04:19.063654 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Aug 13 01:04:19.063663 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Aug 13 01:04:19.063670 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Aug 13 01:04:19.063677 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Aug 13 01:04:19.063684 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Aug 13 01:04:19.063691 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Aug 13 01:04:19.063698 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Aug 13 01:04:19.063705 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Aug 13 01:04:19.063712 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Aug 13 01:04:19.063719 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Aug 13 01:04:19.063727 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Aug 13 01:04:19.063734 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Aug 13 01:04:19.063741 kernel: iommu: Default domain type: Translated Aug 13 01:04:19.063748 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 01:04:19.063891 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Aug 13 01:04:19.064039 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Aug 13 01:04:19.064145 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Aug 13 01:04:19.064156 kernel: vgaarb: loaded Aug 13 01:04:19.064166 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 01:04:19.064173 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 01:04:19.064180 kernel: PTP clock support registered Aug 13 01:04:19.064187 kernel: PCI: Using ACPI for IRQ routing Aug 13 01:04:19.064194 kernel: PCI: pci_cache_line_size set to 64 bytes Aug 13 01:04:19.064201 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Aug 13 01:04:19.064208 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Aug 13 01:04:19.064215 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Aug 13 01:04:19.064234 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Aug 13 01:04:19.064267 kernel: clocksource: Switched to clocksource kvm-clock Aug 13 01:04:19.064275 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 01:04:19.064282 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 01:04:19.064289 kernel: pnp: PnP ACPI init Aug 13 01:04:19.064406 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Aug 13 01:04:19.064418 kernel: pnp: PnP ACPI: found 6 devices Aug 13 01:04:19.064425 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 01:04:19.064432 kernel: NET: Registered PF_INET protocol family Aug 13 01:04:19.064442 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 01:04:19.064449 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Aug 13 01:04:19.064456 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 01:04:19.064463 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Aug 13 01:04:19.064470 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Aug 13 01:04:19.064477 kernel: TCP: Hash tables configured (established 32768 bind 32768) Aug 13 01:04:19.064484 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 01:04:19.064491 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 01:04:19.064498 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 01:04:19.064506 kernel: NET: Registered PF_XDP protocol family Aug 13 01:04:19.064580 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Aug 13 01:04:19.064648 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Aug 13 01:04:19.064719 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Aug 13 01:04:19.064795 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Aug 13 01:04:19.064874 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Aug 13 01:04:19.064944 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Aug 13 01:04:19.064954 kernel: PCI: CLS 0 bytes, default 64 Aug 13 01:04:19.064964 kernel: Initialise system trusted keyrings Aug 13 01:04:19.064971 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Aug 13 01:04:19.064978 kernel: Key type asymmetric registered Aug 13 01:04:19.064985 kernel: Asymmetric key parser 'x509' registered Aug 13 01:04:19.064992 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Aug 13 01:04:19.064999 kernel: io scheduler mq-deadline registered Aug 13 01:04:19.065006 kernel: io scheduler kyber registered Aug 13 01:04:19.065013 kernel: io scheduler bfq registered Aug 13 01:04:19.065020 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 01:04:19.065027 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Aug 13 01:04:19.065035 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Aug 13 01:04:19.065042 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Aug 13 01:04:19.065050 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 01:04:19.065057 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 01:04:19.065064 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Aug 13 01:04:19.065071 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Aug 13 01:04:19.065078 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Aug 13 01:04:19.065165 kernel: rtc_cmos 00:04: RTC can wake from S4 Aug 13 01:04:19.065178 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Aug 13 01:04:19.065291 kernel: rtc_cmos 00:04: registered as rtc0 Aug 13 01:04:19.065365 kernel: rtc_cmos 00:04: setting system clock to 2025-08-13T01:04:18 UTC (1755047058) Aug 13 01:04:19.065435 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Aug 13 01:04:19.065445 kernel: NET: Registered PF_INET6 protocol family Aug 13 01:04:19.065452 kernel: Segment Routing with IPv6 Aug 13 01:04:19.065459 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 01:04:19.065466 kernel: NET: Registered PF_PACKET protocol family Aug 13 01:04:19.065476 kernel: Key type dns_resolver registered Aug 13 01:04:19.065482 kernel: IPI shorthand broadcast: enabled Aug 13 01:04:19.065489 kernel: sched_clock: Marking stable (464199388, 153552987)->(686051636, -68299261) Aug 13 01:04:19.065496 kernel: registered taskstats version 1 Aug 13 01:04:19.065503 kernel: Loading compiled-in X.509 certificates Aug 13 01:04:19.065511 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.189-flatcar: 1d5a64b5798e654719a8bd91d683e7e9894bd433' Aug 13 01:04:19.065518 kernel: Key type .fscrypt registered Aug 13 01:04:19.065524 kernel: Key type fscrypt-provisioning registered Aug 13 01:04:19.065531 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 01:04:19.065540 kernel: ima: Allocated hash algorithm: sha1 Aug 13 01:04:19.065546 kernel: ima: No architecture policies found Aug 13 01:04:19.065553 kernel: clk: Disabling unused clocks Aug 13 01:04:19.065560 kernel: Freeing unused kernel image (initmem) memory: 47488K Aug 13 01:04:19.065567 kernel: Write protecting the kernel read-only data: 28672k Aug 13 01:04:19.065574 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Aug 13 01:04:19.065581 kernel: Freeing unused kernel image (rodata/data gap) memory: 604K Aug 13 01:04:19.065588 kernel: Run /init as init process Aug 13 01:04:19.065595 kernel: with arguments: Aug 13 01:04:19.065603 kernel: /init Aug 13 01:04:19.065610 kernel: with environment: Aug 13 01:04:19.065617 kernel: HOME=/ Aug 13 01:04:19.065623 kernel: TERM=linux Aug 13 01:04:19.065630 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 01:04:19.065639 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Aug 13 01:04:19.065648 systemd[1]: Detected virtualization kvm. Aug 13 01:04:19.065656 systemd[1]: Detected architecture x86-64. Aug 13 01:04:19.065665 systemd[1]: Running in initrd. Aug 13 01:04:19.065672 systemd[1]: No hostname configured, using default hostname. Aug 13 01:04:19.065679 systemd[1]: Hostname set to . Aug 13 01:04:19.065687 systemd[1]: Initializing machine ID from VM UUID. Aug 13 01:04:19.065695 systemd[1]: Queued start job for default target initrd.target. Aug 13 01:04:19.065702 systemd[1]: Started systemd-ask-password-console.path. Aug 13 01:04:19.065709 systemd[1]: Reached target cryptsetup.target. Aug 13 01:04:19.065717 systemd[1]: Reached target paths.target. Aug 13 01:04:19.065725 systemd[1]: Reached target slices.target. Aug 13 01:04:19.065733 systemd[1]: Reached target swap.target. Aug 13 01:04:19.065746 systemd[1]: Reached target timers.target. Aug 13 01:04:19.065755 systemd[1]: Listening on iscsid.socket. Aug 13 01:04:19.065763 systemd[1]: Listening on iscsiuio.socket. Aug 13 01:04:19.065771 systemd[1]: Listening on systemd-journald-audit.socket. Aug 13 01:04:19.065779 systemd[1]: Listening on systemd-journald-dev-log.socket. Aug 13 01:04:19.065788 systemd[1]: Listening on systemd-journald.socket. Aug 13 01:04:19.065796 systemd[1]: Listening on systemd-networkd.socket. Aug 13 01:04:19.065814 systemd[1]: Listening on systemd-udevd-control.socket. Aug 13 01:04:19.065821 systemd[1]: Listening on systemd-udevd-kernel.socket. Aug 13 01:04:19.065829 systemd[1]: Reached target sockets.target. Aug 13 01:04:19.065837 systemd[1]: Starting kmod-static-nodes.service... Aug 13 01:04:19.065844 systemd[1]: Finished network-cleanup.service. Aug 13 01:04:19.065852 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 01:04:19.065861 systemd[1]: Starting systemd-journald.service... Aug 13 01:04:19.065869 systemd[1]: Starting systemd-modules-load.service... Aug 13 01:04:19.065877 systemd[1]: Starting systemd-resolved.service... Aug 13 01:04:19.065884 systemd[1]: Starting systemd-vconsole-setup.service... Aug 13 01:04:19.065892 systemd[1]: Finished kmod-static-nodes.service. Aug 13 01:04:19.065903 systemd-journald[197]: Journal started Aug 13 01:04:19.065946 systemd-journald[197]: Runtime Journal (/run/log/journal/723d83f0fa2746858b59d10f233075a2) is 6.0M, max 48.5M, 42.5M free. Aug 13 01:04:19.064276 systemd-modules-load[198]: Inserted module 'overlay' Aug 13 01:04:19.100159 systemd[1]: Started systemd-journald.service. Aug 13 01:04:19.100201 kernel: audit: type=1130 audit(1755047059.094:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.094000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.079327 systemd-resolved[199]: Positive Trust Anchors: Aug 13 01:04:19.104489 kernel: audit: type=1130 audit(1755047059.100:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.100000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.079336 systemd-resolved[199]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 01:04:19.104000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.079364 systemd-resolved[199]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Aug 13 01:04:19.117516 kernel: audit: type=1130 audit(1755047059.104:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.117534 kernel: audit: type=1130 audit(1755047059.108:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.108000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.081741 systemd-resolved[199]: Defaulting to hostname 'linux'. Aug 13 01:04:19.121902 kernel: audit: type=1130 audit(1755047059.118:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.118000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.101302 systemd[1]: Started systemd-resolved.service. Aug 13 01:04:19.105391 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 01:04:19.108578 systemd[1]: Finished systemd-vconsole-setup.service. Aug 13 01:04:19.118487 systemd[1]: Reached target nss-lookup.target. Aug 13 01:04:19.123376 systemd[1]: Starting dracut-cmdline-ask.service... Aug 13 01:04:19.124762 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Aug 13 01:04:19.132262 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 01:04:19.132641 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Aug 13 01:04:19.136282 kernel: audit: type=1130 audit(1755047059.132:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.132000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.140071 systemd[1]: Finished dracut-cmdline-ask.service. Aug 13 01:04:19.141702 systemd[1]: Starting dracut-cmdline.service... Aug 13 01:04:19.140000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.145257 kernel: audit: type=1130 audit(1755047059.140:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.146413 systemd-modules-load[198]: Inserted module 'br_netfilter' Aug 13 01:04:19.147315 kernel: Bridge firewalling registered Aug 13 01:04:19.150831 dracut-cmdline[216]: dracut-dracut-053 Aug 13 01:04:19.152917 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8f8aacd9fbcdd713563d390e899e90bedf5577e4b1b261b4e57687d87edd6b57 Aug 13 01:04:19.170266 kernel: SCSI subsystem initialized Aug 13 01:04:19.180750 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 01:04:19.180774 kernel: device-mapper: uevent: version 1.0.3 Aug 13 01:04:19.182010 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Aug 13 01:04:19.184740 systemd-modules-load[198]: Inserted module 'dm_multipath' Aug 13 01:04:19.186360 systemd[1]: Finished systemd-modules-load.service. Aug 13 01:04:19.190726 kernel: audit: type=1130 audit(1755047059.186:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.186000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.187106 systemd[1]: Starting systemd-sysctl.service... Aug 13 01:04:19.196823 systemd[1]: Finished systemd-sysctl.service. Aug 13 01:04:19.197000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.201269 kernel: audit: type=1130 audit(1755047059.197:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.222270 kernel: Loading iSCSI transport class v2.0-870. Aug 13 01:04:19.238264 kernel: iscsi: registered transport (tcp) Aug 13 01:04:19.260280 kernel: iscsi: registered transport (qla4xxx) Aug 13 01:04:19.260312 kernel: QLogic iSCSI HBA Driver Aug 13 01:04:19.293155 systemd[1]: Finished dracut-cmdline.service. Aug 13 01:04:19.292000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.294033 systemd[1]: Starting dracut-pre-udev.service... Aug 13 01:04:19.341269 kernel: raid6: avx2x4 gen() 30484 MB/s Aug 13 01:04:19.358267 kernel: raid6: avx2x4 xor() 7973 MB/s Aug 13 01:04:19.394261 kernel: raid6: avx2x2 gen() 32223 MB/s Aug 13 01:04:19.411261 kernel: raid6: avx2x2 xor() 19220 MB/s Aug 13 01:04:19.428259 kernel: raid6: avx2x1 gen() 26182 MB/s Aug 13 01:04:19.495264 kernel: raid6: avx2x1 xor() 15271 MB/s Aug 13 01:04:19.512260 kernel: raid6: sse2x4 gen() 14685 MB/s Aug 13 01:04:19.529260 kernel: raid6: sse2x4 xor() 7616 MB/s Aug 13 01:04:19.546265 kernel: raid6: sse2x2 gen() 16214 MB/s Aug 13 01:04:19.599289 kernel: raid6: sse2x2 xor() 9717 MB/s Aug 13 01:04:19.616265 kernel: raid6: sse2x1 gen() 12288 MB/s Aug 13 01:04:19.633623 kernel: raid6: sse2x1 xor() 7672 MB/s Aug 13 01:04:19.633637 kernel: raid6: using algorithm avx2x2 gen() 32223 MB/s Aug 13 01:04:19.633647 kernel: raid6: .... xor() 19220 MB/s, rmw enabled Aug 13 01:04:19.634319 kernel: raid6: using avx2x2 recovery algorithm Aug 13 01:04:19.647264 kernel: xor: automatically using best checksumming function avx Aug 13 01:04:19.740267 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Aug 13 01:04:19.749986 systemd[1]: Finished dracut-pre-udev.service. Aug 13 01:04:19.750000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.751000 audit: BPF prog-id=7 op=LOAD Aug 13 01:04:19.751000 audit: BPF prog-id=8 op=LOAD Aug 13 01:04:19.752205 systemd[1]: Starting systemd-udevd.service... Aug 13 01:04:19.764840 systemd-udevd[399]: Using default interface naming scheme 'v252'. Aug 13 01:04:19.768958 systemd[1]: Started systemd-udevd.service. Aug 13 01:04:19.799000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.801354 systemd[1]: Starting dracut-pre-trigger.service... Aug 13 01:04:19.811912 dracut-pre-trigger[406]: rd.md=0: removing MD RAID activation Aug 13 01:04:19.838153 systemd[1]: Finished dracut-pre-trigger.service. Aug 13 01:04:19.839053 systemd[1]: Starting systemd-udev-trigger.service... Aug 13 01:04:19.838000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.877325 systemd[1]: Finished systemd-udev-trigger.service. Aug 13 01:04:19.893000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:19.910265 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Aug 13 01:04:19.934899 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 01:04:19.934914 kernel: GPT:9289727 != 19775487 Aug 13 01:04:19.934922 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 01:04:19.934931 kernel: GPT:9289727 != 19775487 Aug 13 01:04:19.934940 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 01:04:19.934948 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 01:04:19.937268 kernel: libata version 3.00 loaded. Aug 13 01:04:19.938265 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 01:04:19.949950 kernel: AVX2 version of gcm_enc/dec engaged. Aug 13 01:04:19.949974 kernel: AES CTR mode by8 optimization enabled Aug 13 01:04:19.962264 kernel: ahci 0000:00:1f.2: version 3.0 Aug 13 01:04:19.973033 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Aug 13 01:04:19.973049 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Aug 13 01:04:19.973143 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Aug 13 01:04:19.973227 kernel: scsi host0: ahci Aug 13 01:04:19.973364 kernel: scsi host1: ahci Aug 13 01:04:19.973456 kernel: scsi host2: ahci Aug 13 01:04:19.973552 kernel: scsi host3: ahci Aug 13 01:04:19.973647 kernel: scsi host4: ahci Aug 13 01:04:19.973736 kernel: scsi host5: ahci Aug 13 01:04:19.973844 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Aug 13 01:04:19.973854 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Aug 13 01:04:19.973864 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Aug 13 01:04:19.973872 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Aug 13 01:04:19.973881 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Aug 13 01:04:19.973892 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Aug 13 01:04:19.967449 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Aug 13 01:04:20.072499 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (440) Aug 13 01:04:20.012486 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Aug 13 01:04:20.069314 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Aug 13 01:04:20.076555 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Aug 13 01:04:20.083226 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Aug 13 01:04:20.085307 systemd[1]: Starting disk-uuid.service... Aug 13 01:04:20.243331 disk-uuid[524]: Primary Header is updated. Aug 13 01:04:20.243331 disk-uuid[524]: Secondary Entries is updated. Aug 13 01:04:20.243331 disk-uuid[524]: Secondary Header is updated. Aug 13 01:04:20.247141 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 01:04:20.250258 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 01:04:20.254266 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 01:04:20.278270 kernel: ata1: SATA link down (SStatus 0 SControl 300) Aug 13 01:04:20.278325 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Aug 13 01:04:20.279610 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Aug 13 01:04:20.279637 kernel: ata3.00: applying bridge limits Aug 13 01:04:20.281880 kernel: ata2: SATA link down (SStatus 0 SControl 300) Aug 13 01:04:20.281940 kernel: ata3.00: configured for UDMA/100 Aug 13 01:04:20.282258 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Aug 13 01:04:20.286255 kernel: ata5: SATA link down (SStatus 0 SControl 300) Aug 13 01:04:20.287268 kernel: ata4: SATA link down (SStatus 0 SControl 300) Aug 13 01:04:20.287294 kernel: ata6: SATA link down (SStatus 0 SControl 300) Aug 13 01:04:20.333327 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Aug 13 01:04:20.351001 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Aug 13 01:04:20.351022 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Aug 13 01:04:21.324042 disk-uuid[525]: The operation has completed successfully. Aug 13 01:04:21.325618 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Aug 13 01:04:21.345340 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 01:04:21.345000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.345000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.345435 systemd[1]: Finished disk-uuid.service. Aug 13 01:04:21.351825 systemd[1]: Starting verity-setup.service... Aug 13 01:04:21.366281 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Aug 13 01:04:21.391299 systemd[1]: Found device dev-mapper-usr.device. Aug 13 01:04:21.394470 systemd[1]: Mounting sysusr-usr.mount... Aug 13 01:04:21.399339 systemd[1]: Finished verity-setup.service. Aug 13 01:04:21.398000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.562298 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Aug 13 01:04:21.562680 systemd[1]: Mounted sysusr-usr.mount. Aug 13 01:04:21.562926 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Aug 13 01:04:21.563857 systemd[1]: Starting ignition-setup.service... Aug 13 01:04:21.565781 systemd[1]: Starting parse-ip-for-networkd.service... Aug 13 01:04:21.574369 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Aug 13 01:04:21.574414 kernel: BTRFS info (device vda6): using free space tree Aug 13 01:04:21.574429 kernel: BTRFS info (device vda6): has skinny extents Aug 13 01:04:21.585081 systemd[1]: mnt-oem.mount: Deactivated successfully. Aug 13 01:04:21.596000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.595455 systemd[1]: Finished ignition-setup.service. Aug 13 01:04:21.597281 systemd[1]: Starting ignition-fetch-offline.service... Aug 13 01:04:21.674475 systemd[1]: Finished parse-ip-for-networkd.service. Aug 13 01:04:21.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.678000 audit: BPF prog-id=9 op=LOAD Aug 13 01:04:21.679674 systemd[1]: Starting systemd-networkd.service... Aug 13 01:04:21.754441 systemd-networkd[710]: lo: Link UP Aug 13 01:04:21.754452 systemd-networkd[710]: lo: Gained carrier Aug 13 01:04:21.755118 systemd-networkd[710]: Enumeration completed Aug 13 01:04:21.755412 systemd-networkd[710]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 01:04:21.757973 systemd-networkd[710]: eth0: Link UP Aug 13 01:04:21.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.757977 systemd-networkd[710]: eth0: Gained carrier Aug 13 01:04:21.758371 systemd[1]: Started systemd-networkd.service. Aug 13 01:04:21.760087 systemd[1]: Reached target network.target. Aug 13 01:04:21.762852 systemd[1]: Starting iscsiuio.service... Aug 13 01:04:21.772049 ignition[649]: Ignition 2.14.0 Aug 13 01:04:21.772062 ignition[649]: Stage: fetch-offline Aug 13 01:04:21.772161 ignition[649]: no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:21.772174 ignition[649]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:21.772318 ignition[649]: parsed url from cmdline: "" Aug 13 01:04:21.772322 ignition[649]: no config URL provided Aug 13 01:04:21.772327 ignition[649]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 01:04:21.772334 ignition[649]: no config at "/usr/lib/ignition/user.ign" Aug 13 01:04:21.772354 ignition[649]: op(1): [started] loading QEMU firmware config module Aug 13 01:04:21.772359 ignition[649]: op(1): executing: "modprobe" "qemu_fw_cfg" Aug 13 01:04:21.777211 ignition[649]: op(1): [finished] loading QEMU firmware config module Aug 13 01:04:21.822593 ignition[649]: parsing config with SHA512: dbf4dbe13ef38e969c1fb99908d2197f9e2cec20bd36c52b52c1ef7636f0b13adb475e440c3431ac2bdcce5dd8cb1eee5f74d11f818e33f34272aa706e258b48 Aug 13 01:04:21.831630 unknown[649]: fetched base config from "system" Aug 13 01:04:21.831641 unknown[649]: fetched user config from "qemu" Aug 13 01:04:21.834000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.832157 ignition[649]: fetch-offline: fetch-offline passed Aug 13 01:04:21.833400 systemd[1]: Finished ignition-fetch-offline.service. Aug 13 01:04:21.832276 ignition[649]: Ignition finished successfully Aug 13 01:04:21.835091 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 01:04:21.836125 systemd[1]: Starting ignition-kargs.service... Aug 13 01:04:21.854552 systemd[1]: Started iscsiuio.service. Aug 13 01:04:21.855000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.856387 systemd-networkd[710]: eth0: DHCPv4 address 10.0.0.132/16, gateway 10.0.0.1 acquired from 10.0.0.1 Aug 13 01:04:21.858293 systemd[1]: Starting iscsid.service... Aug 13 01:04:21.890090 ignition[721]: Ignition 2.14.0 Aug 13 01:04:21.890102 ignition[721]: Stage: kargs Aug 13 01:04:21.899000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.900704 iscsid[728]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Aug 13 01:04:21.900704 iscsid[728]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Aug 13 01:04:21.900704 iscsid[728]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Aug 13 01:04:21.900704 iscsid[728]: If using hardware iscsi like qla4xxx this message can be ignored. Aug 13 01:04:21.900704 iscsid[728]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Aug 13 01:04:21.900704 iscsid[728]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Aug 13 01:04:21.896873 systemd[1]: Started iscsid.service. Aug 13 01:04:21.890274 ignition[721]: no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:21.913000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.916000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.901212 systemd[1]: Starting dracut-initqueue.service... Aug 13 01:04:21.890283 ignition[721]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:21.913456 systemd[1]: Finished ignition-kargs.service. Aug 13 01:04:21.891682 ignition[721]: kargs: kargs passed Aug 13 01:04:21.914262 systemd[1]: Finished dracut-initqueue.service. Aug 13 01:04:21.891732 ignition[721]: Ignition finished successfully Aug 13 01:04:21.916424 systemd[1]: Reached target remote-fs-pre.target. Aug 13 01:04:21.917463 systemd[1]: Reached target remote-cryptsetup.target. Aug 13 01:04:21.918313 systemd[1]: Reached target remote-fs.target. Aug 13 01:04:21.920496 systemd[1]: Starting dracut-pre-mount.service... Aug 13 01:04:21.922750 systemd[1]: Starting ignition-disks.service... Aug 13 01:04:21.935439 ignition[738]: Ignition 2.14.0 Aug 13 01:04:21.935450 ignition[738]: Stage: disks Aug 13 01:04:21.935591 ignition[738]: no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:21.935603 ignition[738]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:21.938060 systemd[1]: Finished ignition-disks.service. Aug 13 01:04:21.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:21.937088 ignition[738]: disks: disks passed Aug 13 01:04:21.940149 systemd[1]: Reached target initrd-root-device.target. Aug 13 01:04:21.937131 ignition[738]: Ignition finished successfully Aug 13 01:04:21.941890 systemd[1]: Reached target local-fs-pre.target. Aug 13 01:04:21.942837 systemd[1]: Reached target local-fs.target. Aug 13 01:04:21.942907 systemd[1]: Reached target sysinit.target. Aug 13 01:04:21.943091 systemd[1]: Reached target basic.target. Aug 13 01:04:22.004113 systemd[1]: Finished dracut-pre-mount.service. Aug 13 01:04:22.005000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:22.006678 systemd[1]: Starting systemd-fsck-root.service... Aug 13 01:04:22.018947 systemd-fsck[751]: ROOT: clean, 629/553520 files, 56027/553472 blocks Aug 13 01:04:22.024956 systemd[1]: Finished systemd-fsck-root.service. Aug 13 01:04:22.026000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:22.026885 systemd[1]: Mounting sysroot.mount... Aug 13 01:04:22.034261 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Aug 13 01:04:22.034682 systemd[1]: Mounted sysroot.mount. Aug 13 01:04:22.035444 systemd[1]: Reached target initrd-root-fs.target. Aug 13 01:04:22.037823 systemd[1]: Mounting sysroot-usr.mount... Aug 13 01:04:22.038866 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Aug 13 01:04:22.038894 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 01:04:22.038914 systemd[1]: Reached target ignition-diskful.target. Aug 13 01:04:22.041313 systemd[1]: Mounted sysroot-usr.mount. Aug 13 01:04:22.043223 systemd[1]: Starting initrd-setup-root.service... Aug 13 01:04:22.049575 initrd-setup-root[761]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 01:04:22.054192 initrd-setup-root[769]: cut: /sysroot/etc/group: No such file or directory Aug 13 01:04:22.057928 initrd-setup-root[777]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 01:04:22.062128 initrd-setup-root[785]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 01:04:22.091807 systemd[1]: Finished initrd-setup-root.service. Aug 13 01:04:22.093000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:22.094468 systemd[1]: Starting ignition-mount.service... Aug 13 01:04:22.096909 systemd[1]: Starting sysroot-boot.service... Aug 13 01:04:22.101094 bash[802]: umount: /sysroot/usr/share/oem: not mounted. Aug 13 01:04:22.192132 ignition[803]: INFO : Ignition 2.14.0 Aug 13 01:04:22.192132 ignition[803]: INFO : Stage: mount Aug 13 01:04:22.194846 ignition[803]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:22.194846 ignition[803]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:22.194846 ignition[803]: INFO : mount: mount passed Aug 13 01:04:22.194846 ignition[803]: INFO : Ignition finished successfully Aug 13 01:04:22.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:22.193997 systemd[1]: Finished ignition-mount.service. Aug 13 01:04:22.202609 systemd[1]: Finished sysroot-boot.service. Aug 13 01:04:22.203000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:22.502307 systemd[1]: Mounting sysroot-usr-share-oem.mount... Aug 13 01:04:22.512036 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (813) Aug 13 01:04:22.512064 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Aug 13 01:04:22.512081 kernel: BTRFS info (device vda6): using free space tree Aug 13 01:04:22.512815 kernel: BTRFS info (device vda6): has skinny extents Aug 13 01:04:22.517399 systemd[1]: Mounted sysroot-usr-share-oem.mount. Aug 13 01:04:22.520230 systemd[1]: Starting ignition-files.service... Aug 13 01:04:22.536316 ignition[833]: INFO : Ignition 2.14.0 Aug 13 01:04:22.536316 ignition[833]: INFO : Stage: files Aug 13 01:04:22.538458 ignition[833]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:22.538458 ignition[833]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:22.538458 ignition[833]: DEBUG : files: compiled without relabeling support, skipping Aug 13 01:04:22.542436 ignition[833]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 01:04:22.542436 ignition[833]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 01:04:22.545480 ignition[833]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 01:04:22.547035 ignition[833]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 01:04:22.548884 unknown[833]: wrote ssh authorized keys file for user: core Aug 13 01:04:22.550106 ignition[833]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 01:04:22.551989 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 01:04:22.554061 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Aug 13 01:04:22.611954 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 01:04:22.820135 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Aug 13 01:04:22.822360 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 01:04:22.822360 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Aug 13 01:04:22.919504 systemd-networkd[710]: eth0: Gained IPv6LL Aug 13 01:04:23.041012 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Aug 13 01:04:23.293434 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 01:04:23.295508 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Aug 13 01:04:23.615977 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Aug 13 01:04:24.201316 ignition[833]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Aug 13 01:04:24.201316 ignition[833]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Aug 13 01:04:24.204642 ignition[833]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 01:04:24.206458 ignition[833]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 01:04:24.206458 ignition[833]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Aug 13 01:04:24.206458 ignition[833]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Aug 13 01:04:24.210589 ignition[833]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Aug 13 01:04:24.210589 ignition[833]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Aug 13 01:04:24.210589 ignition[833]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Aug 13 01:04:24.210589 ignition[833]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Aug 13 01:04:24.217027 ignition[833]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 01:04:24.217027 ignition[833]: INFO : files: op(11): [started] setting preset to disabled for "coreos-metadata.service" Aug 13 01:04:24.219777 ignition[833]: INFO : files: op(11): op(12): [started] removing enablement symlink(s) for "coreos-metadata.service" Aug 13 01:04:24.246732 ignition[833]: INFO : files: op(11): op(12): [finished] removing enablement symlink(s) for "coreos-metadata.service" Aug 13 01:04:24.249199 ignition[833]: INFO : files: op(11): [finished] setting preset to disabled for "coreos-metadata.service" Aug 13 01:04:24.249199 ignition[833]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 01:04:24.249199 ignition[833]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 01:04:24.249199 ignition[833]: INFO : files: files passed Aug 13 01:04:24.249199 ignition[833]: INFO : Ignition finished successfully Aug 13 01:04:24.273293 kernel: kauditd_printk_skb: 25 callbacks suppressed Aug 13 01:04:24.273318 kernel: audit: type=1130 audit(1755047064.248:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.273330 kernel: audit: type=1130 audit(1755047064.260:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.273341 kernel: audit: type=1130 audit(1755047064.264:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.273352 kernel: audit: type=1131 audit(1755047064.264:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.248000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.260000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.264000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.248377 systemd[1]: Finished ignition-files.service. Aug 13 01:04:24.250390 systemd[1]: Starting initrd-setup-root-after-ignition.service... Aug 13 01:04:24.255381 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Aug 13 01:04:24.278962 initrd-setup-root-after-ignition[857]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Aug 13 01:04:24.256109 systemd[1]: Starting ignition-quench.service... Aug 13 01:04:24.281867 initrd-setup-root-after-ignition[860]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 01:04:24.257973 systemd[1]: Finished initrd-setup-root-after-ignition.service. Aug 13 01:04:24.260619 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 01:04:24.260704 systemd[1]: Finished ignition-quench.service. Aug 13 01:04:24.287000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.265163 systemd[1]: Reached target ignition-complete.target. Aug 13 01:04:24.295834 kernel: audit: type=1130 audit(1755047064.287:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.295854 kernel: audit: type=1131 audit(1755047064.287:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.287000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.273903 systemd[1]: Starting initrd-parse-etc.service... Aug 13 01:04:24.286422 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 01:04:24.286497 systemd[1]: Finished initrd-parse-etc.service. Aug 13 01:04:24.287950 systemd[1]: Reached target initrd-fs.target. Aug 13 01:04:24.294311 systemd[1]: Reached target initrd.target. Aug 13 01:04:24.295850 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Aug 13 01:04:24.296744 systemd[1]: Starting dracut-pre-pivot.service... Aug 13 01:04:24.307355 systemd[1]: Finished dracut-pre-pivot.service. Aug 13 01:04:24.312422 kernel: audit: type=1130 audit(1755047064.307:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.307000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.308862 systemd[1]: Starting initrd-cleanup.service... Aug 13 01:04:24.318079 systemd[1]: Stopped target nss-lookup.target. Aug 13 01:04:24.318994 systemd[1]: Stopped target remote-cryptsetup.target. Aug 13 01:04:24.320594 systemd[1]: Stopped target timers.target. Aug 13 01:04:24.322198 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 01:04:24.328067 kernel: audit: type=1131 audit(1755047064.323:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.323000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.322345 systemd[1]: Stopped dracut-pre-pivot.service. Aug 13 01:04:24.323779 systemd[1]: Stopped target initrd.target. Aug 13 01:04:24.328184 systemd[1]: Stopped target basic.target. Aug 13 01:04:24.329773 systemd[1]: Stopped target ignition-complete.target. Aug 13 01:04:24.331349 systemd[1]: Stopped target ignition-diskful.target. Aug 13 01:04:24.332867 systemd[1]: Stopped target initrd-root-device.target. Aug 13 01:04:24.334560 systemd[1]: Stopped target remote-fs.target. Aug 13 01:04:24.336169 systemd[1]: Stopped target remote-fs-pre.target. Aug 13 01:04:24.337845 systemd[1]: Stopped target sysinit.target. Aug 13 01:04:24.339325 systemd[1]: Stopped target local-fs.target. Aug 13 01:04:24.340876 systemd[1]: Stopped target local-fs-pre.target. Aug 13 01:04:24.342419 systemd[1]: Stopped target swap.target. Aug 13 01:04:24.349739 kernel: audit: type=1131 audit(1755047064.345:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.345000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.343866 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 01:04:24.343955 systemd[1]: Stopped dracut-pre-mount.service. Aug 13 01:04:24.356071 kernel: audit: type=1131 audit(1755047064.351:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.351000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.345536 systemd[1]: Stopped target cryptsetup.target. Aug 13 01:04:24.355000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.349761 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 01:04:24.349847 systemd[1]: Stopped dracut-initqueue.service. Aug 13 01:04:24.351740 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 01:04:24.351825 systemd[1]: Stopped ignition-fetch-offline.service. Aug 13 01:04:24.356205 systemd[1]: Stopped target paths.target. Aug 13 01:04:24.357765 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 01:04:24.361278 systemd[1]: Stopped systemd-ask-password-console.path. Aug 13 01:04:24.362641 systemd[1]: Stopped target slices.target. Aug 13 01:04:24.367000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.364104 systemd[1]: Stopped target sockets.target. Aug 13 01:04:24.369000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.365984 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 01:04:24.366073 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Aug 13 01:04:24.373865 iscsid[728]: iscsid shutting down. Aug 13 01:04:24.367884 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 01:04:24.375000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.367965 systemd[1]: Stopped ignition-files.service. Aug 13 01:04:24.370145 systemd[1]: Stopping ignition-mount.service... Aug 13 01:04:24.379000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.372144 systemd[1]: Stopping iscsid.service... Aug 13 01:04:24.381000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.373770 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 01:04:24.374035 systemd[1]: Stopped kmod-static-nodes.service. Aug 13 01:04:24.376801 systemd[1]: Stopping sysroot-boot.service... Aug 13 01:04:24.386901 ignition[874]: INFO : Ignition 2.14.0 Aug 13 01:04:24.386901 ignition[874]: INFO : Stage: umount Aug 13 01:04:24.386901 ignition[874]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 01:04:24.386901 ignition[874]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Aug 13 01:04:24.386901 ignition[874]: INFO : umount: umount passed Aug 13 01:04:24.386901 ignition[874]: INFO : Ignition finished successfully Aug 13 01:04:24.385000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.387000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.387000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.389000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.442000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.442000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.442000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.447000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.377743 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 01:04:24.377913 systemd[1]: Stopped systemd-udev-trigger.service. Aug 13 01:04:24.379909 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 01:04:24.380033 systemd[1]: Stopped dracut-pre-trigger.service. Aug 13 01:04:24.384319 systemd[1]: iscsid.service: Deactivated successfully. Aug 13 01:04:24.384466 systemd[1]: Stopped iscsid.service. Aug 13 01:04:24.386873 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 01:04:24.386948 systemd[1]: Finished initrd-cleanup.service. Aug 13 01:04:24.461000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.388061 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 01:04:24.388133 systemd[1]: Stopped ignition-mount.service. Aug 13 01:04:24.390082 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 01:04:24.390115 systemd[1]: Closed iscsid.socket. Aug 13 01:04:24.468000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.443081 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 01:04:24.468000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.470000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.443142 systemd[1]: Stopped ignition-disks.service. Aug 13 01:04:24.443815 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 01:04:24.443852 systemd[1]: Stopped ignition-kargs.service. Aug 13 01:04:24.474000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.443937 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 01:04:24.443966 systemd[1]: Stopped ignition-setup.service. Aug 13 01:04:24.444183 systemd[1]: Stopping iscsiuio.service... Aug 13 01:04:24.478000 audit: BPF prog-id=6 op=UNLOAD Aug 13 01:04:24.445106 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 01:04:24.447016 systemd[1]: iscsiuio.service: Deactivated successfully. Aug 13 01:04:24.447088 systemd[1]: Stopped iscsiuio.service. Aug 13 01:04:24.448084 systemd[1]: Stopped target network.target. Aug 13 01:04:24.449989 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 01:04:24.482000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.450019 systemd[1]: Closed iscsiuio.socket. Aug 13 01:04:24.451510 systemd[1]: Stopping systemd-networkd.service... Aug 13 01:04:24.453146 systemd[1]: Stopping systemd-resolved.service... Aug 13 01:04:24.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.458268 systemd-networkd[710]: eth0: DHCPv6 lease lost Aug 13 01:04:24.488000 audit: BPF prog-id=9 op=UNLOAD Aug 13 01:04:24.489000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.459614 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 01:04:24.491000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.459734 systemd[1]: Stopped systemd-networkd.service. Aug 13 01:04:24.463474 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 01:04:24.463501 systemd[1]: Closed systemd-networkd.socket. Aug 13 01:04:24.465616 systemd[1]: Stopping network-cleanup.service... Aug 13 01:04:24.494000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.498000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.466489 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 01:04:24.466531 systemd[1]: Stopped parse-ip-for-networkd.service. Aug 13 01:04:24.468088 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 01:04:24.468120 systemd[1]: Stopped systemd-sysctl.service. Aug 13 01:04:24.503000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.503000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.468949 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 01:04:24.468981 systemd[1]: Stopped systemd-modules-load.service. Aug 13 01:04:24.470760 systemd[1]: Stopping systemd-udevd.service... Aug 13 01:04:24.472255 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Aug 13 01:04:24.472608 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 01:04:24.472699 systemd[1]: Stopped systemd-resolved.service. Aug 13 01:04:24.481120 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 01:04:24.481228 systemd[1]: Stopped systemd-udevd.service. Aug 13 01:04:24.482633 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 01:04:24.482674 systemd[1]: Closed systemd-udevd-control.socket. Aug 13 01:04:24.484387 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 01:04:24.484421 systemd[1]: Closed systemd-udevd-kernel.socket. Aug 13 01:04:24.485823 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 01:04:24.485856 systemd[1]: Stopped dracut-pre-udev.service. Aug 13 01:04:24.487413 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 01:04:24.487450 systemd[1]: Stopped dracut-cmdline.service. Aug 13 01:04:24.489560 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 01:04:24.489595 systemd[1]: Stopped dracut-cmdline-ask.service. Aug 13 01:04:24.492094 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Aug 13 01:04:24.493170 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 01:04:24.493211 systemd[1]: Stopped systemd-vconsole-setup.service. Aug 13 01:04:24.496585 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 01:04:24.496782 systemd[1]: Stopped network-cleanup.service. Aug 13 01:04:24.501817 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 01:04:24.501960 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Aug 13 01:04:24.795128 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 01:04:24.795231 systemd[1]: Stopped sysroot-boot.service. Aug 13 01:04:24.796000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.797096 systemd[1]: Reached target initrd-switch-root.target. Aug 13 01:04:24.799000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:24.798555 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 01:04:24.798594 systemd[1]: Stopped initrd-setup-root.service. Aug 13 01:04:24.800064 systemd[1]: Starting initrd-switch-root.service... Aug 13 01:04:24.815606 systemd[1]: Switching root. Aug 13 01:04:24.835078 systemd-journald[197]: Journal stopped Aug 13 01:04:29.891565 systemd-journald[197]: Received SIGTERM from PID 1 (systemd). Aug 13 01:04:29.891610 kernel: SELinux: Class mctp_socket not defined in policy. Aug 13 01:04:29.891627 kernel: SELinux: Class anon_inode not defined in policy. Aug 13 01:04:29.891643 kernel: SELinux: the above unknown classes and permissions will be allowed Aug 13 01:04:29.891653 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 01:04:29.891663 kernel: SELinux: policy capability open_perms=1 Aug 13 01:04:29.891672 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 01:04:29.891682 kernel: SELinux: policy capability always_check_network=0 Aug 13 01:04:29.891691 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 01:04:29.891701 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 01:04:29.891710 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 01:04:29.891724 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 01:04:29.891735 systemd[1]: Successfully loaded SELinux policy in 44.685ms. Aug 13 01:04:29.891755 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.469ms. Aug 13 01:04:29.891767 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Aug 13 01:04:29.891778 systemd[1]: Detected virtualization kvm. Aug 13 01:04:29.891796 systemd[1]: Detected architecture x86-64. Aug 13 01:04:29.891806 systemd[1]: Detected first boot. Aug 13 01:04:29.891816 systemd[1]: Initializing machine ID from VM UUID. Aug 13 01:04:29.891831 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Aug 13 01:04:29.891844 systemd[1]: Populated /etc with preset unit settings. Aug 13 01:04:29.891858 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 01:04:29.891871 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 01:04:29.891883 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 01:04:29.891899 kernel: kauditd_printk_skb: 47 callbacks suppressed Aug 13 01:04:29.891909 kernel: audit: type=1334 audit(1755047069.649:86): prog-id=12 op=LOAD Aug 13 01:04:29.891918 kernel: audit: type=1334 audit(1755047069.649:87): prog-id=3 op=UNLOAD Aug 13 01:04:29.891929 kernel: audit: type=1334 audit(1755047069.651:88): prog-id=13 op=LOAD Aug 13 01:04:29.891939 kernel: audit: type=1334 audit(1755047069.652:89): prog-id=14 op=LOAD Aug 13 01:04:29.891949 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 01:04:29.891959 kernel: audit: type=1334 audit(1755047069.652:90): prog-id=4 op=UNLOAD Aug 13 01:04:29.891968 kernel: audit: type=1334 audit(1755047069.652:91): prog-id=5 op=UNLOAD Aug 13 01:04:29.891982 kernel: audit: type=1131 audit(1755047069.653:92): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.891992 systemd[1]: Stopped initrd-switch-root.service. Aug 13 01:04:29.892003 kernel: audit: type=1130 audit(1755047069.663:93): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.892013 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 01:04:29.892023 kernel: audit: type=1131 audit(1755047069.663:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.892041 systemd[1]: Created slice system-addon\x2dconfig.slice. Aug 13 01:04:29.892052 systemd[1]: Created slice system-addon\x2drun.slice. Aug 13 01:04:29.892062 kernel: audit: type=1334 audit(1755047069.674:95): prog-id=12 op=UNLOAD Aug 13 01:04:29.892076 systemd[1]: Created slice system-getty.slice. Aug 13 01:04:29.892086 systemd[1]: Created slice system-modprobe.slice. Aug 13 01:04:29.892097 systemd[1]: Created slice system-serial\x2dgetty.slice. Aug 13 01:04:29.892107 systemd[1]: Created slice system-system\x2dcloudinit.slice. Aug 13 01:04:29.892118 systemd[1]: Created slice system-systemd\x2dfsck.slice. Aug 13 01:04:29.892128 systemd[1]: Created slice user.slice. Aug 13 01:04:29.892138 systemd[1]: Started systemd-ask-password-console.path. Aug 13 01:04:29.892148 systemd[1]: Started systemd-ask-password-wall.path. Aug 13 01:04:29.892163 systemd[1]: Set up automount boot.automount. Aug 13 01:04:29.892174 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Aug 13 01:04:29.892184 systemd[1]: Stopped target initrd-switch-root.target. Aug 13 01:04:29.892194 systemd[1]: Stopped target initrd-fs.target. Aug 13 01:04:29.892205 systemd[1]: Stopped target initrd-root-fs.target. Aug 13 01:04:29.892215 systemd[1]: Reached target integritysetup.target. Aug 13 01:04:29.892226 systemd[1]: Reached target remote-cryptsetup.target. Aug 13 01:04:29.892247 systemd[1]: Reached target remote-fs.target. Aug 13 01:04:29.892258 systemd[1]: Reached target slices.target. Aug 13 01:04:29.892274 systemd[1]: Reached target swap.target. Aug 13 01:04:29.892297 systemd[1]: Reached target torcx.target. Aug 13 01:04:29.892307 systemd[1]: Reached target veritysetup.target. Aug 13 01:04:29.892318 systemd[1]: Listening on systemd-coredump.socket. Aug 13 01:04:29.892328 systemd[1]: Listening on systemd-initctl.socket. Aug 13 01:04:29.892338 systemd[1]: Listening on systemd-networkd.socket. Aug 13 01:04:29.892348 systemd[1]: Listening on systemd-udevd-control.socket. Aug 13 01:04:29.892358 systemd[1]: Listening on systemd-udevd-kernel.socket. Aug 13 01:04:29.892368 systemd[1]: Listening on systemd-userdbd.socket. Aug 13 01:04:29.892378 systemd[1]: Mounting dev-hugepages.mount... Aug 13 01:04:29.892393 systemd[1]: Mounting dev-mqueue.mount... Aug 13 01:04:29.892403 systemd[1]: Mounting media.mount... Aug 13 01:04:29.892413 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:29.892423 systemd[1]: Mounting sys-kernel-debug.mount... Aug 13 01:04:29.892433 systemd[1]: Mounting sys-kernel-tracing.mount... Aug 13 01:04:29.892443 systemd[1]: Mounting tmp.mount... Aug 13 01:04:29.892454 systemd[1]: Starting flatcar-tmpfiles.service... Aug 13 01:04:29.892464 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 01:04:29.892479 systemd[1]: Starting kmod-static-nodes.service... Aug 13 01:04:29.892489 systemd[1]: Starting modprobe@configfs.service... Aug 13 01:04:29.892509 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 01:04:29.892520 systemd[1]: Starting modprobe@drm.service... Aug 13 01:04:29.892538 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 01:04:29.892548 systemd[1]: Starting modprobe@fuse.service... Aug 13 01:04:29.892558 systemd[1]: Starting modprobe@loop.service... Aug 13 01:04:29.892568 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 01:04:29.892578 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 01:04:29.892594 systemd[1]: Stopped systemd-fsck-root.service. Aug 13 01:04:29.892604 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 01:04:29.892619 kernel: fuse: init (API version 7.34) Aug 13 01:04:29.892628 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 01:04:29.892639 systemd[1]: Stopped systemd-journald.service. Aug 13 01:04:29.892649 kernel: loop: module loaded Aug 13 01:04:29.892659 systemd[1]: Starting systemd-journald.service... Aug 13 01:04:29.892669 systemd[1]: Starting systemd-modules-load.service... Aug 13 01:04:29.892679 systemd[1]: Starting systemd-network-generator.service... Aug 13 01:04:29.892690 systemd[1]: Starting systemd-remount-fs.service... Aug 13 01:04:29.892704 systemd[1]: Starting systemd-udev-trigger.service... Aug 13 01:04:29.892714 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 01:04:29.892725 systemd[1]: Stopped verity-setup.service. Aug 13 01:04:29.892735 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:29.892748 systemd-journald[990]: Journal started Aug 13 01:04:29.892794 systemd-journald[990]: Runtime Journal (/run/log/journal/723d83f0fa2746858b59d10f233075a2) is 6.0M, max 48.5M, 42.5M free. Aug 13 01:04:24.905000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 01:04:26.050000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Aug 13 01:04:26.050000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Aug 13 01:04:26.050000 audit: BPF prog-id=10 op=LOAD Aug 13 01:04:26.050000 audit: BPF prog-id=10 op=UNLOAD Aug 13 01:04:26.051000 audit: BPF prog-id=11 op=LOAD Aug 13 01:04:26.051000 audit: BPF prog-id=11 op=UNLOAD Aug 13 01:04:26.083000 audit[909]: AVC avc: denied { associate } for pid=909 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Aug 13 01:04:26.083000 audit[909]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0001825d2 a1=c000186708 a2=c000190a00 a3=32 items=0 ppid=892 pid=909 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 01:04:26.083000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Aug 13 01:04:26.085000 audit[909]: AVC avc: denied { associate } for pid=909 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Aug 13 01:04:26.085000 audit[909]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0001826b9 a2=1ed a3=0 items=2 ppid=892 pid=909 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 01:04:26.085000 audit: CWD cwd="/" Aug 13 01:04:26.085000 audit: PATH item=0 name=(null) inode=2 dev=00:1b mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:26.085000 audit: PATH item=1 name=(null) inode=3 dev=00:1b mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:26.085000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Aug 13 01:04:29.649000 audit: BPF prog-id=12 op=LOAD Aug 13 01:04:29.649000 audit: BPF prog-id=3 op=UNLOAD Aug 13 01:04:29.651000 audit: BPF prog-id=13 op=LOAD Aug 13 01:04:29.652000 audit: BPF prog-id=14 op=LOAD Aug 13 01:04:29.652000 audit: BPF prog-id=4 op=UNLOAD Aug 13 01:04:29.652000 audit: BPF prog-id=5 op=UNLOAD Aug 13 01:04:29.653000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.663000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.663000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.674000 audit: BPF prog-id=12 op=UNLOAD Aug 13 01:04:29.775000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.778000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.780000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.780000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.781000 audit: BPF prog-id=15 op=LOAD Aug 13 01:04:29.781000 audit: BPF prog-id=16 op=LOAD Aug 13 01:04:29.782000 audit: BPF prog-id=17 op=LOAD Aug 13 01:04:29.782000 audit: BPF prog-id=13 op=UNLOAD Aug 13 01:04:29.782000 audit: BPF prog-id=14 op=UNLOAD Aug 13 01:04:29.890000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Aug 13 01:04:29.890000 audit[990]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffdc435abf0 a2=4000 a3=7ffdc435ac8c items=0 ppid=1 pid=990 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 01:04:29.890000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Aug 13 01:04:29.891000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.647884 systemd[1]: Queued start job for default target multi-user.target. Aug 13 01:04:26.081701 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 01:04:29.647895 systemd[1]: Unnecessary job was removed for dev-vda6.device. Aug 13 01:04:26.081978 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Aug 13 01:04:29.653460 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 01:04:26.082000 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Aug 13 01:04:26.082040 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Aug 13 01:04:26.082052 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="skipped missing lower profile" missing profile=oem Aug 13 01:04:26.082086 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Aug 13 01:04:26.082097 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Aug 13 01:04:26.082339 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Aug 13 01:04:26.082388 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Aug 13 01:04:26.082400 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Aug 13 01:04:26.083184 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Aug 13 01:04:26.083256 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Aug 13 01:04:26.083289 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.8: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.8 Aug 13 01:04:26.083318 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Aug 13 01:04:26.083353 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.8: no such file or directory" path=/var/lib/torcx/store/3510.3.8 Aug 13 01:04:26.083378 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:26Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Aug 13 01:04:29.375200 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Aug 13 01:04:29.375469 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Aug 13 01:04:29.375601 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Aug 13 01:04:29.375776 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Aug 13 01:04:29.375821 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Aug 13 01:04:29.375892 /usr/lib/systemd/system-generators/torcx-generator[909]: time="2025-08-13T01:04:29Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Aug 13 01:04:29.896354 systemd[1]: Started systemd-journald.service. Aug 13 01:04:29.896000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.897109 systemd[1]: Mounted dev-hugepages.mount. Aug 13 01:04:29.898919 systemd[1]: Mounted dev-mqueue.mount. Aug 13 01:04:29.899742 systemd[1]: Mounted media.mount. Aug 13 01:04:29.900613 systemd[1]: Mounted sys-kernel-debug.mount. Aug 13 01:04:29.901636 systemd[1]: Mounted sys-kernel-tracing.mount. Aug 13 01:04:29.902550 systemd[1]: Mounted tmp.mount. Aug 13 01:04:29.903481 systemd[1]: Finished flatcar-tmpfiles.service. Aug 13 01:04:29.904697 systemd[1]: Finished kmod-static-nodes.service. Aug 13 01:04:29.904000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.905000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.905811 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 01:04:29.905924 systemd[1]: Finished modprobe@configfs.service. Aug 13 01:04:29.906000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.906000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.907378 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 01:04:29.907491 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 01:04:29.908000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.908000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.908685 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 01:04:29.908806 systemd[1]: Finished modprobe@drm.service. Aug 13 01:04:29.909000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.909000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.909955 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 01:04:29.910064 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 01:04:29.910000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.910000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.911206 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 01:04:29.911505 systemd[1]: Finished modprobe@fuse.service. Aug 13 01:04:29.912000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.912000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.912606 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 01:04:29.912797 systemd[1]: Finished modprobe@loop.service. Aug 13 01:04:29.913000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.913000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.914035 systemd[1]: Finished systemd-modules-load.service. Aug 13 01:04:29.914000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.915513 systemd[1]: Finished systemd-network-generator.service. Aug 13 01:04:29.916000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.916891 systemd[1]: Finished systemd-remount-fs.service. Aug 13 01:04:29.917000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.918337 systemd[1]: Reached target network-pre.target. Aug 13 01:04:29.920608 systemd[1]: Mounting sys-fs-fuse-connections.mount... Aug 13 01:04:29.922885 systemd[1]: Mounting sys-kernel-config.mount... Aug 13 01:04:29.923632 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 01:04:29.925675 systemd[1]: Starting systemd-hwdb-update.service... Aug 13 01:04:29.928150 systemd[1]: Starting systemd-journal-flush.service... Aug 13 01:04:29.929207 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 01:04:29.930557 systemd[1]: Starting systemd-random-seed.service... Aug 13 01:04:29.931577 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 01:04:29.932822 systemd[1]: Starting systemd-sysctl.service... Aug 13 01:04:29.933440 systemd-journald[990]: Time spent on flushing to /var/log/journal/723d83f0fa2746858b59d10f233075a2 is 13.887ms for 1106 entries. Aug 13 01:04:29.933440 systemd-journald[990]: System Journal (/var/log/journal/723d83f0fa2746858b59d10f233075a2) is 8.0M, max 195.6M, 187.6M free. Aug 13 01:04:30.300521 systemd-journald[990]: Received client request to flush runtime journal. Aug 13 01:04:29.973000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.196000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.200000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:29.937867 systemd[1]: Starting systemd-sysusers.service... Aug 13 01:04:29.942289 systemd[1]: Mounted sys-fs-fuse-connections.mount. Aug 13 01:04:29.943436 systemd[1]: Mounted sys-kernel-config.mount. Aug 13 01:04:30.301295 udevadm[1012]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Aug 13 01:04:29.972926 systemd[1]: Finished systemd-udev-trigger.service. Aug 13 01:04:29.974997 systemd[1]: Starting systemd-udev-settle.service... Aug 13 01:04:30.191007 systemd[1]: Finished systemd-random-seed.service. Aug 13 01:04:30.192330 systemd[1]: Reached target first-boot-complete.target. Aug 13 01:04:30.195788 systemd[1]: Finished systemd-sysctl.service. Aug 13 01:04:30.199544 systemd[1]: Finished systemd-sysusers.service. Aug 13 01:04:30.301729 systemd[1]: Finished systemd-journal-flush.service. Aug 13 01:04:30.302000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.921826 systemd[1]: Finished systemd-hwdb-update.service. Aug 13 01:04:30.922000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.923000 audit: BPF prog-id=18 op=LOAD Aug 13 01:04:30.923000 audit: BPF prog-id=19 op=LOAD Aug 13 01:04:30.923000 audit: BPF prog-id=7 op=UNLOAD Aug 13 01:04:30.923000 audit: BPF prog-id=8 op=UNLOAD Aug 13 01:04:30.924508 systemd[1]: Starting systemd-udevd.service... Aug 13 01:04:30.942566 systemd-udevd[1015]: Using default interface naming scheme 'v252'. Aug 13 01:04:30.956503 systemd[1]: Started systemd-udevd.service. Aug 13 01:04:30.957000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:30.958000 audit: BPF prog-id=20 op=LOAD Aug 13 01:04:30.959449 systemd[1]: Starting systemd-networkd.service... Aug 13 01:04:30.963000 audit: BPF prog-id=21 op=LOAD Aug 13 01:04:30.963000 audit: BPF prog-id=22 op=LOAD Aug 13 01:04:30.963000 audit: BPF prog-id=23 op=LOAD Aug 13 01:04:30.964302 systemd[1]: Starting systemd-userdbd.service... Aug 13 01:04:30.996090 systemd[1]: Started systemd-userdbd.service. Aug 13 01:04:30.996000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.014125 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Aug 13 01:04:31.027860 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Aug 13 01:04:31.046276 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Aug 13 01:04:31.048164 systemd-networkd[1020]: lo: Link UP Aug 13 01:04:31.048180 systemd-networkd[1020]: lo: Gained carrier Aug 13 01:04:31.048785 systemd-networkd[1020]: Enumeration completed Aug 13 01:04:31.048906 systemd[1]: Started systemd-networkd.service. Aug 13 01:04:31.048918 systemd-networkd[1020]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 01:04:31.049000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.051172 systemd-networkd[1020]: eth0: Link UP Aug 13 01:04:31.051185 systemd-networkd[1020]: eth0: Gained carrier Aug 13 01:04:31.053306 kernel: ACPI: button: Power Button [PWRF] Aug 13 01:04:31.064444 systemd-networkd[1020]: eth0: DHCPv4 address 10.0.0.132/16, gateway 10.0.0.1 acquired from 10.0.0.1 Aug 13 01:04:31.066000 audit[1030]: AVC avc: denied { confidentiality } for pid=1030 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Aug 13 01:04:31.066000 audit[1030]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=55677610e5d0 a1=338ac a2=7f74de478bc5 a3=5 items=110 ppid=1015 pid=1030 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 01:04:31.066000 audit: CWD cwd="/" Aug 13 01:04:31.066000 audit: PATH item=0 name=(null) inode=44 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=1 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=2 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=3 name=(null) inode=14649 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=4 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=5 name=(null) inode=14650 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=6 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=7 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=8 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=9 name=(null) inode=14652 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=10 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=11 name=(null) inode=14653 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=12 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=13 name=(null) inode=14654 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=14 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=15 name=(null) inode=14655 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=16 name=(null) inode=14651 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=17 name=(null) inode=14656 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=18 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=19 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=20 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=21 name=(null) inode=14658 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=22 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=23 name=(null) inode=14659 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=24 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=25 name=(null) inode=14660 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=26 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=27 name=(null) inode=14661 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=28 name=(null) inode=14657 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=29 name=(null) inode=14662 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=30 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=31 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=32 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=33 name=(null) inode=14664 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=34 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=35 name=(null) inode=14665 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=36 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=37 name=(null) inode=14666 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=38 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=39 name=(null) inode=14667 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=40 name=(null) inode=14663 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=41 name=(null) inode=14668 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=42 name=(null) inode=14648 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=43 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=44 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=45 name=(null) inode=14670 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=46 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=47 name=(null) inode=14671 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=48 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=49 name=(null) inode=14672 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=50 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=51 name=(null) inode=14673 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=52 name=(null) inode=14669 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=53 name=(null) inode=14674 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=54 name=(null) inode=44 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=55 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=56 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=57 name=(null) inode=14676 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=58 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=59 name=(null) inode=14677 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=60 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=61 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=62 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=63 name=(null) inode=14679 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=64 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=65 name=(null) inode=14680 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=66 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=67 name=(null) inode=14681 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=68 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=69 name=(null) inode=14682 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=70 name=(null) inode=14678 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=71 name=(null) inode=14683 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=72 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=73 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=74 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=75 name=(null) inode=14685 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=76 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=77 name=(null) inode=14686 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=78 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=79 name=(null) inode=14687 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=80 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=81 name=(null) inode=14688 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=82 name=(null) inode=14684 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=83 name=(null) inode=14689 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=84 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=85 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=86 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=87 name=(null) inode=14691 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=88 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=89 name=(null) inode=14692 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=90 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=91 name=(null) inode=14693 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=92 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=93 name=(null) inode=14694 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=94 name=(null) inode=14690 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=95 name=(null) inode=14695 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=96 name=(null) inode=14675 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=97 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=98 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=99 name=(null) inode=14697 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=100 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=101 name=(null) inode=14698 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=102 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=103 name=(null) inode=14699 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=104 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=105 name=(null) inode=14700 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=106 name=(null) inode=14696 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=107 name=(null) inode=14701 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PATH item=109 name=(null) inode=11948 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 01:04:31.066000 audit: PROCTITLE proctitle="(udev-worker)" Aug 13 01:04:31.082383 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Aug 13 01:04:31.084262 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Aug 13 01:04:31.090037 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Aug 13 01:04:31.090178 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Aug 13 01:04:31.118264 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 01:04:31.162283 kernel: kvm: Nested Virtualization enabled Aug 13 01:04:31.162456 kernel: SVM: kvm: Nested Paging enabled Aug 13 01:04:31.162494 kernel: SVM: Virtual VMLOAD VMSAVE supported Aug 13 01:04:31.162525 kernel: SVM: Virtual GIF supported Aug 13 01:04:31.183271 kernel: EDAC MC: Ver: 3.0.0 Aug 13 01:04:31.209658 systemd[1]: Finished systemd-udev-settle.service. Aug 13 01:04:31.210000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.211820 systemd[1]: Starting lvm2-activation-early.service... Aug 13 01:04:31.219775 lvm[1050]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 01:04:31.250150 systemd[1]: Finished lvm2-activation-early.service. Aug 13 01:04:31.250000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.251129 systemd[1]: Reached target cryptsetup.target. Aug 13 01:04:31.252794 systemd[1]: Starting lvm2-activation.service... Aug 13 01:04:31.256472 lvm[1051]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 01:04:31.285982 systemd[1]: Finished lvm2-activation.service. Aug 13 01:04:31.286000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.286884 systemd[1]: Reached target local-fs-pre.target. Aug 13 01:04:31.287708 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 01:04:31.287731 systemd[1]: Reached target local-fs.target. Aug 13 01:04:31.288510 systemd[1]: Reached target machines.target. Aug 13 01:04:31.290193 systemd[1]: Starting ldconfig.service... Aug 13 01:04:31.291110 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.291163 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:31.292115 systemd[1]: Starting systemd-boot-update.service... Aug 13 01:04:31.293931 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Aug 13 01:04:31.295910 systemd[1]: Starting systemd-machine-id-commit.service... Aug 13 01:04:31.298087 systemd[1]: Starting systemd-sysext.service... Aug 13 01:04:31.299162 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1053 (bootctl) Aug 13 01:04:31.300076 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Aug 13 01:04:31.314323 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Aug 13 01:04:31.314000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.316050 systemd[1]: Unmounting usr-share-oem.mount... Aug 13 01:04:31.321597 systemd[1]: usr-share-oem.mount: Deactivated successfully. Aug 13 01:04:31.321740 systemd[1]: Unmounted usr-share-oem.mount. Aug 13 01:04:31.336268 kernel: loop0: detected capacity change from 0 to 221472 Aug 13 01:04:31.353599 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 01:04:31.354322 systemd[1]: Finished systemd-machine-id-commit.service. Aug 13 01:04:31.353000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.358273 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 01:04:31.374840 systemd-fsck[1061]: fsck.fat 4.2 (2021-01-31) Aug 13 01:04:31.374840 systemd-fsck[1061]: /dev/vda1: 789 files, 119324/258078 clusters Aug 13 01:04:31.376275 kernel: loop1: detected capacity change from 0 to 221472 Aug 13 01:04:31.376735 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Aug 13 01:04:31.378000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.379822 systemd[1]: Mounting boot.mount... Aug 13 01:04:31.384709 (sd-sysext)[1066]: Using extensions 'kubernetes'. Aug 13 01:04:31.385645 (sd-sysext)[1066]: Merged extensions into '/usr'. Aug 13 01:04:31.394685 systemd[1]: Mounted boot.mount. Aug 13 01:04:31.398623 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:31.400221 systemd[1]: Mounting usr-share-oem.mount... Aug 13 01:04:31.401170 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.402270 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 01:04:31.404041 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 01:04:31.406052 systemd[1]: Starting modprobe@loop.service... Aug 13 01:04:31.407125 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.407225 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:31.411000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.407335 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:31.409656 systemd[1]: Mounted usr-share-oem.mount. Aug 13 01:04:31.410956 systemd[1]: Finished systemd-boot-update.service. Aug 13 01:04:31.412349 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 01:04:31.412473 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 01:04:31.413000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.413000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.413879 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 01:04:31.413980 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 01:04:31.415000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.415000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.415602 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 01:04:31.415700 systemd[1]: Finished modprobe@loop.service. Aug 13 01:04:31.416000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.416000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.417793 systemd[1]: Finished systemd-sysext.service. Aug 13 01:04:31.418000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.420859 systemd[1]: Starting ensure-sysext.service... Aug 13 01:04:31.421768 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 01:04:31.421817 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.422824 systemd[1]: Starting systemd-tmpfiles-setup.service... Aug 13 01:04:31.426939 systemd[1]: Reloading. Aug 13 01:04:31.441415 systemd-tmpfiles[1074]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Aug 13 01:04:31.443147 systemd-tmpfiles[1074]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 01:04:31.447208 systemd-tmpfiles[1074]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 01:04:31.540185 ldconfig[1052]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 01:04:31.633550 /usr/lib/systemd/system-generators/torcx-generator[1129]: time="2025-08-13T01:04:31Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 01:04:31.633583 /usr/lib/systemd/system-generators/torcx-generator[1129]: time="2025-08-13T01:04:31Z" level=info msg="torcx already run" Aug 13 01:04:31.662915 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 01:04:31.662938 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 01:04:31.682457 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 01:04:31.735000 audit: BPF prog-id=24 op=LOAD Aug 13 01:04:31.735000 audit: BPF prog-id=20 op=UNLOAD Aug 13 01:04:31.737000 audit: BPF prog-id=25 op=LOAD Aug 13 01:04:31.737000 audit: BPF prog-id=15 op=UNLOAD Aug 13 01:04:31.738000 audit: BPF prog-id=26 op=LOAD Aug 13 01:04:31.738000 audit: BPF prog-id=27 op=LOAD Aug 13 01:04:31.738000 audit: BPF prog-id=16 op=UNLOAD Aug 13 01:04:31.738000 audit: BPF prog-id=17 op=UNLOAD Aug 13 01:04:31.738000 audit: BPF prog-id=28 op=LOAD Aug 13 01:04:31.738000 audit: BPF prog-id=29 op=LOAD Aug 13 01:04:31.738000 audit: BPF prog-id=18 op=UNLOAD Aug 13 01:04:31.738000 audit: BPF prog-id=19 op=UNLOAD Aug 13 01:04:31.739000 audit: BPF prog-id=30 op=LOAD Aug 13 01:04:31.739000 audit: BPF prog-id=21 op=UNLOAD Aug 13 01:04:31.739000 audit: BPF prog-id=31 op=LOAD Aug 13 01:04:31.739000 audit: BPF prog-id=32 op=LOAD Aug 13 01:04:31.739000 audit: BPF prog-id=22 op=UNLOAD Aug 13 01:04:31.739000 audit: BPF prog-id=23 op=UNLOAD Aug 13 01:04:31.742002 systemd[1]: Finished ldconfig.service. Aug 13 01:04:31.742000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.743827 systemd[1]: Finished systemd-tmpfiles-setup.service. Aug 13 01:04:31.744000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.747059 systemd[1]: Starting audit-rules.service... Aug 13 01:04:31.748630 systemd[1]: Starting clean-ca-certificates.service... Aug 13 01:04:31.752000 audit: BPF prog-id=33 op=LOAD Aug 13 01:04:31.750629 systemd[1]: Starting systemd-journal-catalog-update.service... Aug 13 01:04:31.753078 systemd[1]: Starting systemd-resolved.service... Aug 13 01:04:31.754000 audit: BPF prog-id=34 op=LOAD Aug 13 01:04:31.755482 systemd[1]: Starting systemd-timesyncd.service... Aug 13 01:04:31.757167 systemd[1]: Starting systemd-update-utmp.service... Aug 13 01:04:31.758478 systemd[1]: Finished clean-ca-certificates.service. Aug 13 01:04:31.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.761621 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 01:04:31.761000 audit[1147]: SYSTEM_BOOT pid=1147 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.767063 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.768622 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 01:04:31.770846 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 01:04:31.772738 systemd[1]: Starting modprobe@loop.service... Aug 13 01:04:31.773687 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.773853 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:31.773987 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 01:04:31.775098 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 01:04:31.775277 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 01:04:31.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.776000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.776759 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 01:04:31.776892 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 01:04:31.777000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.777000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.778299 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 01:04:31.778414 systemd[1]: Finished modprobe@loop.service. Aug 13 01:04:31.779000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.779000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.779972 systemd[1]: Finished systemd-journal-catalog-update.service. Aug 13 01:04:31.780000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.781469 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 01:04:31.781641 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.783374 systemd[1]: Starting systemd-update-done.service... Aug 13 01:04:31.784876 systemd[1]: Finished systemd-update-utmp.service. Aug 13 01:04:31.785000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.788030 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.789080 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 01:04:31.791008 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 01:04:31.792866 systemd[1]: Starting modprobe@loop.service... Aug 13 01:04:31.793682 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.793806 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:31.793920 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 01:04:31.795000 systemd[1]: Finished systemd-update-done.service. Aug 13 01:04:31.797000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 01:04:31.797648 augenrules[1160]: No rules Aug 13 01:04:31.797644 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 01:04:31.797822 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 01:04:31.797000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Aug 13 01:04:31.797000 audit[1160]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffcdc060600 a2=420 a3=0 items=0 ppid=1136 pid=1160 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 01:04:31.797000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Aug 13 01:04:31.799197 systemd[1]: Finished audit-rules.service. Aug 13 01:04:31.800384 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 01:04:31.800527 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 01:04:31.801691 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 01:04:31.801788 systemd[1]: Finished modprobe@loop.service. Aug 13 01:04:31.805014 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.806107 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 01:04:31.808135 systemd[1]: Starting modprobe@drm.service... Aug 13 01:04:31.810304 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 01:04:31.812456 systemd[1]: Starting modprobe@loop.service... Aug 13 01:04:31.813251 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.813350 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:31.814614 systemd[1]: Starting systemd-networkd-wait-online.service... Aug 13 01:04:31.815605 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 01:04:31.817177 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 01:04:31.817466 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 01:04:31.818746 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 01:04:31.818885 systemd[1]: Finished modprobe@drm.service. Aug 13 01:04:31.820114 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 01:04:31.820624 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 01:04:31.822016 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 01:04:31.822138 systemd[1]: Finished modprobe@loop.service. Aug 13 01:04:31.823838 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 01:04:31.823983 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 01:04:31.825305 systemd[1]: Finished ensure-sysext.service. Aug 13 01:04:31.832879 systemd-resolved[1142]: Positive Trust Anchors: Aug 13 01:04:31.832888 systemd-resolved[1142]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 01:04:31.832915 systemd-resolved[1142]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Aug 13 01:04:31.834132 systemd[1]: Started systemd-timesyncd.service. Aug 13 01:04:31.835318 systemd[1]: Reached target time-set.target. Aug 13 01:04:32.940869 systemd-timesyncd[1143]: Contacted time server 10.0.0.1:123 (10.0.0.1). Aug 13 01:04:32.940932 systemd-timesyncd[1143]: Initial clock synchronization to Wed 2025-08-13 01:04:32.940779 UTC. Aug 13 01:04:32.946674 systemd-resolved[1142]: Defaulting to hostname 'linux'. Aug 13 01:04:32.948179 systemd[1]: Started systemd-resolved.service. Aug 13 01:04:32.949181 systemd[1]: Reached target network.target. Aug 13 01:04:32.950037 systemd[1]: Reached target nss-lookup.target. Aug 13 01:04:32.950913 systemd[1]: Reached target sysinit.target. Aug 13 01:04:32.951801 systemd[1]: Started motdgen.path. Aug 13 01:04:32.952568 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Aug 13 01:04:32.953825 systemd[1]: Started logrotate.timer. Aug 13 01:04:32.954721 systemd[1]: Started mdadm.timer. Aug 13 01:04:32.955453 systemd[1]: Started systemd-tmpfiles-clean.timer. Aug 13 01:04:32.956410 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 01:04:32.956435 systemd[1]: Reached target paths.target. Aug 13 01:04:32.957217 systemd[1]: Reached target timers.target. Aug 13 01:04:32.958328 systemd[1]: Listening on dbus.socket. Aug 13 01:04:32.960015 systemd[1]: Starting docker.socket... Aug 13 01:04:32.962896 systemd[1]: Listening on sshd.socket. Aug 13 01:04:32.963823 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:32.964192 systemd[1]: Listening on docker.socket. Aug 13 01:04:32.965109 systemd[1]: Reached target sockets.target. Aug 13 01:04:32.965973 systemd[1]: Reached target basic.target. Aug 13 01:04:32.966847 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Aug 13 01:04:32.966875 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Aug 13 01:04:32.967717 systemd[1]: Starting containerd.service... Aug 13 01:04:32.969388 systemd[1]: Starting dbus.service... Aug 13 01:04:32.971210 systemd[1]: Starting enable-oem-cloudinit.service... Aug 13 01:04:32.973453 systemd[1]: Starting extend-filesystems.service... Aug 13 01:04:32.974614 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Aug 13 01:04:32.976377 jq[1178]: false Aug 13 01:04:32.975755 systemd[1]: Starting motdgen.service... Aug 13 01:04:32.977925 systemd[1]: Starting prepare-helm.service... Aug 13 01:04:32.981327 systemd[1]: Starting ssh-key-proc-cmdline.service... Aug 13 01:04:32.983671 systemd[1]: Starting sshd-keygen.service... Aug 13 01:04:32.988365 extend-filesystems[1179]: Found loop1 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found sr0 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda1 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda2 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda3 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found usr Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda4 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda6 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda7 Aug 13 01:04:32.988365 extend-filesystems[1179]: Found vda9 Aug 13 01:04:32.988365 extend-filesystems[1179]: Checking size of /dev/vda9 Aug 13 01:04:33.007017 extend-filesystems[1179]: Resized partition /dev/vda9 Aug 13 01:04:32.996625 dbus-daemon[1177]: [system] SELinux support is enabled Aug 13 01:04:32.990625 systemd[1]: Starting systemd-logind.service... Aug 13 01:04:33.009475 extend-filesystems[1202]: resize2fs 1.46.5 (30-Dec-2021) Aug 13 01:04:33.058084 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Aug 13 01:04:32.993960 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 01:04:32.996695 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 01:04:33.000835 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 01:04:33.002530 systemd[1]: Starting update-engine.service... Aug 13 01:04:33.009421 systemd[1]: Starting update-ssh-keys-after-ignition.service... Aug 13 01:04:33.059317 systemd[1]: Started dbus.service. Aug 13 01:04:33.063445 jq[1203]: true Aug 13 01:04:33.067581 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 01:04:33.067737 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Aug 13 01:04:33.068006 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 01:04:33.068144 systemd[1]: Finished motdgen.service. Aug 13 01:04:33.070352 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 01:04:33.070564 systemd[1]: Finished ssh-key-proc-cmdline.service. Aug 13 01:04:33.082924 systemd-logind[1195]: Watching system buttons on /dev/input/event1 (Power Button) Aug 13 01:04:33.082955 systemd-logind[1195]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Aug 13 01:04:33.084194 tar[1205]: linux-amd64/helm Aug 13 01:04:33.084510 systemd-logind[1195]: New seat seat0. Aug 13 01:04:33.085282 systemd[1]: Created slice system-sshd.slice. Aug 13 01:04:33.086666 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 01:04:33.086690 systemd[1]: Reached target system-config.target. Aug 13 01:04:33.087819 jq[1206]: true Aug 13 01:04:33.088365 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 01:04:33.088382 systemd[1]: Reached target user-config.target. Aug 13 01:04:33.092674 systemd[1]: Started systemd-logind.service. Aug 13 01:04:33.094301 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Aug 13 01:04:33.125014 extend-filesystems[1202]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Aug 13 01:04:33.125014 extend-filesystems[1202]: old_desc_blocks = 1, new_desc_blocks = 1 Aug 13 01:04:33.125014 extend-filesystems[1202]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Aug 13 01:04:33.126123 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 01:04:33.295280 update_engine[1200]: I0813 01:04:33.164088 1200 main.cc:92] Flatcar Update Engine starting Aug 13 01:04:33.295280 update_engine[1200]: I0813 01:04:33.166510 1200 update_check_scheduler.cc:74] Next update check in 4m4s Aug 13 01:04:33.295493 extend-filesystems[1179]: Resized filesystem in /dev/vda9 Aug 13 01:04:33.126450 systemd[1]: Finished extend-filesystems.service. Aug 13 01:04:33.289715 systemd[1]: Started update-engine.service. Aug 13 01:04:33.292746 systemd[1]: Started locksmithd.service. Aug 13 01:04:33.323736 env[1207]: time="2025-08-13T01:04:33.323676026Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Aug 13 01:04:33.345224 env[1207]: time="2025-08-13T01:04:33.345182099Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 01:04:33.345624 env[1207]: time="2025-08-13T01:04:33.345602918Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.347362 env[1207]: time="2025-08-13T01:04:33.347337541Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.189-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 01:04:33.347442 env[1207]: time="2025-08-13T01:04:33.347422330Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.347716 env[1207]: time="2025-08-13T01:04:33.347695131Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 01:04:33.347812 env[1207]: time="2025-08-13T01:04:33.347791983Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.347901 env[1207]: time="2025-08-13T01:04:33.347879527Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Aug 13 01:04:33.347983 env[1207]: time="2025-08-13T01:04:33.347963555Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.348135 env[1207]: time="2025-08-13T01:04:33.348115680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.348502 env[1207]: time="2025-08-13T01:04:33.348483449Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 01:04:33.348691 env[1207]: time="2025-08-13T01:04:33.348669108Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 01:04:33.348769 env[1207]: time="2025-08-13T01:04:33.348749438Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 01:04:33.348900 env[1207]: time="2025-08-13T01:04:33.348879612Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Aug 13 01:04:33.348995 env[1207]: time="2025-08-13T01:04:33.348974370Z" level=info msg="metadata content store policy set" policy=shared Aug 13 01:04:33.427101 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:33.427199 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 01:04:33.519684 locksmithd[1230]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 01:04:33.650305 tar[1205]: linux-amd64/LICENSE Aug 13 01:04:33.650450 tar[1205]: linux-amd64/README.md Aug 13 01:04:33.654982 systemd[1]: Finished prepare-helm.service. Aug 13 01:04:33.714045 env[1207]: time="2025-08-13T01:04:33.713966099Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 01:04:33.714045 env[1207]: time="2025-08-13T01:04:33.714046350Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 01:04:33.714045 env[1207]: time="2025-08-13T01:04:33.714059785Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714144834Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714176634Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714202162Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714215868Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714232509Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714246776Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714290277Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714305345Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714320814Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 01:04:33.714498 env[1207]: time="2025-08-13T01:04:33.714470996Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 01:04:33.714684 env[1207]: time="2025-08-13T01:04:33.714563730Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 01:04:33.715050 env[1207]: time="2025-08-13T01:04:33.714971705Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 01:04:33.715155 env[1207]: time="2025-08-13T01:04:33.715090989Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715155 env[1207]: time="2025-08-13T01:04:33.715126936Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 01:04:33.715291 env[1207]: time="2025-08-13T01:04:33.715269994Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715327 env[1207]: time="2025-08-13T01:04:33.715290903Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715327 env[1207]: time="2025-08-13T01:04:33.715310440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715327 env[1207]: time="2025-08-13T01:04:33.715325468Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715393 env[1207]: time="2025-08-13T01:04:33.715338613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715393 env[1207]: time="2025-08-13T01:04:33.715354162Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715393 env[1207]: time="2025-08-13T01:04:33.715367096Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715393 env[1207]: time="2025-08-13T01:04:33.715377355Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715490 env[1207]: time="2025-08-13T01:04:33.715396381Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 01:04:33.715662 env[1207]: time="2025-08-13T01:04:33.715623737Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715662 env[1207]: time="2025-08-13T01:04:33.715651509Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715716 env[1207]: time="2025-08-13T01:04:33.715669654Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.715716 env[1207]: time="2025-08-13T01:04:33.715681546Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 01:04:33.715716 env[1207]: time="2025-08-13T01:04:33.715700842Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Aug 13 01:04:33.715716 env[1207]: time="2025-08-13T01:04:33.715714137Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 01:04:33.715806 env[1207]: time="2025-08-13T01:04:33.715751727Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Aug 13 01:04:33.715829 env[1207]: time="2025-08-13T01:04:33.715816649Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 01:04:33.716133 env[1207]: time="2025-08-13T01:04:33.716079873Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 01:04:33.717230 env[1207]: time="2025-08-13T01:04:33.716150766Z" level=info msg="Connect containerd service" Aug 13 01:04:33.717230 env[1207]: time="2025-08-13T01:04:33.716208855Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 01:04:33.717230 env[1207]: time="2025-08-13T01:04:33.716944784Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 01:04:33.717230 env[1207]: time="2025-08-13T01:04:33.717225000Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 01:04:33.717351 env[1207]: time="2025-08-13T01:04:33.717286595Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 01:04:33.717351 env[1207]: time="2025-08-13T01:04:33.717339925Z" level=info msg="containerd successfully booted in 0.401481s" Aug 13 01:04:33.717462 systemd[1]: Started containerd.service. Aug 13 01:04:33.719206 bash[1224]: Updated "/home/core/.ssh/authorized_keys" Aug 13 01:04:33.720083 systemd[1]: Finished update-ssh-keys-after-ignition.service. Aug 13 01:04:33.720529 env[1207]: time="2025-08-13T01:04:33.720385676Z" level=info msg="Start subscribing containerd event" Aug 13 01:04:33.720529 env[1207]: time="2025-08-13T01:04:33.720477628Z" level=info msg="Start recovering state" Aug 13 01:04:33.720809 env[1207]: time="2025-08-13T01:04:33.720590831Z" level=info msg="Start event monitor" Aug 13 01:04:33.720809 env[1207]: time="2025-08-13T01:04:33.720712168Z" level=info msg="Start snapshots syncer" Aug 13 01:04:33.721706 env[1207]: time="2025-08-13T01:04:33.720730903Z" level=info msg="Start cni network conf syncer for default" Aug 13 01:04:33.721706 env[1207]: time="2025-08-13T01:04:33.721689761Z" level=info msg="Start streaming server" Aug 13 01:04:33.862408 sshd_keygen[1197]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 01:04:33.882855 systemd[1]: Finished sshd-keygen.service. Aug 13 01:04:33.885285 systemd[1]: Starting issuegen.service... Aug 13 01:04:33.887715 systemd[1]: Started sshd@0-10.0.0.132:22-10.0.0.1:37326.service. Aug 13 01:04:33.890577 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 01:04:33.890714 systemd[1]: Finished issuegen.service. Aug 13 01:04:33.892822 systemd[1]: Starting systemd-user-sessions.service... Aug 13 01:04:33.899331 systemd[1]: Finished systemd-user-sessions.service. Aug 13 01:04:33.901500 systemd[1]: Started getty@tty1.service. Aug 13 01:04:33.903502 systemd[1]: Started serial-getty@ttyS0.service. Aug 13 01:04:33.904807 systemd[1]: Reached target getty.target. Aug 13 01:04:33.923724 sshd[1251]: Accepted publickey for core from 10.0.0.1 port 37326 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:33.925465 sshd[1251]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:33.934561 systemd-logind[1195]: New session 1 of user core. Aug 13 01:04:33.935485 systemd[1]: Created slice user-500.slice. Aug 13 01:04:33.937540 systemd[1]: Starting user-runtime-dir@500.service... Aug 13 01:04:33.946128 systemd[1]: Finished user-runtime-dir@500.service. Aug 13 01:04:33.948288 systemd[1]: Starting user@500.service... Aug 13 01:04:33.951513 (systemd)[1259]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:34.007548 systemd-networkd[1020]: eth0: Gained IPv6LL Aug 13 01:04:34.010054 systemd[1]: Finished systemd-networkd-wait-online.service. Aug 13 01:04:34.011822 systemd[1]: Reached target network-online.target. Aug 13 01:04:34.014991 systemd[1]: Starting kubelet.service... Aug 13 01:04:34.041088 systemd[1259]: Queued start job for default target default.target. Aug 13 01:04:34.041695 systemd[1259]: Reached target paths.target. Aug 13 01:04:34.041716 systemd[1259]: Reached target sockets.target. Aug 13 01:04:34.041727 systemd[1259]: Reached target timers.target. Aug 13 01:04:34.041739 systemd[1259]: Reached target basic.target. Aug 13 01:04:34.041849 systemd[1]: Started user@500.service. Aug 13 01:04:34.041974 systemd[1259]: Reached target default.target. Aug 13 01:04:34.042003 systemd[1259]: Startup finished in 84ms. Aug 13 01:04:34.043636 systemd[1]: Started session-1.scope. Aug 13 01:04:34.097708 systemd[1]: Started sshd@1-10.0.0.132:22-10.0.0.1:37328.service. Aug 13 01:04:34.131938 sshd[1270]: Accepted publickey for core from 10.0.0.1 port 37328 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:34.172511 sshd[1270]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:34.176796 systemd-logind[1195]: New session 2 of user core. Aug 13 01:04:34.177527 systemd[1]: Started session-2.scope. Aug 13 01:04:34.235103 sshd[1270]: pam_unix(sshd:session): session closed for user core Aug 13 01:04:34.238711 systemd[1]: sshd@1-10.0.0.132:22-10.0.0.1:37328.service: Deactivated successfully. Aug 13 01:04:34.239361 systemd[1]: session-2.scope: Deactivated successfully. Aug 13 01:04:34.239958 systemd-logind[1195]: Session 2 logged out. Waiting for processes to exit. Aug 13 01:04:34.241299 systemd[1]: Started sshd@2-10.0.0.132:22-10.0.0.1:37332.service. Aug 13 01:04:34.243716 systemd-logind[1195]: Removed session 2. Aug 13 01:04:34.270352 sshd[1276]: Accepted publickey for core from 10.0.0.1 port 37332 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:34.271823 sshd[1276]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:34.276209 systemd-logind[1195]: New session 3 of user core. Aug 13 01:04:34.276603 systemd[1]: Started session-3.scope. Aug 13 01:04:34.388063 sshd[1276]: pam_unix(sshd:session): session closed for user core Aug 13 01:04:34.390641 systemd[1]: sshd@2-10.0.0.132:22-10.0.0.1:37332.service: Deactivated successfully. Aug 13 01:04:34.391611 systemd-logind[1195]: Session 3 logged out. Waiting for processes to exit. Aug 13 01:04:34.391661 systemd[1]: session-3.scope: Deactivated successfully. Aug 13 01:04:34.392488 systemd-logind[1195]: Removed session 3. Aug 13 01:04:35.597476 systemd[1]: Started kubelet.service. Aug 13 01:04:35.599077 systemd[1]: Reached target multi-user.target. Aug 13 01:04:35.601546 systemd[1]: Starting systemd-update-utmp-runlevel.service... Aug 13 01:04:35.609354 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Aug 13 01:04:35.609533 systemd[1]: Finished systemd-update-utmp-runlevel.service. Aug 13 01:04:35.610735 systemd[1]: Startup finished in 748ms (kernel) + 6.067s (initrd) + 9.647s (userspace) = 16.463s. Aug 13 01:04:36.347786 kubelet[1284]: E0813 01:04:36.347704 1284 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 01:04:36.349735 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 01:04:36.349864 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 01:04:36.350104 systemd[1]: kubelet.service: Consumed 2.213s CPU time. Aug 13 01:04:44.392220 systemd[1]: Started sshd@3-10.0.0.132:22-10.0.0.1:50282.service. Aug 13 01:04:44.420691 sshd[1293]: Accepted publickey for core from 10.0.0.1 port 50282 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:44.421665 sshd[1293]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:44.425110 systemd-logind[1195]: New session 4 of user core. Aug 13 01:04:44.425854 systemd[1]: Started session-4.scope. Aug 13 01:04:44.479385 sshd[1293]: pam_unix(sshd:session): session closed for user core Aug 13 01:04:44.482512 systemd[1]: sshd@3-10.0.0.132:22-10.0.0.1:50282.service: Deactivated successfully. Aug 13 01:04:44.483147 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 01:04:44.483834 systemd-logind[1195]: Session 4 logged out. Waiting for processes to exit. Aug 13 01:04:44.485384 systemd[1]: Started sshd@4-10.0.0.132:22-10.0.0.1:50290.service. Aug 13 01:04:44.486191 systemd-logind[1195]: Removed session 4. Aug 13 01:04:44.514073 sshd[1299]: Accepted publickey for core from 10.0.0.1 port 50290 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:44.515184 sshd[1299]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:44.518468 systemd-logind[1195]: New session 5 of user core. Aug 13 01:04:44.519242 systemd[1]: Started session-5.scope. Aug 13 01:04:44.567668 sshd[1299]: pam_unix(sshd:session): session closed for user core Aug 13 01:04:44.570383 systemd[1]: sshd@4-10.0.0.132:22-10.0.0.1:50290.service: Deactivated successfully. Aug 13 01:04:44.570865 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 01:04:44.571335 systemd-logind[1195]: Session 5 logged out. Waiting for processes to exit. Aug 13 01:04:44.572376 systemd[1]: Started sshd@5-10.0.0.132:22-10.0.0.1:50304.service. Aug 13 01:04:44.573019 systemd-logind[1195]: Removed session 5. Aug 13 01:04:44.602653 sshd[1306]: Accepted publickey for core from 10.0.0.1 port 50304 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:44.604197 sshd[1306]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:44.607877 systemd-logind[1195]: New session 6 of user core. Aug 13 01:04:44.608910 systemd[1]: Started session-6.scope. Aug 13 01:04:44.663545 sshd[1306]: pam_unix(sshd:session): session closed for user core Aug 13 01:04:44.665847 systemd[1]: sshd@5-10.0.0.132:22-10.0.0.1:50304.service: Deactivated successfully. Aug 13 01:04:44.666387 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 01:04:44.666927 systemd-logind[1195]: Session 6 logged out. Waiting for processes to exit. Aug 13 01:04:44.667833 systemd[1]: Started sshd@6-10.0.0.132:22-10.0.0.1:50314.service. Aug 13 01:04:44.668483 systemd-logind[1195]: Removed session 6. Aug 13 01:04:44.697181 sshd[1312]: Accepted publickey for core from 10.0.0.1 port 50314 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:04:44.698387 sshd[1312]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:04:44.701511 systemd-logind[1195]: New session 7 of user core. Aug 13 01:04:44.702237 systemd[1]: Started session-7.scope. Aug 13 01:04:44.757568 sudo[1315]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 01:04:44.757762 sudo[1315]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Aug 13 01:04:44.788047 systemd[1]: Starting docker.service... Aug 13 01:04:44.877841 env[1326]: time="2025-08-13T01:04:44.877749631Z" level=info msg="Starting up" Aug 13 01:04:44.879567 env[1326]: time="2025-08-13T01:04:44.879532002Z" level=info msg="parsed scheme: \"unix\"" module=grpc Aug 13 01:04:44.879567 env[1326]: time="2025-08-13T01:04:44.879552451Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Aug 13 01:04:44.879666 env[1326]: time="2025-08-13T01:04:44.879572308Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Aug 13 01:04:44.879666 env[1326]: time="2025-08-13T01:04:44.879584200Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Aug 13 01:04:44.882222 env[1326]: time="2025-08-13T01:04:44.882178304Z" level=info msg="parsed scheme: \"unix\"" module=grpc Aug 13 01:04:44.882222 env[1326]: time="2025-08-13T01:04:44.882213671Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Aug 13 01:04:44.882330 env[1326]: time="2025-08-13T01:04:44.882238267Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Aug 13 01:04:44.882330 env[1326]: time="2025-08-13T01:04:44.882250640Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Aug 13 01:04:44.887038 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2793245944-merged.mount: Deactivated successfully. Aug 13 01:04:44.917790 env[1326]: time="2025-08-13T01:04:44.917595682Z" level=info msg="Loading containers: start." Aug 13 01:04:45.046301 kernel: Initializing XFRM netlink socket Aug 13 01:04:45.086373 env[1326]: time="2025-08-13T01:04:45.086322084Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Aug 13 01:04:45.142067 systemd-networkd[1020]: docker0: Link UP Aug 13 01:04:45.767375 env[1326]: time="2025-08-13T01:04:45.767325174Z" level=info msg="Loading containers: done." Aug 13 01:04:45.784539 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck830731211-merged.mount: Deactivated successfully. Aug 13 01:04:46.014193 env[1326]: time="2025-08-13T01:04:46.014083406Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 01:04:46.014488 env[1326]: time="2025-08-13T01:04:46.014433893Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Aug 13 01:04:46.014570 env[1326]: time="2025-08-13T01:04:46.014548668Z" level=info msg="Daemon has completed initialization" Aug 13 01:04:46.444725 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 01:04:46.444960 systemd[1]: Stopped kubelet.service. Aug 13 01:04:46.445018 systemd[1]: kubelet.service: Consumed 2.213s CPU time. Aug 13 01:04:46.446472 systemd[1]: Starting kubelet.service... Aug 13 01:04:46.446749 systemd[1]: Started docker.service. Aug 13 01:04:46.454362 env[1326]: time="2025-08-13T01:04:46.454275927Z" level=info msg="API listen on /run/docker.sock" Aug 13 01:04:46.670432 systemd[1]: Started kubelet.service. Aug 13 01:04:46.804695 kubelet[1456]: E0813 01:04:46.804560 1456 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 01:04:46.807712 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 01:04:46.807842 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 01:04:47.431425 env[1207]: time="2025-08-13T01:04:47.431359989Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\"" Aug 13 01:04:48.940743 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2365839110.mount: Deactivated successfully. Aug 13 01:04:50.833173 env[1207]: time="2025-08-13T01:04:50.833103608Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:50.835031 env[1207]: time="2025-08-13T01:04:50.834992349Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ea7fa3cfabed1b85e7de8e0a02356b6dcb7708442d6e4600d68abaebe1e9b1fc,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:50.837093 env[1207]: time="2025-08-13T01:04:50.837044998Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:50.838977 env[1207]: time="2025-08-13T01:04:50.838925564Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:50.839689 env[1207]: time="2025-08-13T01:04:50.839652427Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\" returns image reference \"sha256:ea7fa3cfabed1b85e7de8e0a02356b6dcb7708442d6e4600d68abaebe1e9b1fc\"" Aug 13 01:04:50.840904 env[1207]: time="2025-08-13T01:04:50.840868597Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\"" Aug 13 01:04:53.285761 env[1207]: time="2025-08-13T01:04:53.285693343Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:53.287790 env[1207]: time="2025-08-13T01:04:53.287755379Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c057eceea4b436b01f9ce394734cfb06f13b2a3688c3983270e99743370b6051,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:53.290885 env[1207]: time="2025-08-13T01:04:53.290855822Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:53.292842 env[1207]: time="2025-08-13T01:04:53.292784198Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:53.293769 env[1207]: time="2025-08-13T01:04:53.293727497Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\" returns image reference \"sha256:c057eceea4b436b01f9ce394734cfb06f13b2a3688c3983270e99743370b6051\"" Aug 13 01:04:53.294462 env[1207]: time="2025-08-13T01:04:53.294438510Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\"" Aug 13 01:04:55.554544 env[1207]: time="2025-08-13T01:04:55.554478484Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:55.556419 env[1207]: time="2025-08-13T01:04:55.556355965Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:64e6a0b453108c87da0bb61473b35fd54078119a09edc56a4c8cb31602437c58,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:55.558239 env[1207]: time="2025-08-13T01:04:55.558211864Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:55.559951 env[1207]: time="2025-08-13T01:04:55.559917973Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:55.561200 env[1207]: time="2025-08-13T01:04:55.561154522Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\" returns image reference \"sha256:64e6a0b453108c87da0bb61473b35fd54078119a09edc56a4c8cb31602437c58\"" Aug 13 01:04:55.562872 env[1207]: time="2025-08-13T01:04:55.562833229Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\"" Aug 13 01:04:56.944815 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 01:04:56.945023 systemd[1]: Stopped kubelet.service. Aug 13 01:04:56.946816 systemd[1]: Starting kubelet.service... Aug 13 01:04:57.040945 systemd[1]: Started kubelet.service. Aug 13 01:04:57.090335 kubelet[1471]: E0813 01:04:57.090244 1471 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 01:04:57.092184 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 01:04:57.092317 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 01:04:58.126967 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3917749125.mount: Deactivated successfully. Aug 13 01:04:59.571532 env[1207]: time="2025-08-13T01:04:59.571449143Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:59.574704 env[1207]: time="2025-08-13T01:04:59.574667187Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:0cec28fd5c3c446ec52e2886ddea38bf7f7e17755aa5d0095d50d3df5914a8fd,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:59.576347 env[1207]: time="2025-08-13T01:04:59.576316239Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:59.577892 env[1207]: time="2025-08-13T01:04:59.577835237Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:04:59.578282 env[1207]: time="2025-08-13T01:04:59.578233364Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\" returns image reference \"sha256:0cec28fd5c3c446ec52e2886ddea38bf7f7e17755aa5d0095d50d3df5914a8fd\"" Aug 13 01:04:59.578877 env[1207]: time="2025-08-13T01:04:59.578850741Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Aug 13 01:05:00.164687 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4259279734.mount: Deactivated successfully. Aug 13 01:05:01.434411 env[1207]: time="2025-08-13T01:05:01.434343419Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:01.436378 env[1207]: time="2025-08-13T01:05:01.436325105Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:01.438106 env[1207]: time="2025-08-13T01:05:01.438073684Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:01.439740 env[1207]: time="2025-08-13T01:05:01.439716564Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:01.440452 env[1207]: time="2025-08-13T01:05:01.440423369Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Aug 13 01:05:01.441062 env[1207]: time="2025-08-13T01:05:01.441040106Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 01:05:02.009902 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount117139669.mount: Deactivated successfully. Aug 13 01:05:02.013125 env[1207]: time="2025-08-13T01:05:02.013076704Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:02.014893 env[1207]: time="2025-08-13T01:05:02.014850781Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:02.016575 env[1207]: time="2025-08-13T01:05:02.016518157Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:02.017853 env[1207]: time="2025-08-13T01:05:02.017809949Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:02.018267 env[1207]: time="2025-08-13T01:05:02.018220739Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Aug 13 01:05:02.018980 env[1207]: time="2025-08-13T01:05:02.018951309Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Aug 13 01:05:02.654670 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount39394714.mount: Deactivated successfully. Aug 13 01:05:06.499631 env[1207]: time="2025-08-13T01:05:06.499563887Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:06.503975 env[1207]: time="2025-08-13T01:05:06.503909387Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:06.506169 env[1207]: time="2025-08-13T01:05:06.506130380Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:06.509294 env[1207]: time="2025-08-13T01:05:06.509238169Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:06.510081 env[1207]: time="2025-08-13T01:05:06.510040452Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Aug 13 01:05:07.194638 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Aug 13 01:05:07.194804 systemd[1]: Stopped kubelet.service. Aug 13 01:05:07.196120 systemd[1]: Starting kubelet.service... Aug 13 01:05:07.294703 systemd[1]: Started kubelet.service. Aug 13 01:05:07.327322 kubelet[1503]: E0813 01:05:07.327240 1503 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 01:05:07.329183 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 01:05:07.329319 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 01:05:09.157552 systemd[1]: Stopped kubelet.service. Aug 13 01:05:09.159547 systemd[1]: Starting kubelet.service... Aug 13 01:05:09.180513 systemd[1]: Reloading. Aug 13 01:05:09.247010 /usr/lib/systemd/system-generators/torcx-generator[1538]: time="2025-08-13T01:05:09Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 01:05:09.247389 /usr/lib/systemd/system-generators/torcx-generator[1538]: time="2025-08-13T01:05:09Z" level=info msg="torcx already run" Aug 13 01:05:09.485425 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 01:05:09.485443 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 01:05:09.502566 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 01:05:09.579310 systemd[1]: Started kubelet.service. Aug 13 01:05:09.580609 systemd[1]: Stopping kubelet.service... Aug 13 01:05:09.580870 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 01:05:09.581051 systemd[1]: Stopped kubelet.service. Aug 13 01:05:09.582471 systemd[1]: Starting kubelet.service... Aug 13 01:05:09.674675 systemd[1]: Started kubelet.service. Aug 13 01:05:09.721907 kubelet[1585]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 01:05:09.721907 kubelet[1585]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 01:05:09.721907 kubelet[1585]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 01:05:09.722573 kubelet[1585]: I0813 01:05:09.721967 1585 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 01:05:10.252748 kubelet[1585]: I0813 01:05:10.248955 1585 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 01:05:10.252748 kubelet[1585]: I0813 01:05:10.249009 1585 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 01:05:10.252748 kubelet[1585]: I0813 01:05:10.249318 1585 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 01:05:10.288462 kubelet[1585]: E0813 01:05:10.288197 1585 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.132:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:10.289158 kubelet[1585]: I0813 01:05:10.288957 1585 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 01:05:10.308701 kubelet[1585]: E0813 01:05:10.308594 1585 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 01:05:10.308701 kubelet[1585]: I0813 01:05:10.308672 1585 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 01:05:10.318041 kubelet[1585]: I0813 01:05:10.317976 1585 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 01:05:10.326599 kubelet[1585]: I0813 01:05:10.326529 1585 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 01:05:10.331663 kubelet[1585]: I0813 01:05:10.331578 1585 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 01:05:10.331923 kubelet[1585]: I0813 01:05:10.331642 1585 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 01:05:10.332045 kubelet[1585]: I0813 01:05:10.331933 1585 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 01:05:10.332045 kubelet[1585]: I0813 01:05:10.331946 1585 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 01:05:10.332149 kubelet[1585]: I0813 01:05:10.332122 1585 state_mem.go:36] "Initialized new in-memory state store" Aug 13 01:05:10.338123 kubelet[1585]: I0813 01:05:10.338062 1585 kubelet.go:408] "Attempting to sync node with API server" Aug 13 01:05:10.338210 kubelet[1585]: I0813 01:05:10.338140 1585 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 01:05:10.338275 kubelet[1585]: I0813 01:05:10.338213 1585 kubelet.go:314] "Adding apiserver pod source" Aug 13 01:05:10.338275 kubelet[1585]: I0813 01:05:10.338271 1585 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 01:05:10.356364 kubelet[1585]: I0813 01:05:10.356311 1585 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Aug 13 01:05:10.356685 kubelet[1585]: W0813 01:05:10.356583 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:10.356747 kubelet[1585]: E0813 01:05:10.356721 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:10.356889 kubelet[1585]: I0813 01:05:10.356836 1585 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 01:05:10.357350 kubelet[1585]: W0813 01:05:10.357227 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:10.357423 kubelet[1585]: E0813 01:05:10.357343 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:10.357520 kubelet[1585]: W0813 01:05:10.357486 1585 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 01:05:10.367806 kubelet[1585]: I0813 01:05:10.365471 1585 server.go:1274] "Started kubelet" Aug 13 01:05:10.367923 kubelet[1585]: I0813 01:05:10.367880 1585 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 01:05:10.373191 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Aug 13 01:05:10.373282 kubelet[1585]: I0813 01:05:10.373015 1585 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 01:05:10.374698 kubelet[1585]: I0813 01:05:10.374008 1585 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 01:05:10.374698 kubelet[1585]: I0813 01:05:10.374021 1585 server.go:449] "Adding debug handlers to kubelet server" Aug 13 01:05:10.374698 kubelet[1585]: E0813 01:05:10.373062 1585 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.132:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.132:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.185b2e0c5967abd5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-08-13 01:05:10.365391829 +0000 UTC m=+0.687090613,LastTimestamp:2025-08-13 01:05:10.365391829 +0000 UTC m=+0.687090613,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Aug 13 01:05:10.374698 kubelet[1585]: I0813 01:05:10.374338 1585 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 01:05:10.374945 kubelet[1585]: I0813 01:05:10.374739 1585 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 01:05:10.375973 kubelet[1585]: I0813 01:05:10.375062 1585 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 01:05:10.375973 kubelet[1585]: I0813 01:05:10.375183 1585 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 01:05:10.375973 kubelet[1585]: I0813 01:05:10.375241 1585 reconciler.go:26] "Reconciler: start to sync state" Aug 13 01:05:10.375973 kubelet[1585]: W0813 01:05:10.375582 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:10.375973 kubelet[1585]: E0813 01:05:10.375628 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:10.375973 kubelet[1585]: E0813 01:05:10.375837 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.381446 kubelet[1585]: E0813 01:05:10.381415 1585 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 01:05:10.381860 kubelet[1585]: E0813 01:05:10.381820 1585 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="200ms" Aug 13 01:05:10.382892 kubelet[1585]: I0813 01:05:10.382852 1585 factory.go:221] Registration of the containerd container factory successfully Aug 13 01:05:10.382892 kubelet[1585]: I0813 01:05:10.382886 1585 factory.go:221] Registration of the systemd container factory successfully Aug 13 01:05:10.383059 kubelet[1585]: I0813 01:05:10.383013 1585 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 01:05:10.418973 kubelet[1585]: I0813 01:05:10.418667 1585 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 01:05:10.418973 kubelet[1585]: I0813 01:05:10.418694 1585 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 01:05:10.418973 kubelet[1585]: I0813 01:05:10.418729 1585 state_mem.go:36] "Initialized new in-memory state store" Aug 13 01:05:10.432879 kubelet[1585]: I0813 01:05:10.432772 1585 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 01:05:10.435632 kubelet[1585]: I0813 01:05:10.435600 1585 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 01:05:10.435714 kubelet[1585]: I0813 01:05:10.435663 1585 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 01:05:10.435714 kubelet[1585]: I0813 01:05:10.435701 1585 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 01:05:10.435787 kubelet[1585]: E0813 01:05:10.435755 1585 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 01:05:10.476863 kubelet[1585]: E0813 01:05:10.476754 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.536102 kubelet[1585]: E0813 01:05:10.535888 1585 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Aug 13 01:05:10.577366 kubelet[1585]: E0813 01:05:10.577279 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.583084 kubelet[1585]: E0813 01:05:10.583024 1585 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="400ms" Aug 13 01:05:10.678461 kubelet[1585]: E0813 01:05:10.678352 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.736964 kubelet[1585]: E0813 01:05:10.736865 1585 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Aug 13 01:05:10.779503 kubelet[1585]: E0813 01:05:10.779413 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.880828 kubelet[1585]: E0813 01:05:10.880436 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.981070 kubelet[1585]: E0813 01:05:10.980933 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:10.983756 kubelet[1585]: E0813 01:05:10.983681 1585 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="800ms" Aug 13 01:05:11.081419 kubelet[1585]: E0813 01:05:11.081291 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.139907 kubelet[1585]: E0813 01:05:11.137557 1585 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Aug 13 01:05:11.182353 kubelet[1585]: E0813 01:05:11.182230 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.282809 kubelet[1585]: E0813 01:05:11.282735 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.383323 kubelet[1585]: E0813 01:05:11.383271 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.483922 kubelet[1585]: E0813 01:05:11.483878 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.519469 kubelet[1585]: W0813 01:05:11.519405 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:11.519469 kubelet[1585]: E0813 01:05:11.519464 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.132:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:11.583973 kubelet[1585]: E0813 01:05:11.583948 1585 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Aug 13 01:05:11.601052 kubelet[1585]: W0813 01:05:11.601031 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:11.601111 kubelet[1585]: E0813 01:05:11.601056 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:11.601782 kubelet[1585]: I0813 01:05:11.601760 1585 policy_none.go:49] "None policy: Start" Aug 13 01:05:11.602364 kubelet[1585]: I0813 01:05:11.602348 1585 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 01:05:11.602417 kubelet[1585]: I0813 01:05:11.602377 1585 state_mem.go:35] "Initializing new in-memory state store" Aug 13 01:05:11.609840 systemd[1]: Created slice kubepods.slice. Aug 13 01:05:11.613424 systemd[1]: Created slice kubepods-burstable.slice. Aug 13 01:05:11.615975 systemd[1]: Created slice kubepods-besteffort.slice. Aug 13 01:05:11.621958 kubelet[1585]: I0813 01:05:11.621917 1585 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 01:05:11.622133 kubelet[1585]: I0813 01:05:11.622112 1585 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 01:05:11.622201 kubelet[1585]: I0813 01:05:11.622140 1585 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 01:05:11.622468 kubelet[1585]: I0813 01:05:11.622437 1585 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 01:05:11.623209 kubelet[1585]: W0813 01:05:11.623158 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:11.623289 kubelet[1585]: E0813 01:05:11.623222 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:11.623381 kubelet[1585]: E0813 01:05:11.623362 1585 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Aug 13 01:05:11.630208 kubelet[1585]: W0813 01:05:11.630143 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:11.630208 kubelet[1585]: E0813 01:05:11.630211 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.132:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:11.724020 kubelet[1585]: I0813 01:05:11.723962 1585 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:11.724338 kubelet[1585]: E0813 01:05:11.724302 1585 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost" Aug 13 01:05:11.784045 kubelet[1585]: E0813 01:05:11.783958 1585 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="1.6s" Aug 13 01:05:11.925894 kubelet[1585]: I0813 01:05:11.925848 1585 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:11.926165 kubelet[1585]: E0813 01:05:11.926132 1585 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost" Aug 13 01:05:11.943746 systemd[1]: Created slice kubepods-burstable-pod4d5290f0ff42efabb9babf0d29c7f9e3.slice. Aug 13 01:05:11.960272 systemd[1]: Created slice kubepods-burstable-pod407c569889bb86d746b0274843003fd0.slice. Aug 13 01:05:11.971715 systemd[1]: Created slice kubepods-burstable-pod27e4a50e94f48ec00f6bd509cb48ed05.slice. Aug 13 01:05:11.982215 kubelet[1585]: I0813 01:05:11.982183 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:11.982303 kubelet[1585]: I0813 01:05:11.982224 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/27e4a50e94f48ec00f6bd509cb48ed05-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"27e4a50e94f48ec00f6bd509cb48ed05\") " pod="kube-system/kube-scheduler-localhost" Aug 13 01:05:11.982303 kubelet[1585]: I0813 01:05:11.982248 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:11.982303 kubelet[1585]: I0813 01:05:11.982284 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:11.982422 kubelet[1585]: I0813 01:05:11.982305 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:11.982422 kubelet[1585]: I0813 01:05:11.982327 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:11.982422 kubelet[1585]: I0813 01:05:11.982349 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:11.982422 kubelet[1585]: I0813 01:05:11.982368 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:11.982422 kubelet[1585]: I0813 01:05:11.982413 1585 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:12.259500 kubelet[1585]: E0813 01:05:12.259459 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:12.260279 env[1207]: time="2025-08-13T01:05:12.260217347Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:4d5290f0ff42efabb9babf0d29c7f9e3,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:12.270468 kubelet[1585]: E0813 01:05:12.270433 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:12.271034 env[1207]: time="2025-08-13T01:05:12.270990203Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:407c569889bb86d746b0274843003fd0,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:12.273191 kubelet[1585]: E0813 01:05:12.273176 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:12.273439 env[1207]: time="2025-08-13T01:05:12.273412937Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:27e4a50e94f48ec00f6bd509cb48ed05,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:12.327943 kubelet[1585]: I0813 01:05:12.327926 1585 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:12.328327 kubelet[1585]: E0813 01:05:12.328289 1585 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost" Aug 13 01:05:12.377827 kubelet[1585]: E0813 01:05:12.377793 1585 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.132:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:12.822759 kubelet[1585]: W0813 01:05:12.822689 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:12.823084 kubelet[1585]: E0813 01:05:12.822766 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.132:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:13.130022 kubelet[1585]: I0813 01:05:13.129894 1585 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:13.130403 kubelet[1585]: E0813 01:05:13.130344 1585 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.132:6443/api/v1/nodes\": dial tcp 10.0.0.132:6443: connect: connection refused" node="localhost" Aug 13 01:05:13.385085 kubelet[1585]: E0813 01:05:13.384939 1585 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.132:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.132:6443: connect: connection refused" interval="3.2s" Aug 13 01:05:13.777563 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1930164576.mount: Deactivated successfully. Aug 13 01:05:13.778795 kubelet[1585]: W0813 01:05:13.778758 1585 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.132:6443: connect: connection refused Aug 13 01:05:13.778885 kubelet[1585]: E0813 01:05:13.778808 1585 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.132:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.132:6443: connect: connection refused" logger="UnhandledError" Aug 13 01:05:13.783607 env[1207]: time="2025-08-13T01:05:13.783570375Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.786610 env[1207]: time="2025-08-13T01:05:13.786556747Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.787789 env[1207]: time="2025-08-13T01:05:13.787720466Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.790194 env[1207]: time="2025-08-13T01:05:13.790155307Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.791903 env[1207]: time="2025-08-13T01:05:13.791871258Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.793341 env[1207]: time="2025-08-13T01:05:13.793308828Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.795474 env[1207]: time="2025-08-13T01:05:13.795434440Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.797079 env[1207]: time="2025-08-13T01:05:13.797032486Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.798396 env[1207]: time="2025-08-13T01:05:13.798372530Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.800044 env[1207]: time="2025-08-13T01:05:13.800011414Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.801578 env[1207]: time="2025-08-13T01:05:13.801543234Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.802886 env[1207]: time="2025-08-13T01:05:13.802854093Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:13.822352 env[1207]: time="2025-08-13T01:05:13.822274832Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:13.822352 env[1207]: time="2025-08-13T01:05:13.822312934Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:13.822352 env[1207]: time="2025-08-13T01:05:13.822323765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:13.822601 env[1207]: time="2025-08-13T01:05:13.822494791Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/95d555f1936dc1247bcd2849f0df94f28bfdb091aae474f14f274fdadcf0be8d pid=1629 runtime=io.containerd.runc.v2 Aug 13 01:05:13.834438 systemd[1]: Started cri-containerd-95d555f1936dc1247bcd2849f0df94f28bfdb091aae474f14f274fdadcf0be8d.scope. Aug 13 01:05:13.842944 env[1207]: time="2025-08-13T01:05:13.842099571Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:13.842944 env[1207]: time="2025-08-13T01:05:13.842132043Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:13.842944 env[1207]: time="2025-08-13T01:05:13.842141170Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:13.842944 env[1207]: time="2025-08-13T01:05:13.842238015Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e055d09278ee282a155c0b194bf6877a747ead0b9c55a084bd89f7829cd01806 pid=1662 runtime=io.containerd.runc.v2 Aug 13 01:05:13.846063 env[1207]: time="2025-08-13T01:05:13.845986810Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:13.846147 env[1207]: time="2025-08-13T01:05:13.846073686Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:13.846147 env[1207]: time="2025-08-13T01:05:13.846098302Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:13.847321 env[1207]: time="2025-08-13T01:05:13.846460033Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/925815f6efd5cd6cc7df117e4fc6d8a5c8ed9eb6ad723492f46f6c3503687239 pid=1680 runtime=io.containerd.runc.v2 Aug 13 01:05:13.857499 systemd[1]: Started cri-containerd-e055d09278ee282a155c0b194bf6877a747ead0b9c55a084bd89f7829cd01806.scope. Aug 13 01:05:13.862787 systemd[1]: Started cri-containerd-925815f6efd5cd6cc7df117e4fc6d8a5c8ed9eb6ad723492f46f6c3503687239.scope. Aug 13 01:05:13.874185 env[1207]: time="2025-08-13T01:05:13.874131819Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:407c569889bb86d746b0274843003fd0,Namespace:kube-system,Attempt:0,} returns sandbox id \"95d555f1936dc1247bcd2849f0df94f28bfdb091aae474f14f274fdadcf0be8d\"" Aug 13 01:05:13.875529 kubelet[1585]: E0813 01:05:13.875306 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:13.877381 env[1207]: time="2025-08-13T01:05:13.877344764Z" level=info msg="CreateContainer within sandbox \"95d555f1936dc1247bcd2849f0df94f28bfdb091aae474f14f274fdadcf0be8d\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 01:05:13.897190 env[1207]: time="2025-08-13T01:05:13.897139745Z" level=info msg="CreateContainer within sandbox \"95d555f1936dc1247bcd2849f0df94f28bfdb091aae474f14f274fdadcf0be8d\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b814b030c09e6af95ca478248cf5d8b4dcd9e3c8f5590d16e85a140093c455a5\"" Aug 13 01:05:13.897971 env[1207]: time="2025-08-13T01:05:13.897942275Z" level=info msg="StartContainer for \"b814b030c09e6af95ca478248cf5d8b4dcd9e3c8f5590d16e85a140093c455a5\"" Aug 13 01:05:13.900440 env[1207]: time="2025-08-13T01:05:13.900414568Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:4d5290f0ff42efabb9babf0d29c7f9e3,Namespace:kube-system,Attempt:0,} returns sandbox id \"e055d09278ee282a155c0b194bf6877a747ead0b9c55a084bd89f7829cd01806\"" Aug 13 01:05:13.901340 kubelet[1585]: E0813 01:05:13.901104 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:13.902429 env[1207]: time="2025-08-13T01:05:13.902400965Z" level=info msg="CreateContainer within sandbox \"e055d09278ee282a155c0b194bf6877a747ead0b9c55a084bd89f7829cd01806\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 01:05:13.904509 env[1207]: time="2025-08-13T01:05:13.904477883Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:27e4a50e94f48ec00f6bd509cb48ed05,Namespace:kube-system,Attempt:0,} returns sandbox id \"925815f6efd5cd6cc7df117e4fc6d8a5c8ed9eb6ad723492f46f6c3503687239\"" Aug 13 01:05:13.905139 kubelet[1585]: E0813 01:05:13.905031 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:13.906573 env[1207]: time="2025-08-13T01:05:13.906537890Z" level=info msg="CreateContainer within sandbox \"925815f6efd5cd6cc7df117e4fc6d8a5c8ed9eb6ad723492f46f6c3503687239\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 01:05:13.916645 env[1207]: time="2025-08-13T01:05:13.916599058Z" level=info msg="CreateContainer within sandbox \"e055d09278ee282a155c0b194bf6877a747ead0b9c55a084bd89f7829cd01806\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0c021a154be37b4dafe3bc183e461280cfc39943445c054ce30f62a707b6ac8b\"" Aug 13 01:05:13.917103 env[1207]: time="2025-08-13T01:05:13.917062571Z" level=info msg="StartContainer for \"0c021a154be37b4dafe3bc183e461280cfc39943445c054ce30f62a707b6ac8b\"" Aug 13 01:05:13.919792 systemd[1]: Started cri-containerd-b814b030c09e6af95ca478248cf5d8b4dcd9e3c8f5590d16e85a140093c455a5.scope. Aug 13 01:05:13.927555 env[1207]: time="2025-08-13T01:05:13.927511999Z" level=info msg="CreateContainer within sandbox \"925815f6efd5cd6cc7df117e4fc6d8a5c8ed9eb6ad723492f46f6c3503687239\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"b25f806a2fcd160b1023e7106bdb3dcbdbf6a4202cc3c45c85c019177d4a63c2\"" Aug 13 01:05:13.928118 env[1207]: time="2025-08-13T01:05:13.928094911Z" level=info msg="StartContainer for \"b25f806a2fcd160b1023e7106bdb3dcbdbf6a4202cc3c45c85c019177d4a63c2\"" Aug 13 01:05:13.935031 systemd[1]: Started cri-containerd-0c021a154be37b4dafe3bc183e461280cfc39943445c054ce30f62a707b6ac8b.scope. Aug 13 01:05:13.945510 systemd[1]: Started cri-containerd-b25f806a2fcd160b1023e7106bdb3dcbdbf6a4202cc3c45c85c019177d4a63c2.scope. Aug 13 01:05:13.964909 env[1207]: time="2025-08-13T01:05:13.964864389Z" level=info msg="StartContainer for \"b814b030c09e6af95ca478248cf5d8b4dcd9e3c8f5590d16e85a140093c455a5\" returns successfully" Aug 13 01:05:13.979410 env[1207]: time="2025-08-13T01:05:13.979098210Z" level=info msg="StartContainer for \"0c021a154be37b4dafe3bc183e461280cfc39943445c054ce30f62a707b6ac8b\" returns successfully" Aug 13 01:05:13.993025 env[1207]: time="2025-08-13T01:05:13.992979992Z" level=info msg="StartContainer for \"b25f806a2fcd160b1023e7106bdb3dcbdbf6a4202cc3c45c85c019177d4a63c2\" returns successfully" Aug 13 01:05:14.445439 kubelet[1585]: E0813 01:05:14.445406 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:14.447115 kubelet[1585]: E0813 01:05:14.447091 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:14.448656 kubelet[1585]: E0813 01:05:14.448637 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:14.732354 kubelet[1585]: I0813 01:05:14.732205 1585 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:15.099434 kubelet[1585]: I0813 01:05:15.099303 1585 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Aug 13 01:05:15.353709 kubelet[1585]: I0813 01:05:15.353554 1585 apiserver.go:52] "Watching apiserver" Aug 13 01:05:15.375820 kubelet[1585]: I0813 01:05:15.375787 1585 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 01:05:15.453957 kubelet[1585]: E0813 01:05:15.453649 1585 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Aug 13 01:05:15.453957 kubelet[1585]: E0813 01:05:15.453754 1585 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:15.453957 kubelet[1585]: E0813 01:05:15.453801 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:15.453957 kubelet[1585]: E0813 01:05:15.453903 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:18.151784 update_engine[1200]: I0813 01:05:18.151700 1200 update_attempter.cc:509] Updating boot flags... Aug 13 01:05:19.864021 kubelet[1585]: E0813 01:05:19.863985 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:20.456361 kubelet[1585]: E0813 01:05:20.456304 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:20.456565 kubelet[1585]: I0813 01:05:20.456374 1585 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.456352624 podStartE2EDuration="1.456352624s" podCreationTimestamp="2025-08-13 01:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:20.455232992 +0000 UTC m=+10.776931796" watchObservedRunningTime="2025-08-13 01:05:20.456352624 +0000 UTC m=+10.778051408" Aug 13 01:05:20.456812 kubelet[1585]: E0813 01:05:20.456764 1585 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:20.615729 systemd[1]: Reloading. Aug 13 01:05:20.672069 /usr/lib/systemd/system-generators/torcx-generator[1904]: time="2025-08-13T01:05:20Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 01:05:20.672095 /usr/lib/systemd/system-generators/torcx-generator[1904]: time="2025-08-13T01:05:20Z" level=info msg="torcx already run" Aug 13 01:05:20.733147 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 01:05:20.733165 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 01:05:20.750369 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 01:05:20.841360 systemd[1]: Stopping kubelet.service... Aug 13 01:05:20.864678 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 01:05:20.864870 systemd[1]: Stopped kubelet.service. Aug 13 01:05:20.864919 systemd[1]: kubelet.service: Consumed 1.178s CPU time. Aug 13 01:05:20.866576 systemd[1]: Starting kubelet.service... Aug 13 01:05:20.958542 systemd[1]: Started kubelet.service. Aug 13 01:05:20.995591 kubelet[1948]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 01:05:20.995591 kubelet[1948]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 01:05:20.995591 kubelet[1948]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 01:05:20.995591 kubelet[1948]: I0813 01:05:20.995563 1948 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 01:05:21.001775 kubelet[1948]: I0813 01:05:21.001366 1948 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 01:05:21.001775 kubelet[1948]: I0813 01:05:21.001399 1948 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 01:05:21.001775 kubelet[1948]: I0813 01:05:21.001685 1948 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 01:05:21.002899 kubelet[1948]: I0813 01:05:21.002872 1948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Aug 13 01:05:21.004551 kubelet[1948]: I0813 01:05:21.004525 1948 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 01:05:21.010751 kubelet[1948]: E0813 01:05:21.010717 1948 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 01:05:21.010751 kubelet[1948]: I0813 01:05:21.010741 1948 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 01:05:21.017546 kubelet[1948]: I0813 01:05:21.016405 1948 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 01:05:21.017546 kubelet[1948]: I0813 01:05:21.017009 1948 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 01:05:21.017546 kubelet[1948]: I0813 01:05:21.017135 1948 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 01:05:21.017546 kubelet[1948]: I0813 01:05:21.017158 1948 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 01:05:21.017793 kubelet[1948]: I0813 01:05:21.017421 1948 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 01:05:21.017793 kubelet[1948]: I0813 01:05:21.017428 1948 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 01:05:21.017793 kubelet[1948]: I0813 01:05:21.017452 1948 state_mem.go:36] "Initialized new in-memory state store" Aug 13 01:05:21.018351 kubelet[1948]: I0813 01:05:21.018331 1948 kubelet.go:408] "Attempting to sync node with API server" Aug 13 01:05:21.018440 kubelet[1948]: I0813 01:05:21.018424 1948 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 01:05:21.018542 kubelet[1948]: I0813 01:05:21.018527 1948 kubelet.go:314] "Adding apiserver pod source" Aug 13 01:05:21.018629 kubelet[1948]: I0813 01:05:21.018614 1948 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 01:05:21.019694 kubelet[1948]: I0813 01:05:21.019679 1948 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Aug 13 01:05:21.020111 kubelet[1948]: I0813 01:05:21.020094 1948 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 01:05:21.020561 kubelet[1948]: I0813 01:05:21.020546 1948 server.go:1274] "Started kubelet" Aug 13 01:05:21.023497 kubelet[1948]: I0813 01:05:21.023484 1948 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 01:05:21.026809 kubelet[1948]: I0813 01:05:21.026774 1948 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 01:05:21.027497 kubelet[1948]: I0813 01:05:21.027473 1948 server.go:449] "Adding debug handlers to kubelet server" Aug 13 01:05:21.028224 kubelet[1948]: I0813 01:05:21.028197 1948 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 01:05:21.028551 kubelet[1948]: I0813 01:05:21.028536 1948 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 01:05:21.029574 kubelet[1948]: I0813 01:05:21.029560 1948 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 01:05:21.033405 kubelet[1948]: E0813 01:05:21.033374 1948 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 01:05:21.034296 kubelet[1948]: I0813 01:05:21.034271 1948 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 01:05:21.034459 kubelet[1948]: I0813 01:05:21.034443 1948 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 01:05:21.034548 kubelet[1948]: I0813 01:05:21.034526 1948 factory.go:221] Registration of the systemd container factory successfully Aug 13 01:05:21.034624 kubelet[1948]: I0813 01:05:21.034602 1948 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 01:05:21.034771 kubelet[1948]: I0813 01:05:21.034757 1948 reconciler.go:26] "Reconciler: start to sync state" Aug 13 01:05:21.036243 kubelet[1948]: I0813 01:05:21.036228 1948 factory.go:221] Registration of the containerd container factory successfully Aug 13 01:05:21.042394 kubelet[1948]: I0813 01:05:21.042358 1948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 01:05:21.043335 kubelet[1948]: I0813 01:05:21.043316 1948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 01:05:21.043335 kubelet[1948]: I0813 01:05:21.043335 1948 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 01:05:21.043421 kubelet[1948]: I0813 01:05:21.043351 1948 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 01:05:21.043421 kubelet[1948]: E0813 01:05:21.043385 1948 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 01:05:21.062484 kubelet[1948]: I0813 01:05:21.062449 1948 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 01:05:21.062484 kubelet[1948]: I0813 01:05:21.062464 1948 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 01:05:21.062484 kubelet[1948]: I0813 01:05:21.062481 1948 state_mem.go:36] "Initialized new in-memory state store" Aug 13 01:05:21.062681 kubelet[1948]: I0813 01:05:21.062605 1948 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 01:05:21.062681 kubelet[1948]: I0813 01:05:21.062614 1948 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 01:05:21.062681 kubelet[1948]: I0813 01:05:21.062630 1948 policy_none.go:49] "None policy: Start" Aug 13 01:05:21.063067 kubelet[1948]: I0813 01:05:21.063051 1948 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 01:05:21.063120 kubelet[1948]: I0813 01:05:21.063071 1948 state_mem.go:35] "Initializing new in-memory state store" Aug 13 01:05:21.063190 kubelet[1948]: I0813 01:05:21.063180 1948 state_mem.go:75] "Updated machine memory state" Aug 13 01:05:21.066793 kubelet[1948]: I0813 01:05:21.066763 1948 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 01:05:21.066972 kubelet[1948]: I0813 01:05:21.066948 1948 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 01:05:21.067046 kubelet[1948]: I0813 01:05:21.066968 1948 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 01:05:21.067175 kubelet[1948]: I0813 01:05:21.067152 1948 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 01:05:21.152398 kubelet[1948]: E0813 01:05:21.152326 1948 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:21.152398 kubelet[1948]: E0813 01:05:21.152378 1948 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.175668 kubelet[1948]: I0813 01:05:21.175598 1948 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Aug 13 01:05:21.182185 kubelet[1948]: I0813 01:05:21.182131 1948 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Aug 13 01:05:21.182374 kubelet[1948]: I0813 01:05:21.182242 1948 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Aug 13 01:05:21.236345 kubelet[1948]: I0813 01:05:21.236284 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.236345 kubelet[1948]: I0813 01:05:21.236330 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.236345 kubelet[1948]: I0813 01:05:21.236371 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/27e4a50e94f48ec00f6bd509cb48ed05-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"27e4a50e94f48ec00f6bd509cb48ed05\") " pod="kube-system/kube-scheduler-localhost" Aug 13 01:05:21.236633 kubelet[1948]: I0813 01:05:21.236397 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:21.236633 kubelet[1948]: I0813 01:05:21.236415 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.236633 kubelet[1948]: I0813 01:05:21.236506 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.236633 kubelet[1948]: I0813 01:05:21.236545 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/407c569889bb86d746b0274843003fd0-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"407c569889bb86d746b0274843003fd0\") " pod="kube-system/kube-controller-manager-localhost" Aug 13 01:05:21.236633 kubelet[1948]: I0813 01:05:21.236564 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:21.236807 kubelet[1948]: I0813 01:05:21.236581 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4d5290f0ff42efabb9babf0d29c7f9e3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"4d5290f0ff42efabb9babf0d29c7f9e3\") " pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:21.452594 kubelet[1948]: E0813 01:05:21.452524 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:21.452840 kubelet[1948]: E0813 01:05:21.452718 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:21.452840 kubelet[1948]: E0813 01:05:21.452754 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:21.614736 sudo[1984]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Aug 13 01:05:21.615047 sudo[1984]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Aug 13 01:05:22.019942 kubelet[1948]: I0813 01:05:22.019897 1948 apiserver.go:52] "Watching apiserver" Aug 13 01:05:22.035202 kubelet[1948]: I0813 01:05:22.035167 1948 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 01:05:22.056894 kubelet[1948]: E0813 01:05:22.056845 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:22.057330 kubelet[1948]: E0813 01:05:22.057302 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:22.079913 sudo[1984]: pam_unix(sudo:session): session closed for user root Aug 13 01:05:22.238026 kubelet[1948]: E0813 01:05:22.237963 1948 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Aug 13 01:05:22.238237 kubelet[1948]: E0813 01:05:22.238178 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:22.248515 kubelet[1948]: I0813 01:05:22.248432 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.248404992 podStartE2EDuration="1.248404992s" podCreationTimestamp="2025-08-13 01:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:22.238083949 +0000 UTC m=+1.275643618" watchObservedRunningTime="2025-08-13 01:05:22.248404992 +0000 UTC m=+1.285964650" Aug 13 01:05:22.268723 kubelet[1948]: I0813 01:05:22.268626 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=2.268603896 podStartE2EDuration="2.268603896s" podCreationTimestamp="2025-08-13 01:05:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:22.259555363 +0000 UTC m=+1.297115021" watchObservedRunningTime="2025-08-13 01:05:22.268603896 +0000 UTC m=+1.306163554" Aug 13 01:05:23.058250 kubelet[1948]: E0813 01:05:23.058209 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:23.671894 sudo[1315]: pam_unix(sudo:session): session closed for user root Aug 13 01:05:23.673372 sshd[1312]: pam_unix(sshd:session): session closed for user core Aug 13 01:05:23.675962 systemd[1]: sshd@6-10.0.0.132:22-10.0.0.1:50314.service: Deactivated successfully. Aug 13 01:05:23.676773 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 01:05:23.676914 systemd[1]: session-7.scope: Consumed 4.699s CPU time. Aug 13 01:05:23.677617 systemd-logind[1195]: Session 7 logged out. Waiting for processes to exit. Aug 13 01:05:23.678423 systemd-logind[1195]: Removed session 7. Aug 13 01:05:24.059677 kubelet[1948]: E0813 01:05:24.059633 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:25.787392 kubelet[1948]: I0813 01:05:25.787349 1948 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 01:05:25.788128 kubelet[1948]: I0813 01:05:25.787883 1948 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 01:05:25.788173 env[1207]: time="2025-08-13T01:05:25.787708099Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 01:05:26.814328 kubelet[1948]: W0813 01:05:26.814282 1948 reflector.go:561] object-"kube-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object Aug 13 01:05:26.814874 kubelet[1948]: E0813 01:05:26.814842 1948 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Aug 13 01:05:26.814962 kubelet[1948]: W0813 01:05:26.814717 1948 reflector.go:561] object-"kube-system"/"kube-proxy": failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object Aug 13 01:05:26.815073 kubelet[1948]: E0813 01:05:26.815049 1948 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"kube-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-proxy\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Aug 13 01:05:26.819877 systemd[1]: Created slice kubepods-besteffort-pod2bf28fb4_6187_4497_bfa7_8e66db955672.slice. Aug 13 01:05:26.832282 systemd[1]: Created slice kubepods-burstable-poda903bf21_d1fd_4b11_b47b_bf64f5bb48a6.slice. Aug 13 01:05:26.872079 kubelet[1948]: I0813 01:05:26.872028 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sql4l\" (UniqueName: \"kubernetes.io/projected/2bf28fb4-6187-4497-bfa7-8e66db955672-kube-api-access-sql4l\") pod \"kube-proxy-75r4n\" (UID: \"2bf28fb4-6187-4497-bfa7-8e66db955672\") " pod="kube-system/kube-proxy-75r4n" Aug 13 01:05:26.872471 kubelet[1948]: I0813 01:05:26.872446 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-kernel\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.872631 kubelet[1948]: I0813 01:05:26.872613 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2bf28fb4-6187-4497-bfa7-8e66db955672-xtables-lock\") pod \"kube-proxy-75r4n\" (UID: \"2bf28fb4-6187-4497-bfa7-8e66db955672\") " pod="kube-system/kube-proxy-75r4n" Aug 13 01:05:26.872770 kubelet[1948]: I0813 01:05:26.872750 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-etc-cni-netd\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.872892 kubelet[1948]: I0813 01:05:26.872871 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-net\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873023 kubelet[1948]: I0813 01:05:26.873002 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2bf28fb4-6187-4497-bfa7-8e66db955672-kube-proxy\") pod \"kube-proxy-75r4n\" (UID: \"2bf28fb4-6187-4497-bfa7-8e66db955672\") " pod="kube-system/kube-proxy-75r4n" Aug 13 01:05:26.873149 kubelet[1948]: I0813 01:05:26.873125 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-cgroup\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873287 kubelet[1948]: I0813 01:05:26.873249 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hubble-tls\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873416 kubelet[1948]: I0813 01:05:26.873397 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-bpf-maps\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873525 kubelet[1948]: I0813 01:05:26.873504 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cni-path\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873658 kubelet[1948]: I0813 01:05:26.873637 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zkdj\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-kube-api-access-5zkdj\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873772 kubelet[1948]: I0813 01:05:26.873752 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-xtables-lock\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.873887 kubelet[1948]: I0813 01:05:26.873869 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hostproc\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.874005 kubelet[1948]: I0813 01:05:26.873975 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-lib-modules\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.874125 kubelet[1948]: I0813 01:05:26.874102 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-config-path\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.874268 kubelet[1948]: I0813 01:05:26.874232 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2bf28fb4-6187-4497-bfa7-8e66db955672-lib-modules\") pod \"kube-proxy-75r4n\" (UID: \"2bf28fb4-6187-4497-bfa7-8e66db955672\") " pod="kube-system/kube-proxy-75r4n" Aug 13 01:05:26.874375 kubelet[1948]: I0813 01:05:26.874358 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-run\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.874472 kubelet[1948]: I0813 01:05:26.874455 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-clustermesh-secrets\") pod \"cilium-c8fc8\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " pod="kube-system/cilium-c8fc8" Aug 13 01:05:26.882470 systemd[1]: Created slice kubepods-besteffort-podc0e2881b_250b_43b5_aa9b_437c0aa094c6.slice. Aug 13 01:05:26.975565 kubelet[1948]: I0813 01:05:26.975510 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0e2881b-250b-43b5-aa9b-437c0aa094c6-cilium-config-path\") pod \"cilium-operator-5d85765b45-qwm68\" (UID: \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\") " pod="kube-system/cilium-operator-5d85765b45-qwm68" Aug 13 01:05:26.975565 kubelet[1948]: I0813 01:05:26.975557 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjpm\" (UniqueName: \"kubernetes.io/projected/c0e2881b-250b-43b5-aa9b-437c0aa094c6-kube-api-access-2tjpm\") pod \"cilium-operator-5d85765b45-qwm68\" (UID: \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\") " pod="kube-system/cilium-operator-5d85765b45-qwm68" Aug 13 01:05:26.976028 kubelet[1948]: I0813 01:05:26.975983 1948 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Aug 13 01:05:27.160615 kubelet[1948]: E0813 01:05:27.160465 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:27.790171 kubelet[1948]: E0813 01:05:27.790136 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:27.790847 env[1207]: time="2025-08-13T01:05:27.790787355Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-qwm68,Uid:c0e2881b-250b-43b5-aa9b-437c0aa094c6,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:27.976535 kubelet[1948]: E0813 01:05:27.976483 1948 configmap.go:193] Couldn't get configMap kube-system/kube-proxy: failed to sync configmap cache: timed out waiting for the condition Aug 13 01:05:27.976941 kubelet[1948]: E0813 01:05:27.976617 1948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2bf28fb4-6187-4497-bfa7-8e66db955672-kube-proxy podName:2bf28fb4-6187-4497-bfa7-8e66db955672 nodeName:}" failed. No retries permitted until 2025-08-13 01:05:28.476596883 +0000 UTC m=+7.514156541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/2bf28fb4-6187-4497-bfa7-8e66db955672-kube-proxy") pod "kube-proxy-75r4n" (UID: "2bf28fb4-6187-4497-bfa7-8e66db955672") : failed to sync configmap cache: timed out waiting for the condition Aug 13 01:05:28.038191 kubelet[1948]: E0813 01:05:28.038140 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:28.038612 env[1207]: time="2025-08-13T01:05:28.038560016Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c8fc8,Uid:a903bf21-d1fd-4b11-b47b-bf64f5bb48a6,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:28.065471 kubelet[1948]: E0813 01:05:28.065353 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:28.089509 env[1207]: time="2025-08-13T01:05:28.089433917Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:28.089509 env[1207]: time="2025-08-13T01:05:28.089473432Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:28.089509 env[1207]: time="2025-08-13T01:05:28.089482729Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:28.089759 env[1207]: time="2025-08-13T01:05:28.089628263Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8 pid=2048 runtime=io.containerd.runc.v2 Aug 13 01:05:28.092312 env[1207]: time="2025-08-13T01:05:28.090665781Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:28.092312 env[1207]: time="2025-08-13T01:05:28.090698272Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:28.092312 env[1207]: time="2025-08-13T01:05:28.090712108Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:28.092312 env[1207]: time="2025-08-13T01:05:28.090885515Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f pid=2049 runtime=io.containerd.runc.v2 Aug 13 01:05:28.107226 systemd[1]: Started cri-containerd-be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f.scope. Aug 13 01:05:28.111514 systemd[1]: Started cri-containerd-7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8.scope. Aug 13 01:05:28.135527 env[1207]: time="2025-08-13T01:05:28.135464239Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c8fc8,Uid:a903bf21-d1fd-4b11-b47b-bf64f5bb48a6,Namespace:kube-system,Attempt:0,} returns sandbox id \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\"" Aug 13 01:05:28.136909 kubelet[1948]: E0813 01:05:28.136884 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:28.138621 env[1207]: time="2025-08-13T01:05:28.138587451Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Aug 13 01:05:28.159816 env[1207]: time="2025-08-13T01:05:28.158855474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-qwm68,Uid:c0e2881b-250b-43b5-aa9b-437c0aa094c6,Namespace:kube-system,Attempt:0,} returns sandbox id \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\"" Aug 13 01:05:28.159988 kubelet[1948]: E0813 01:05:28.159643 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:28.629495 kubelet[1948]: E0813 01:05:28.629437 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:28.630203 env[1207]: time="2025-08-13T01:05:28.630133606Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-75r4n,Uid:2bf28fb4-6187-4497-bfa7-8e66db955672,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:29.926957 env[1207]: time="2025-08-13T01:05:29.926561986Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:29.926957 env[1207]: time="2025-08-13T01:05:29.926654661Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:29.926957 env[1207]: time="2025-08-13T01:05:29.926686271Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:29.926957 env[1207]: time="2025-08-13T01:05:29.926825112Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9d8eb5be9a8e618677cf10b8204318e3638b661cb20d1eee25394eda0da4bef3 pid=2122 runtime=io.containerd.runc.v2 Aug 13 01:05:29.984877 systemd[1]: Started cri-containerd-9d8eb5be9a8e618677cf10b8204318e3638b661cb20d1eee25394eda0da4bef3.scope. Aug 13 01:05:30.066724 env[1207]: time="2025-08-13T01:05:30.062868249Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-75r4n,Uid:2bf28fb4-6187-4497-bfa7-8e66db955672,Namespace:kube-system,Attempt:0,} returns sandbox id \"9d8eb5be9a8e618677cf10b8204318e3638b661cb20d1eee25394eda0da4bef3\"" Aug 13 01:05:30.066955 kubelet[1948]: E0813 01:05:30.065970 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:30.121652 env[1207]: time="2025-08-13T01:05:30.119178337Z" level=info msg="CreateContainer within sandbox \"9d8eb5be9a8e618677cf10b8204318e3638b661cb20d1eee25394eda0da4bef3\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 01:05:30.155643 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount810906053.mount: Deactivated successfully. Aug 13 01:05:30.185410 env[1207]: time="2025-08-13T01:05:30.182603529Z" level=info msg="CreateContainer within sandbox \"9d8eb5be9a8e618677cf10b8204318e3638b661cb20d1eee25394eda0da4bef3\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"1ef364c54e86d231321c511d62f28b742d6a5c0b2139ad5d55fb3d629b68f1a3\"" Aug 13 01:05:30.185410 env[1207]: time="2025-08-13T01:05:30.184067799Z" level=info msg="StartContainer for \"1ef364c54e86d231321c511d62f28b742d6a5c0b2139ad5d55fb3d629b68f1a3\"" Aug 13 01:05:30.247493 systemd[1]: Started cri-containerd-1ef364c54e86d231321c511d62f28b742d6a5c0b2139ad5d55fb3d629b68f1a3.scope. Aug 13 01:05:30.286298 env[1207]: time="2025-08-13T01:05:30.282751647Z" level=info msg="StartContainer for \"1ef364c54e86d231321c511d62f28b742d6a5c0b2139ad5d55fb3d629b68f1a3\" returns successfully" Aug 13 01:05:31.106481 kubelet[1948]: E0813 01:05:31.106428 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:31.125493 kubelet[1948]: E0813 01:05:31.125440 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:31.125493 kubelet[1948]: E0813 01:05:31.125506 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:31.172272 kubelet[1948]: I0813 01:05:31.172168 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-75r4n" podStartSLOduration=5.172124049 podStartE2EDuration="5.172124049s" podCreationTimestamp="2025-08-13 01:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:31.171942086 +0000 UTC m=+10.209501744" watchObservedRunningTime="2025-08-13 01:05:31.172124049 +0000 UTC m=+10.209683707" Aug 13 01:05:32.001277 kubelet[1948]: E0813 01:05:32.000157 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:32.126946 kubelet[1948]: E0813 01:05:32.126883 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:32.127601 kubelet[1948]: E0813 01:05:32.127038 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:36.850921 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1291862464.mount: Deactivated successfully. Aug 13 01:05:42.885653 env[1207]: time="2025-08-13T01:05:42.885586082Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:42.929835 env[1207]: time="2025-08-13T01:05:42.929746674Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:42.970314 env[1207]: time="2025-08-13T01:05:42.970227790Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:42.971083 env[1207]: time="2025-08-13T01:05:42.971025590Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Aug 13 01:05:42.972513 env[1207]: time="2025-08-13T01:05:42.972474834Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Aug 13 01:05:42.973848 env[1207]: time="2025-08-13T01:05:42.973812960Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 01:05:43.150001 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount848169538.mount: Deactivated successfully. Aug 13 01:05:43.263306 env[1207]: time="2025-08-13T01:05:43.263222772Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\"" Aug 13 01:05:43.263743 env[1207]: time="2025-08-13T01:05:43.263704938Z" level=info msg="StartContainer for \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\"" Aug 13 01:05:43.280372 systemd[1]: Started cri-containerd-616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2.scope. Aug 13 01:05:43.338545 systemd[1]: cri-containerd-616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2.scope: Deactivated successfully. Aug 13 01:05:43.413068 env[1207]: time="2025-08-13T01:05:43.412918885Z" level=info msg="StartContainer for \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\" returns successfully" Aug 13 01:05:44.148056 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2-rootfs.mount: Deactivated successfully. Aug 13 01:05:44.223444 kubelet[1948]: E0813 01:05:44.223386 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:44.777802 env[1207]: time="2025-08-13T01:05:44.777722126Z" level=info msg="shim disconnected" id=616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2 Aug 13 01:05:44.777802 env[1207]: time="2025-08-13T01:05:44.777787108Z" level=warning msg="cleaning up after shim disconnected" id=616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2 namespace=k8s.io Aug 13 01:05:44.777802 env[1207]: time="2025-08-13T01:05:44.777796457Z" level=info msg="cleaning up dead shim" Aug 13 01:05:44.785827 env[1207]: time="2025-08-13T01:05:44.785769169Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:05:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2378 runtime=io.containerd.runc.v2\n" Aug 13 01:05:45.225806 kubelet[1948]: E0813 01:05:45.225769 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:45.227184 env[1207]: time="2025-08-13T01:05:45.227126334Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 01:05:46.698019 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1220614987.mount: Deactivated successfully. Aug 13 01:05:46.964950 env[1207]: time="2025-08-13T01:05:46.964801009Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\"" Aug 13 01:05:46.965449 env[1207]: time="2025-08-13T01:05:46.965360531Z" level=info msg="StartContainer for \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\"" Aug 13 01:05:46.983322 systemd[1]: Started cri-containerd-570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053.scope. Aug 13 01:05:47.019319 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 01:05:47.019522 systemd[1]: Stopped systemd-sysctl.service. Aug 13 01:05:47.021277 systemd[1]: Stopping systemd-sysctl.service... Aug 13 01:05:47.022703 systemd[1]: Starting systemd-sysctl.service... Aug 13 01:05:47.024907 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Aug 13 01:05:47.025654 systemd[1]: cri-containerd-570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053.scope: Deactivated successfully. Aug 13 01:05:47.030924 systemd[1]: Finished systemd-sysctl.service. Aug 13 01:05:47.372788 env[1207]: time="2025-08-13T01:05:47.372340780Z" level=info msg="StartContainer for \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\" returns successfully" Aug 13 01:05:47.410329 env[1207]: time="2025-08-13T01:05:47.410232037Z" level=info msg="shim disconnected" id=570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053 Aug 13 01:05:47.410329 env[1207]: time="2025-08-13T01:05:47.410323570Z" level=warning msg="cleaning up after shim disconnected" id=570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053 namespace=k8s.io Aug 13 01:05:47.410329 env[1207]: time="2025-08-13T01:05:47.410338578Z" level=info msg="cleaning up dead shim" Aug 13 01:05:47.418455 env[1207]: time="2025-08-13T01:05:47.418372960Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:05:47Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2441 runtime=io.containerd.runc.v2\n" Aug 13 01:05:47.695491 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053-rootfs.mount: Deactivated successfully. Aug 13 01:05:48.271199 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2446481709.mount: Deactivated successfully. Aug 13 01:05:48.379096 kubelet[1948]: E0813 01:05:48.379042 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:48.382139 env[1207]: time="2025-08-13T01:05:48.382069133Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 01:05:48.406822 env[1207]: time="2025-08-13T01:05:48.406728289Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\"" Aug 13 01:05:48.407548 env[1207]: time="2025-08-13T01:05:48.407472598Z" level=info msg="StartContainer for \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\"" Aug 13 01:05:48.423800 systemd[1]: Started cri-containerd-6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6.scope. Aug 13 01:05:48.453852 systemd[1]: cri-containerd-6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6.scope: Deactivated successfully. Aug 13 01:05:48.455852 env[1207]: time="2025-08-13T01:05:48.455801842Z" level=info msg="StartContainer for \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\" returns successfully" Aug 13 01:05:48.531390 env[1207]: time="2025-08-13T01:05:48.531275590Z" level=info msg="shim disconnected" id=6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6 Aug 13 01:05:48.531643 env[1207]: time="2025-08-13T01:05:48.531595371Z" level=warning msg="cleaning up after shim disconnected" id=6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6 namespace=k8s.io Aug 13 01:05:48.531643 env[1207]: time="2025-08-13T01:05:48.531617093Z" level=info msg="cleaning up dead shim" Aug 13 01:05:48.541500 env[1207]: time="2025-08-13T01:05:48.541437047Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:05:48Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2497 runtime=io.containerd.runc.v2\n" Aug 13 01:05:49.382544 kubelet[1948]: E0813 01:05:49.382491 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:49.388826 env[1207]: time="2025-08-13T01:05:49.388771186Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 01:05:49.669185 env[1207]: time="2025-08-13T01:05:49.669037501Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:49.837491 env[1207]: time="2025-08-13T01:05:49.837411495Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\"" Aug 13 01:05:49.837834 env[1207]: time="2025-08-13T01:05:49.837789775Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:49.838445 env[1207]: time="2025-08-13T01:05:49.838369755Z" level=info msg="StartContainer for \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\"" Aug 13 01:05:49.858769 env[1207]: time="2025-08-13T01:05:49.858072038Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 01:05:49.858769 env[1207]: time="2025-08-13T01:05:49.858298152Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Aug 13 01:05:49.858968 systemd[1]: run-containerd-runc-k8s.io-2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0-runc.Yqma8m.mount: Deactivated successfully. Aug 13 01:05:49.860812 systemd[1]: Started cri-containerd-2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0.scope. Aug 13 01:05:49.862569 env[1207]: time="2025-08-13T01:05:49.862521323Z" level=info msg="CreateContainer within sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Aug 13 01:05:49.886176 env[1207]: time="2025-08-13T01:05:49.886046194Z" level=info msg="CreateContainer within sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\"" Aug 13 01:05:49.886286 systemd[1]: cri-containerd-2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0.scope: Deactivated successfully. Aug 13 01:05:49.888054 env[1207]: time="2025-08-13T01:05:49.886942628Z" level=info msg="StartContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\"" Aug 13 01:05:49.888054 env[1207]: time="2025-08-13T01:05:49.887562662Z" level=info msg="StartContainer for \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\" returns successfully" Aug 13 01:05:49.903317 systemd[1]: Started cri-containerd-2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f.scope. Aug 13 01:05:50.106466 env[1207]: time="2025-08-13T01:05:50.106350147Z" level=info msg="StartContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" returns successfully" Aug 13 01:05:50.106930 env[1207]: time="2025-08-13T01:05:50.106881604Z" level=info msg="shim disconnected" id=2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0 Aug 13 01:05:50.106930 env[1207]: time="2025-08-13T01:05:50.106920338Z" level=warning msg="cleaning up after shim disconnected" id=2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0 namespace=k8s.io Aug 13 01:05:50.106930 env[1207]: time="2025-08-13T01:05:50.106930557Z" level=info msg="cleaning up dead shim" Aug 13 01:05:50.124004 env[1207]: time="2025-08-13T01:05:50.122830494Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:05:50Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2591 runtime=io.containerd.runc.v2\n" Aug 13 01:05:50.385423 kubelet[1948]: E0813 01:05:50.385304 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:50.388295 kubelet[1948]: E0813 01:05:50.388248 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:50.390027 env[1207]: time="2025-08-13T01:05:50.389982278Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 01:05:50.432396 env[1207]: time="2025-08-13T01:05:50.432323178Z" level=info msg="CreateContainer within sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\"" Aug 13 01:05:50.432899 env[1207]: time="2025-08-13T01:05:50.432860116Z" level=info msg="StartContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\"" Aug 13 01:05:50.453017 kubelet[1948]: I0813 01:05:50.452856 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-qwm68" podStartSLOduration=2.753571423 podStartE2EDuration="24.452837291s" podCreationTimestamp="2025-08-13 01:05:26 +0000 UTC" firstStartedPulling="2025-08-13 01:05:28.160884823 +0000 UTC m=+7.198444481" lastFinishedPulling="2025-08-13 01:05:49.860150681 +0000 UTC m=+28.897710349" observedRunningTime="2025-08-13 01:05:50.437890743 +0000 UTC m=+29.475450391" watchObservedRunningTime="2025-08-13 01:05:50.452837291 +0000 UTC m=+29.490396949" Aug 13 01:05:50.461155 systemd[1]: Started cri-containerd-ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864.scope. Aug 13 01:05:50.512393 env[1207]: time="2025-08-13T01:05:50.512347482Z" level=info msg="StartContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" returns successfully" Aug 13 01:05:50.769107 kubelet[1948]: I0813 01:05:50.769035 1948 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Aug 13 01:05:50.810178 systemd[1]: Created slice kubepods-burstable-podb6525e0c_f24b_42f2_af17_d0236a33377a.slice. Aug 13 01:05:50.817381 systemd[1]: Created slice kubepods-burstable-pod0fe0f047_443a_48e7_bffa_81ce0432c16d.slice. Aug 13 01:05:50.850623 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3905044653.mount: Deactivated successfully. Aug 13 01:05:50.850726 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0-rootfs.mount: Deactivated successfully. Aug 13 01:05:50.919399 kubelet[1948]: I0813 01:05:50.919353 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6525e0c-f24b-42f2-af17-d0236a33377a-config-volume\") pod \"coredns-7c65d6cfc9-f4sq5\" (UID: \"b6525e0c-f24b-42f2-af17-d0236a33377a\") " pod="kube-system/coredns-7c65d6cfc9-f4sq5" Aug 13 01:05:50.919399 kubelet[1948]: I0813 01:05:50.919395 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe0f047-443a-48e7-bffa-81ce0432c16d-config-volume\") pod \"coredns-7c65d6cfc9-xrpx8\" (UID: \"0fe0f047-443a-48e7-bffa-81ce0432c16d\") " pod="kube-system/coredns-7c65d6cfc9-xrpx8" Aug 13 01:05:50.919627 kubelet[1948]: I0813 01:05:50.919412 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2cbr\" (UniqueName: \"kubernetes.io/projected/b6525e0c-f24b-42f2-af17-d0236a33377a-kube-api-access-q2cbr\") pod \"coredns-7c65d6cfc9-f4sq5\" (UID: \"b6525e0c-f24b-42f2-af17-d0236a33377a\") " pod="kube-system/coredns-7c65d6cfc9-f4sq5" Aug 13 01:05:50.919627 kubelet[1948]: I0813 01:05:50.919431 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dmb6\" (UniqueName: \"kubernetes.io/projected/0fe0f047-443a-48e7-bffa-81ce0432c16d-kube-api-access-8dmb6\") pod \"coredns-7c65d6cfc9-xrpx8\" (UID: \"0fe0f047-443a-48e7-bffa-81ce0432c16d\") " pod="kube-system/coredns-7c65d6cfc9-xrpx8" Aug 13 01:05:51.115776 kubelet[1948]: E0813 01:05:51.115615 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:51.116383 env[1207]: time="2025-08-13T01:05:51.116333413Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-f4sq5,Uid:b6525e0c-f24b-42f2-af17-d0236a33377a,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:51.120242 kubelet[1948]: E0813 01:05:51.120188 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:51.120883 env[1207]: time="2025-08-13T01:05:51.120818153Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-xrpx8,Uid:0fe0f047-443a-48e7-bffa-81ce0432c16d,Namespace:kube-system,Attempt:0,}" Aug 13 01:05:51.393807 kubelet[1948]: E0813 01:05:51.393377 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:51.393807 kubelet[1948]: E0813 01:05:51.393528 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:51.415769 kubelet[1948]: I0813 01:05:51.414801 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-c8fc8" podStartSLOduration=10.580569316 podStartE2EDuration="25.4147793s" podCreationTimestamp="2025-08-13 01:05:26 +0000 UTC" firstStartedPulling="2025-08-13 01:05:28.137975467 +0000 UTC m=+7.175535125" lastFinishedPulling="2025-08-13 01:05:42.972185451 +0000 UTC m=+22.009745109" observedRunningTime="2025-08-13 01:05:51.41386877 +0000 UTC m=+30.451428428" watchObservedRunningTime="2025-08-13 01:05:51.4147793 +0000 UTC m=+30.452338958" Aug 13 01:05:52.395309 kubelet[1948]: E0813 01:05:52.395251 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:52.560429 systemd-networkd[1020]: cilium_host: Link UP Aug 13 01:05:52.566382 systemd-networkd[1020]: cilium_net: Link UP Aug 13 01:05:52.575901 systemd-networkd[1020]: cilium_net: Gained carrier Aug 13 01:05:52.605359 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Aug 13 01:05:52.605537 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Aug 13 01:05:52.605707 systemd-networkd[1020]: cilium_host: Gained carrier Aug 13 01:05:52.699859 systemd-networkd[1020]: cilium_vxlan: Link UP Aug 13 01:05:52.699870 systemd-networkd[1020]: cilium_vxlan: Gained carrier Aug 13 01:05:52.703538 systemd-networkd[1020]: cilium_host: Gained IPv6LL Aug 13 01:05:52.893300 kernel: NET: Registered PF_ALG protocol family Aug 13 01:05:53.396806 kubelet[1948]: E0813 01:05:53.396763 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:53.445226 systemd-networkd[1020]: lxc_health: Link UP Aug 13 01:05:53.457297 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Aug 13 01:05:53.457483 systemd-networkd[1020]: lxc_health: Gained carrier Aug 13 01:05:53.559461 systemd-networkd[1020]: cilium_net: Gained IPv6LL Aug 13 01:05:53.711444 systemd-networkd[1020]: lxc9655a263bf2b: Link UP Aug 13 01:05:53.744620 kernel: eth0: renamed from tmpa44c5 Aug 13 01:05:53.758298 kernel: eth0: renamed from tmp38ae8 Aug 13 01:05:53.758489 systemd-networkd[1020]: lxc0bfdd9f2e035: Link UP Aug 13 01:05:53.766319 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 01:05:53.766415 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc9655a263bf2b: link becomes ready Aug 13 01:05:53.766448 systemd-networkd[1020]: lxc9655a263bf2b: Gained carrier Aug 13 01:05:53.769292 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc0bfdd9f2e035: link becomes ready Aug 13 01:05:53.772126 systemd-networkd[1020]: lxc0bfdd9f2e035: Gained carrier Aug 13 01:05:53.879400 systemd-networkd[1020]: cilium_vxlan: Gained IPv6LL Aug 13 01:05:54.399013 kubelet[1948]: E0813 01:05:54.398949 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:55.159439 systemd-networkd[1020]: lxc9655a263bf2b: Gained IPv6LL Aug 13 01:05:55.400936 kubelet[1948]: E0813 01:05:55.400883 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:55.415390 systemd-networkd[1020]: lxc_health: Gained IPv6LL Aug 13 01:05:55.479420 systemd-networkd[1020]: lxc0bfdd9f2e035: Gained IPv6LL Aug 13 01:05:56.401990 kubelet[1948]: E0813 01:05:56.401956 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:56.443050 systemd[1]: Started sshd@7-10.0.0.132:22-10.0.0.1:52696.service. Aug 13 01:05:56.527276 sshd[3145]: Accepted publickey for core from 10.0.0.1 port 52696 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:05:56.528598 sshd[3145]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:05:56.532580 systemd-logind[1195]: New session 8 of user core. Aug 13 01:05:56.533648 systemd[1]: Started session-8.scope. Aug 13 01:05:56.709537 sshd[3145]: pam_unix(sshd:session): session closed for user core Aug 13 01:05:56.712426 systemd[1]: sshd@7-10.0.0.132:22-10.0.0.1:52696.service: Deactivated successfully. Aug 13 01:05:56.713321 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 01:05:56.714032 systemd-logind[1195]: Session 8 logged out. Waiting for processes to exit. Aug 13 01:05:56.714886 systemd-logind[1195]: Removed session 8. Aug 13 01:05:57.645142 env[1207]: time="2025-08-13T01:05:57.644155125Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:57.645142 env[1207]: time="2025-08-13T01:05:57.644202344Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:57.645142 env[1207]: time="2025-08-13T01:05:57.644212783Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:57.645142 env[1207]: time="2025-08-13T01:05:57.644467000Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/a44c5f214bcbf0ae0ab6c1691f132577f7d73e3d34f4e1441fedbad45e054853 pid=3172 runtime=io.containerd.runc.v2 Aug 13 01:05:57.651767 env[1207]: time="2025-08-13T01:05:57.647967330Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:05:57.651767 env[1207]: time="2025-08-13T01:05:57.648027753Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:05:57.651767 env[1207]: time="2025-08-13T01:05:57.648051157Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:05:57.651767 env[1207]: time="2025-08-13T01:05:57.648206609Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe pid=3189 runtime=io.containerd.runc.v2 Aug 13 01:05:57.664931 systemd[1]: run-containerd-runc-k8s.io-38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe-runc.1KunZe.mount: Deactivated successfully. Aug 13 01:05:57.667972 systemd[1]: Started cri-containerd-38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe.scope. Aug 13 01:05:57.668813 systemd[1]: Started cri-containerd-a44c5f214bcbf0ae0ab6c1691f132577f7d73e3d34f4e1441fedbad45e054853.scope. Aug 13 01:05:57.680868 systemd-resolved[1142]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Aug 13 01:05:57.682297 systemd-resolved[1142]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Aug 13 01:05:57.707582 env[1207]: time="2025-08-13T01:05:57.707523035Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-xrpx8,Uid:0fe0f047-443a-48e7-bffa-81ce0432c16d,Namespace:kube-system,Attempt:0,} returns sandbox id \"38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe\"" Aug 13 01:05:57.708327 kubelet[1948]: E0813 01:05:57.708293 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:57.710822 env[1207]: time="2025-08-13T01:05:57.710751845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-f4sq5,Uid:b6525e0c-f24b-42f2-af17-d0236a33377a,Namespace:kube-system,Attempt:0,} returns sandbox id \"a44c5f214bcbf0ae0ab6c1691f132577f7d73e3d34f4e1441fedbad45e054853\"" Aug 13 01:05:57.711761 kubelet[1948]: E0813 01:05:57.711607 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:57.716795 env[1207]: time="2025-08-13T01:05:57.716764109Z" level=info msg="CreateContainer within sandbox \"38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 01:05:57.717168 env[1207]: time="2025-08-13T01:05:57.717017144Z" level=info msg="CreateContainer within sandbox \"a44c5f214bcbf0ae0ab6c1691f132577f7d73e3d34f4e1441fedbad45e054853\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 01:05:57.756963 env[1207]: time="2025-08-13T01:05:57.756901639Z" level=info msg="CreateContainer within sandbox \"a44c5f214bcbf0ae0ab6c1691f132577f7d73e3d34f4e1441fedbad45e054853\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c7437bac77d9e71789a9569314d378e78d2a04cdf86a7d9d23d2e4d3ba16dc15\"" Aug 13 01:05:57.758720 env[1207]: time="2025-08-13T01:05:57.758688904Z" level=info msg="StartContainer for \"c7437bac77d9e71789a9569314d378e78d2a04cdf86a7d9d23d2e4d3ba16dc15\"" Aug 13 01:05:57.758977 env[1207]: time="2025-08-13T01:05:57.758915489Z" level=info msg="CreateContainer within sandbox \"38ae87f4a3605805c43533cbcf7055c4cfba1a7d3660b4da56bf492dde0f4afe\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"be8c3215b53c1ff811710842005980e320a64b2bbded9c7a10a127c6eb4eab21\"" Aug 13 01:05:57.759636 env[1207]: time="2025-08-13T01:05:57.759607257Z" level=info msg="StartContainer for \"be8c3215b53c1ff811710842005980e320a64b2bbded9c7a10a127c6eb4eab21\"" Aug 13 01:05:57.779972 systemd[1]: Started cri-containerd-c7437bac77d9e71789a9569314d378e78d2a04cdf86a7d9d23d2e4d3ba16dc15.scope. Aug 13 01:05:57.785624 systemd[1]: Started cri-containerd-be8c3215b53c1ff811710842005980e320a64b2bbded9c7a10a127c6eb4eab21.scope. Aug 13 01:05:57.813642 env[1207]: time="2025-08-13T01:05:57.813572180Z" level=info msg="StartContainer for \"c7437bac77d9e71789a9569314d378e78d2a04cdf86a7d9d23d2e4d3ba16dc15\" returns successfully" Aug 13 01:05:57.820884 env[1207]: time="2025-08-13T01:05:57.820827898Z" level=info msg="StartContainer for \"be8c3215b53c1ff811710842005980e320a64b2bbded9c7a10a127c6eb4eab21\" returns successfully" Aug 13 01:05:58.413848 kubelet[1948]: E0813 01:05:58.413794 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:58.414085 kubelet[1948]: E0813 01:05:58.413863 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:05:58.437991 kubelet[1948]: I0813 01:05:58.437918 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-f4sq5" podStartSLOduration=32.437894502 podStartE2EDuration="32.437894502s" podCreationTimestamp="2025-08-13 01:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:58.427368247 +0000 UTC m=+37.464927905" watchObservedRunningTime="2025-08-13 01:05:58.437894502 +0000 UTC m=+37.475454160" Aug 13 01:05:58.438234 kubelet[1948]: I0813 01:05:58.438014 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-xrpx8" podStartSLOduration=32.438010058 podStartE2EDuration="32.438010058s" podCreationTimestamp="2025-08-13 01:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:05:58.437390165 +0000 UTC m=+37.474949843" watchObservedRunningTime="2025-08-13 01:05:58.438010058 +0000 UTC m=+37.475569716" Aug 13 01:06:01.116731 kubelet[1948]: E0813 01:06:01.116632 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:01.120827 kubelet[1948]: E0813 01:06:01.120776 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:01.418306 kubelet[1948]: E0813 01:06:01.418144 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:01.418306 kubelet[1948]: E0813 01:06:01.418233 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:01.714090 systemd[1]: Started sshd@8-10.0.0.132:22-10.0.0.1:46406.service. Aug 13 01:06:01.745742 sshd[3337]: Accepted publickey for core from 10.0.0.1 port 46406 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:01.746984 sshd[3337]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:01.750708 systemd-logind[1195]: New session 9 of user core. Aug 13 01:06:01.751755 systemd[1]: Started session-9.scope. Aug 13 01:06:01.875668 sshd[3337]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:01.878724 systemd[1]: sshd@8-10.0.0.132:22-10.0.0.1:46406.service: Deactivated successfully. Aug 13 01:06:01.879490 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 01:06:01.880150 systemd-logind[1195]: Session 9 logged out. Waiting for processes to exit. Aug 13 01:06:01.880896 systemd-logind[1195]: Removed session 9. Aug 13 01:06:06.880424 systemd[1]: Started sshd@9-10.0.0.132:22-10.0.0.1:46418.service. Aug 13 01:06:06.912746 sshd[3351]: Accepted publickey for core from 10.0.0.1 port 46418 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:06.914117 sshd[3351]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:06.917524 systemd-logind[1195]: New session 10 of user core. Aug 13 01:06:06.918392 systemd[1]: Started session-10.scope. Aug 13 01:06:07.073049 sshd[3351]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:07.075883 systemd[1]: sshd@9-10.0.0.132:22-10.0.0.1:46418.service: Deactivated successfully. Aug 13 01:06:07.076656 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 01:06:07.077245 systemd-logind[1195]: Session 10 logged out. Waiting for processes to exit. Aug 13 01:06:07.077914 systemd-logind[1195]: Removed session 10. Aug 13 01:06:12.078251 systemd[1]: Started sshd@10-10.0.0.132:22-10.0.0.1:58806.service. Aug 13 01:06:12.111175 sshd[3365]: Accepted publickey for core from 10.0.0.1 port 58806 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:12.112577 sshd[3365]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:12.116599 systemd-logind[1195]: New session 11 of user core. Aug 13 01:06:12.117469 systemd[1]: Started session-11.scope. Aug 13 01:06:12.225149 sshd[3365]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:12.227977 systemd[1]: sshd@10-10.0.0.132:22-10.0.0.1:58806.service: Deactivated successfully. Aug 13 01:06:12.228695 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 01:06:12.229387 systemd-logind[1195]: Session 11 logged out. Waiting for processes to exit. Aug 13 01:06:12.230217 systemd-logind[1195]: Removed session 11. Aug 13 01:06:17.231448 systemd[1]: Started sshd@11-10.0.0.132:22-10.0.0.1:58810.service. Aug 13 01:06:17.262048 sshd[3379]: Accepted publickey for core from 10.0.0.1 port 58810 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:17.263436 sshd[3379]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:17.267107 systemd-logind[1195]: New session 12 of user core. Aug 13 01:06:17.268096 systemd[1]: Started session-12.scope. Aug 13 01:06:17.377325 sshd[3379]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:17.381415 systemd[1]: sshd@11-10.0.0.132:22-10.0.0.1:58810.service: Deactivated successfully. Aug 13 01:06:17.382023 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 01:06:17.382695 systemd-logind[1195]: Session 12 logged out. Waiting for processes to exit. Aug 13 01:06:17.384015 systemd[1]: Started sshd@12-10.0.0.132:22-10.0.0.1:58820.service. Aug 13 01:06:17.385173 systemd-logind[1195]: Removed session 12. Aug 13 01:06:17.414921 sshd[3394]: Accepted publickey for core from 10.0.0.1 port 58820 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:17.416599 sshd[3394]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:17.420981 systemd-logind[1195]: New session 13 of user core. Aug 13 01:06:17.421858 systemd[1]: Started session-13.scope. Aug 13 01:06:17.570817 sshd[3394]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:17.575298 systemd[1]: Started sshd@13-10.0.0.132:22-10.0.0.1:58836.service. Aug 13 01:06:17.575932 systemd[1]: sshd@12-10.0.0.132:22-10.0.0.1:58820.service: Deactivated successfully. Aug 13 01:06:17.576748 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 01:06:17.578323 systemd-logind[1195]: Session 13 logged out. Waiting for processes to exit. Aug 13 01:06:17.579371 systemd-logind[1195]: Removed session 13. Aug 13 01:06:17.613878 sshd[3405]: Accepted publickey for core from 10.0.0.1 port 58836 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:17.615122 sshd[3405]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:17.618500 systemd-logind[1195]: New session 14 of user core. Aug 13 01:06:17.619271 systemd[1]: Started session-14.scope. Aug 13 01:06:17.726958 sshd[3405]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:17.729569 systemd[1]: sshd@13-10.0.0.132:22-10.0.0.1:58836.service: Deactivated successfully. Aug 13 01:06:17.730472 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 01:06:17.731117 systemd-logind[1195]: Session 14 logged out. Waiting for processes to exit. Aug 13 01:06:17.731825 systemd-logind[1195]: Removed session 14. Aug 13 01:06:22.732286 systemd[1]: Started sshd@14-10.0.0.132:22-10.0.0.1:39430.service. Aug 13 01:06:22.767760 sshd[3424]: Accepted publickey for core from 10.0.0.1 port 39430 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:22.769465 sshd[3424]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:22.773176 systemd-logind[1195]: New session 15 of user core. Aug 13 01:06:22.774129 systemd[1]: Started session-15.scope. Aug 13 01:06:22.889276 sshd[3424]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:22.891962 systemd[1]: sshd@14-10.0.0.132:22-10.0.0.1:39430.service: Deactivated successfully. Aug 13 01:06:22.892653 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 01:06:22.893387 systemd-logind[1195]: Session 15 logged out. Waiting for processes to exit. Aug 13 01:06:22.894068 systemd-logind[1195]: Removed session 15. Aug 13 01:06:27.893806 systemd[1]: Started sshd@15-10.0.0.132:22-10.0.0.1:39446.service. Aug 13 01:06:27.921766 sshd[3438]: Accepted publickey for core from 10.0.0.1 port 39446 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:27.922712 sshd[3438]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:27.925887 systemd-logind[1195]: New session 16 of user core. Aug 13 01:06:27.926640 systemd[1]: Started session-16.scope. Aug 13 01:06:28.028915 sshd[3438]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:28.031021 systemd[1]: sshd@15-10.0.0.132:22-10.0.0.1:39446.service: Deactivated successfully. Aug 13 01:06:28.031790 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 01:06:28.032318 systemd-logind[1195]: Session 16 logged out. Waiting for processes to exit. Aug 13 01:06:28.032965 systemd-logind[1195]: Removed session 16. Aug 13 01:06:33.033176 systemd[1]: Started sshd@16-10.0.0.132:22-10.0.0.1:50454.service. Aug 13 01:06:33.061370 sshd[3454]: Accepted publickey for core from 10.0.0.1 port 50454 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:33.062386 sshd[3454]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:33.065503 systemd-logind[1195]: New session 17 of user core. Aug 13 01:06:33.066299 systemd[1]: Started session-17.scope. Aug 13 01:06:33.165213 sshd[3454]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:33.167894 systemd[1]: sshd@16-10.0.0.132:22-10.0.0.1:50454.service: Deactivated successfully. Aug 13 01:06:33.168438 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 01:06:33.168948 systemd-logind[1195]: Session 17 logged out. Waiting for processes to exit. Aug 13 01:06:33.169833 systemd[1]: Started sshd@17-10.0.0.132:22-10.0.0.1:50460.service. Aug 13 01:06:33.170589 systemd-logind[1195]: Removed session 17. Aug 13 01:06:33.197560 sshd[3467]: Accepted publickey for core from 10.0.0.1 port 50460 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:33.198474 sshd[3467]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:33.201527 systemd-logind[1195]: New session 18 of user core. Aug 13 01:06:33.202561 systemd[1]: Started session-18.scope. Aug 13 01:06:33.445736 sshd[3467]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:33.448824 systemd[1]: sshd@17-10.0.0.132:22-10.0.0.1:50460.service: Deactivated successfully. Aug 13 01:06:33.449386 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 01:06:33.449946 systemd-logind[1195]: Session 18 logged out. Waiting for processes to exit. Aug 13 01:06:33.451081 systemd[1]: Started sshd@18-10.0.0.132:22-10.0.0.1:50470.service. Aug 13 01:06:33.451995 systemd-logind[1195]: Removed session 18. Aug 13 01:06:33.482224 sshd[3479]: Accepted publickey for core from 10.0.0.1 port 50470 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:33.483279 sshd[3479]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:33.486702 systemd-logind[1195]: New session 19 of user core. Aug 13 01:06:33.487551 systemd[1]: Started session-19.scope. Aug 13 01:06:34.044408 kubelet[1948]: E0813 01:06:34.044365 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:35.357107 sshd[3479]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:35.360317 systemd[1]: sshd@18-10.0.0.132:22-10.0.0.1:50470.service: Deactivated successfully. Aug 13 01:06:35.360834 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 01:06:35.361421 systemd-logind[1195]: Session 19 logged out. Waiting for processes to exit. Aug 13 01:06:35.363189 systemd[1]: Started sshd@19-10.0.0.132:22-10.0.0.1:50472.service. Aug 13 01:06:35.364403 systemd-logind[1195]: Removed session 19. Aug 13 01:06:35.392247 sshd[3498]: Accepted publickey for core from 10.0.0.1 port 50472 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:35.393504 sshd[3498]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:35.396780 systemd-logind[1195]: New session 20 of user core. Aug 13 01:06:35.397717 systemd[1]: Started session-20.scope. Aug 13 01:06:35.805918 sshd[3498]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:35.809733 systemd[1]: Started sshd@20-10.0.0.132:22-10.0.0.1:50474.service. Aug 13 01:06:35.810300 systemd[1]: sshd@19-10.0.0.132:22-10.0.0.1:50472.service: Deactivated successfully. Aug 13 01:06:35.810839 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 01:06:35.811645 systemd-logind[1195]: Session 20 logged out. Waiting for processes to exit. Aug 13 01:06:35.812360 systemd-logind[1195]: Removed session 20. Aug 13 01:06:35.840018 sshd[3508]: Accepted publickey for core from 10.0.0.1 port 50474 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:35.841203 sshd[3508]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:35.844718 systemd-logind[1195]: New session 21 of user core. Aug 13 01:06:35.845735 systemd[1]: Started session-21.scope. Aug 13 01:06:35.982891 sshd[3508]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:35.985176 systemd[1]: sshd@20-10.0.0.132:22-10.0.0.1:50474.service: Deactivated successfully. Aug 13 01:06:35.985842 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 01:06:35.986422 systemd-logind[1195]: Session 21 logged out. Waiting for processes to exit. Aug 13 01:06:35.987162 systemd-logind[1195]: Removed session 21. Aug 13 01:06:37.044646 kubelet[1948]: E0813 01:06:37.044602 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:40.987875 systemd[1]: Started sshd@21-10.0.0.132:22-10.0.0.1:36394.service. Aug 13 01:06:41.016762 sshd[3522]: Accepted publickey for core from 10.0.0.1 port 36394 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:41.017971 sshd[3522]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:41.022333 systemd-logind[1195]: New session 22 of user core. Aug 13 01:06:41.023267 systemd[1]: Started session-22.scope. Aug 13 01:06:41.044429 kubelet[1948]: E0813 01:06:41.044402 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:41.124025 sshd[3522]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:41.126654 systemd[1]: sshd@21-10.0.0.132:22-10.0.0.1:36394.service: Deactivated successfully. Aug 13 01:06:41.127432 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 01:06:41.127894 systemd-logind[1195]: Session 22 logged out. Waiting for processes to exit. Aug 13 01:06:41.128581 systemd-logind[1195]: Removed session 22. Aug 13 01:06:42.044787 kubelet[1948]: E0813 01:06:42.044746 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:06:46.129040 systemd[1]: Started sshd@22-10.0.0.132:22-10.0.0.1:36406.service. Aug 13 01:06:46.158678 sshd[3536]: Accepted publickey for core from 10.0.0.1 port 36406 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:46.159925 sshd[3536]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:46.164045 systemd-logind[1195]: New session 23 of user core. Aug 13 01:06:46.165200 systemd[1]: Started session-23.scope. Aug 13 01:06:46.293961 sshd[3536]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:46.296179 systemd[1]: sshd@22-10.0.0.132:22-10.0.0.1:36406.service: Deactivated successfully. Aug 13 01:06:46.296951 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 01:06:46.297566 systemd-logind[1195]: Session 23 logged out. Waiting for processes to exit. Aug 13 01:06:46.298346 systemd-logind[1195]: Removed session 23. Aug 13 01:06:51.298238 systemd[1]: Started sshd@23-10.0.0.132:22-10.0.0.1:40010.service. Aug 13 01:06:51.326069 sshd[3552]: Accepted publickey for core from 10.0.0.1 port 40010 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:51.326962 sshd[3552]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:51.330000 systemd-logind[1195]: New session 24 of user core. Aug 13 01:06:51.330722 systemd[1]: Started session-24.scope. Aug 13 01:06:51.499616 sshd[3552]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:51.501724 systemd[1]: sshd@23-10.0.0.132:22-10.0.0.1:40010.service: Deactivated successfully. Aug 13 01:06:51.502373 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 01:06:51.502863 systemd-logind[1195]: Session 24 logged out. Waiting for processes to exit. Aug 13 01:06:51.503548 systemd-logind[1195]: Removed session 24. Aug 13 01:06:56.506274 systemd[1]: Started sshd@24-10.0.0.132:22-10.0.0.1:40012.service. Aug 13 01:06:56.558364 sshd[3566]: Accepted publickey for core from 10.0.0.1 port 40012 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:06:56.561396 sshd[3566]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:06:56.573113 systemd-logind[1195]: New session 25 of user core. Aug 13 01:06:56.574326 systemd[1]: Started session-25.scope. Aug 13 01:06:56.805350 sshd[3566]: pam_unix(sshd:session): session closed for user core Aug 13 01:06:56.810324 systemd[1]: sshd@24-10.0.0.132:22-10.0.0.1:40012.service: Deactivated successfully. Aug 13 01:06:56.811640 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 01:06:56.815046 systemd-logind[1195]: Session 25 logged out. Waiting for processes to exit. Aug 13 01:06:56.818049 systemd-logind[1195]: Removed session 25. Aug 13 01:06:58.045116 kubelet[1948]: E0813 01:06:58.045054 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:01.805971 systemd[1]: Started sshd@25-10.0.0.132:22-10.0.0.1:55220.service. Aug 13 01:07:01.834651 sshd[3581]: Accepted publickey for core from 10.0.0.1 port 55220 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:07:01.835782 sshd[3581]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:07:01.839399 systemd-logind[1195]: New session 26 of user core. Aug 13 01:07:01.840478 systemd[1]: Started session-26.scope. Aug 13 01:07:01.947996 sshd[3581]: pam_unix(sshd:session): session closed for user core Aug 13 01:07:01.951416 systemd[1]: sshd@25-10.0.0.132:22-10.0.0.1:55220.service: Deactivated successfully. Aug 13 01:07:01.952107 systemd[1]: session-26.scope: Deactivated successfully. Aug 13 01:07:01.952820 systemd-logind[1195]: Session 26 logged out. Waiting for processes to exit. Aug 13 01:07:01.954018 systemd[1]: Started sshd@26-10.0.0.132:22-10.0.0.1:55232.service. Aug 13 01:07:01.954999 systemd-logind[1195]: Removed session 26. Aug 13 01:07:01.986314 sshd[3594]: Accepted publickey for core from 10.0.0.1 port 55232 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:07:01.987619 sshd[3594]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:07:01.991519 systemd-logind[1195]: New session 27 of user core. Aug 13 01:07:01.992627 systemd[1]: Started session-27.scope. Aug 13 01:07:03.508574 env[1207]: time="2025-08-13T01:07:03.508514871Z" level=info msg="StopContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" with timeout 30 (s)" Aug 13 01:07:03.509052 env[1207]: time="2025-08-13T01:07:03.508986082Z" level=info msg="Stop container \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" with signal terminated" Aug 13 01:07:03.523928 systemd[1]: run-containerd-runc-k8s.io-ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864-runc.YvEYMP.mount: Deactivated successfully. Aug 13 01:07:03.524545 systemd[1]: cri-containerd-2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f.scope: Deactivated successfully. Aug 13 01:07:03.540635 env[1207]: time="2025-08-13T01:07:03.540563408Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 01:07:03.543017 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f-rootfs.mount: Deactivated successfully. Aug 13 01:07:03.546605 env[1207]: time="2025-08-13T01:07:03.546559987Z" level=info msg="StopContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" with timeout 2 (s)" Aug 13 01:07:03.546872 env[1207]: time="2025-08-13T01:07:03.546841871Z" level=info msg="Stop container \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" with signal terminated" Aug 13 01:07:03.552390 systemd-networkd[1020]: lxc_health: Link DOWN Aug 13 01:07:03.552397 systemd-networkd[1020]: lxc_health: Lost carrier Aug 13 01:07:03.555575 env[1207]: time="2025-08-13T01:07:03.555530996Z" level=info msg="shim disconnected" id=2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f Aug 13 01:07:03.555666 env[1207]: time="2025-08-13T01:07:03.555586271Z" level=warning msg="cleaning up after shim disconnected" id=2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f namespace=k8s.io Aug 13 01:07:03.555666 env[1207]: time="2025-08-13T01:07:03.555595999Z" level=info msg="cleaning up dead shim" Aug 13 01:07:03.562093 env[1207]: time="2025-08-13T01:07:03.562046538Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3646 runtime=io.containerd.runc.v2\n" Aug 13 01:07:03.567591 env[1207]: time="2025-08-13T01:07:03.567551847Z" level=info msg="StopContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" returns successfully" Aug 13 01:07:03.568212 env[1207]: time="2025-08-13T01:07:03.568187521Z" level=info msg="StopPodSandbox for \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\"" Aug 13 01:07:03.568374 env[1207]: time="2025-08-13T01:07:03.568349758Z" level=info msg="Container to stop \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.570183 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8-shm.mount: Deactivated successfully. Aug 13 01:07:03.581749 systemd[1]: cri-containerd-7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8.scope: Deactivated successfully. Aug 13 01:07:03.597502 systemd[1]: cri-containerd-ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864.scope: Deactivated successfully. Aug 13 01:07:03.597738 systemd[1]: cri-containerd-ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864.scope: Consumed 6.191s CPU time. Aug 13 01:07:03.600404 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8-rootfs.mount: Deactivated successfully. Aug 13 01:07:03.611406 env[1207]: time="2025-08-13T01:07:03.611352266Z" level=info msg="shim disconnected" id=7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8 Aug 13 01:07:03.611406 env[1207]: time="2025-08-13T01:07:03.611406930Z" level=warning msg="cleaning up after shim disconnected" id=7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8 namespace=k8s.io Aug 13 01:07:03.611530 env[1207]: time="2025-08-13T01:07:03.611415767Z" level=info msg="cleaning up dead shim" Aug 13 01:07:03.617872 env[1207]: time="2025-08-13T01:07:03.617805911Z" level=info msg="shim disconnected" id=ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864 Aug 13 01:07:03.617872 env[1207]: time="2025-08-13T01:07:03.617865854Z" level=warning msg="cleaning up after shim disconnected" id=ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864 namespace=k8s.io Aug 13 01:07:03.617872 env[1207]: time="2025-08-13T01:07:03.617875503Z" level=info msg="cleaning up dead shim" Aug 13 01:07:03.618919 env[1207]: time="2025-08-13T01:07:03.618880675Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3691 runtime=io.containerd.runc.v2\n" Aug 13 01:07:03.619213 env[1207]: time="2025-08-13T01:07:03.619178710Z" level=info msg="TearDown network for sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" successfully" Aug 13 01:07:03.619213 env[1207]: time="2025-08-13T01:07:03.619202565Z" level=info msg="StopPodSandbox for \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" returns successfully" Aug 13 01:07:03.629486 env[1207]: time="2025-08-13T01:07:03.629446783Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3703 runtime=io.containerd.runc.v2\n" Aug 13 01:07:03.684002 env[1207]: time="2025-08-13T01:07:03.683932556Z" level=info msg="StopContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" returns successfully" Aug 13 01:07:03.684444 env[1207]: time="2025-08-13T01:07:03.684409238Z" level=info msg="StopPodSandbox for \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\"" Aug 13 01:07:03.684585 env[1207]: time="2025-08-13T01:07:03.684468149Z" level=info msg="Container to stop \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.684585 env[1207]: time="2025-08-13T01:07:03.684482497Z" level=info msg="Container to stop \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.684585 env[1207]: time="2025-08-13T01:07:03.684493899Z" level=info msg="Container to stop \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.684585 env[1207]: time="2025-08-13T01:07:03.684505510Z" level=info msg="Container to stop \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.684585 env[1207]: time="2025-08-13T01:07:03.684515399Z" level=info msg="Container to stop \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:03.690006 systemd[1]: cri-containerd-be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f.scope: Deactivated successfully. Aug 13 01:07:03.705060 kubelet[1948]: I0813 01:07:03.705023 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0e2881b-250b-43b5-aa9b-437c0aa094c6-cilium-config-path\") pod \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\" (UID: \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\") " Aug 13 01:07:03.705060 kubelet[1948]: I0813 01:07:03.705061 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjpm\" (UniqueName: \"kubernetes.io/projected/c0e2881b-250b-43b5-aa9b-437c0aa094c6-kube-api-access-2tjpm\") pod \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\" (UID: \"c0e2881b-250b-43b5-aa9b-437c0aa094c6\") " Aug 13 01:07:03.707800 kubelet[1948]: I0813 01:07:03.707751 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e2881b-250b-43b5-aa9b-437c0aa094c6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "c0e2881b-250b-43b5-aa9b-437c0aa094c6" (UID: "c0e2881b-250b-43b5-aa9b-437c0aa094c6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 01:07:03.708882 kubelet[1948]: I0813 01:07:03.708858 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0e2881b-250b-43b5-aa9b-437c0aa094c6-kube-api-access-2tjpm" (OuterVolumeSpecName: "kube-api-access-2tjpm") pod "c0e2881b-250b-43b5-aa9b-437c0aa094c6" (UID: "c0e2881b-250b-43b5-aa9b-437c0aa094c6"). InnerVolumeSpecName "kube-api-access-2tjpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 01:07:03.712024 env[1207]: time="2025-08-13T01:07:03.711976524Z" level=info msg="shim disconnected" id=be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f Aug 13 01:07:03.712118 env[1207]: time="2025-08-13T01:07:03.712027981Z" level=warning msg="cleaning up after shim disconnected" id=be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f namespace=k8s.io Aug 13 01:07:03.712118 env[1207]: time="2025-08-13T01:07:03.712035696Z" level=info msg="cleaning up dead shim" Aug 13 01:07:03.718660 env[1207]: time="2025-08-13T01:07:03.718621801Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3734 runtime=io.containerd.runc.v2\n" Aug 13 01:07:03.718924 env[1207]: time="2025-08-13T01:07:03.718901591Z" level=info msg="TearDown network for sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" successfully" Aug 13 01:07:03.718991 env[1207]: time="2025-08-13T01:07:03.718923652Z" level=info msg="StopPodSandbox for \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" returns successfully" Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806249 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-kernel\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806304 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-etc-cni-netd\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806326 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-cgroup\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806344 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-xtables-lock\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806366 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-config-path\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.806392 kubelet[1948]: I0813 01:07:03.806365 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.806671 kubelet[1948]: I0813 01:07:03.806420 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.806671 kubelet[1948]: I0813 01:07:03.806436 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.806671 kubelet[1948]: I0813 01:07:03.806440 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807076 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-bpf-maps\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807106 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cni-path\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807121 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-run\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807145 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hubble-tls\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807161 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zkdj\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-kube-api-access-5zkdj\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807274 kubelet[1948]: I0813 01:07:03.807172 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hostproc\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807185 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-lib-modules\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807200 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-clustermesh-secrets\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807213 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-net\") pod \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\" (UID: \"a903bf21-d1fd-4b11-b47b-bf64f5bb48a6\") " Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807277 1948 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807293 1948 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807300 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807510 kubelet[1948]: I0813 01:07:03.807309 1948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjpm\" (UniqueName: \"kubernetes.io/projected/c0e2881b-250b-43b5-aa9b-437c0aa094c6-kube-api-access-2tjpm\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807319 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0e2881b-250b-43b5-aa9b-437c0aa094c6-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807326 1948 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-xtables-lock\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807349 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807368 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807381 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cni-path" (OuterVolumeSpecName: "cni-path") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.807693 kubelet[1948]: I0813 01:07:03.807393 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.809691 kubelet[1948]: I0813 01:07:03.807894 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.809691 kubelet[1948]: I0813 01:07:03.808154 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hostproc" (OuterVolumeSpecName: "hostproc") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:03.809691 kubelet[1948]: I0813 01:07:03.808519 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 01:07:03.810077 kubelet[1948]: I0813 01:07:03.810050 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 01:07:03.810900 kubelet[1948]: I0813 01:07:03.810876 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-kube-api-access-5zkdj" (OuterVolumeSpecName: "kube-api-access-5zkdj") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "kube-api-access-5zkdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 01:07:03.811302 kubelet[1948]: I0813 01:07:03.811228 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" (UID: "a903bf21-d1fd-4b11-b47b-bf64f5bb48a6"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907561 1948 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-bpf-maps\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907598 1948 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cni-path\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907606 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-run\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907613 1948 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hubble-tls\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907625 1948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zkdj\" (UniqueName: \"kubernetes.io/projected/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-kube-api-access-5zkdj\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907632 1948 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-hostproc\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907639 1948 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-lib-modules\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.907625 kubelet[1948]: I0813 01:07:03.907645 1948 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.908031 kubelet[1948]: I0813 01:07:03.907652 1948 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:03.908031 kubelet[1948]: I0813 01:07:03.907659 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:04.521707 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864-rootfs.mount: Deactivated successfully. Aug 13 01:07:04.521824 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f-rootfs.mount: Deactivated successfully. Aug 13 01:07:04.521914 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f-shm.mount: Deactivated successfully. Aug 13 01:07:04.522000 systemd[1]: var-lib-kubelet-pods-c0e2881b\x2d250b\x2d43b5\x2daa9b\x2d437c0aa094c6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2tjpm.mount: Deactivated successfully. Aug 13 01:07:04.522073 systemd[1]: var-lib-kubelet-pods-a903bf21\x2dd1fd\x2d4b11\x2db47b\x2dbf64f5bb48a6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d5zkdj.mount: Deactivated successfully. Aug 13 01:07:04.522150 systemd[1]: var-lib-kubelet-pods-a903bf21\x2dd1fd\x2d4b11\x2db47b\x2dbf64f5bb48a6-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Aug 13 01:07:04.522235 systemd[1]: var-lib-kubelet-pods-a903bf21\x2dd1fd\x2d4b11\x2db47b\x2dbf64f5bb48a6-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Aug 13 01:07:04.535016 kubelet[1948]: I0813 01:07:04.534984 1948 scope.go:117] "RemoveContainer" containerID="2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f" Aug 13 01:07:04.536339 env[1207]: time="2025-08-13T01:07:04.536296662Z" level=info msg="RemoveContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\"" Aug 13 01:07:04.541508 systemd[1]: Removed slice kubepods-besteffort-podc0e2881b_250b_43b5_aa9b_437c0aa094c6.slice. Aug 13 01:07:04.546497 systemd[1]: Removed slice kubepods-burstable-poda903bf21_d1fd_4b11_b47b_bf64f5bb48a6.slice. Aug 13 01:07:04.546609 systemd[1]: kubepods-burstable-poda903bf21_d1fd_4b11_b47b_bf64f5bb48a6.slice: Consumed 6.284s CPU time. Aug 13 01:07:04.673629 env[1207]: time="2025-08-13T01:07:04.673565067Z" level=info msg="RemoveContainer for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" returns successfully" Aug 13 01:07:04.673886 kubelet[1948]: I0813 01:07:04.673856 1948 scope.go:117] "RemoveContainer" containerID="2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f" Aug 13 01:07:04.674243 env[1207]: time="2025-08-13T01:07:04.674154954Z" level=error msg="ContainerStatus for \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\": not found" Aug 13 01:07:04.674368 kubelet[1948]: E0813 01:07:04.674346 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\": not found" containerID="2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f" Aug 13 01:07:04.674466 kubelet[1948]: I0813 01:07:04.674389 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f"} err="failed to get container status \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\": rpc error: code = NotFound desc = an error occurred when try to find container \"2b99c505db49baee06eef62efefff329493273348978c39c12236f5985db637f\": not found" Aug 13 01:07:04.674505 kubelet[1948]: I0813 01:07:04.674468 1948 scope.go:117] "RemoveContainer" containerID="ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864" Aug 13 01:07:04.675409 env[1207]: time="2025-08-13T01:07:04.675381966Z" level=info msg="RemoveContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\"" Aug 13 01:07:04.718679 env[1207]: time="2025-08-13T01:07:04.718617391Z" level=info msg="RemoveContainer for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" returns successfully" Aug 13 01:07:04.718919 kubelet[1948]: I0813 01:07:04.718867 1948 scope.go:117] "RemoveContainer" containerID="2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0" Aug 13 01:07:04.720068 env[1207]: time="2025-08-13T01:07:04.720033140Z" level=info msg="RemoveContainer for \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\"" Aug 13 01:07:04.765978 env[1207]: time="2025-08-13T01:07:04.765912228Z" level=info msg="RemoveContainer for \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\" returns successfully" Aug 13 01:07:04.766229 kubelet[1948]: I0813 01:07:04.766196 1948 scope.go:117] "RemoveContainer" containerID="6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6" Aug 13 01:07:04.767507 env[1207]: time="2025-08-13T01:07:04.767238488Z" level=info msg="RemoveContainer for \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\"" Aug 13 01:07:04.924554 env[1207]: time="2025-08-13T01:07:04.924419295Z" level=info msg="RemoveContainer for \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\" returns successfully" Aug 13 01:07:04.924802 kubelet[1948]: I0813 01:07:04.924771 1948 scope.go:117] "RemoveContainer" containerID="570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053" Aug 13 01:07:04.925883 env[1207]: time="2025-08-13T01:07:04.925853850Z" level=info msg="RemoveContainer for \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\"" Aug 13 01:07:05.046180 kubelet[1948]: I0813 01:07:05.046102 1948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" path="/var/lib/kubelet/pods/a903bf21-d1fd-4b11-b47b-bf64f5bb48a6/volumes" Aug 13 01:07:05.046378 env[1207]: time="2025-08-13T01:07:05.046270233Z" level=info msg="RemoveContainer for \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\" returns successfully" Aug 13 01:07:05.046441 kubelet[1948]: I0813 01:07:05.046408 1948 scope.go:117] "RemoveContainer" containerID="616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2" Aug 13 01:07:05.046929 kubelet[1948]: I0813 01:07:05.046879 1948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0e2881b-250b-43b5-aa9b-437c0aa094c6" path="/var/lib/kubelet/pods/c0e2881b-250b-43b5-aa9b-437c0aa094c6/volumes" Aug 13 01:07:05.047579 env[1207]: time="2025-08-13T01:07:05.047335398Z" level=info msg="RemoveContainer for \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\"" Aug 13 01:07:05.115444 env[1207]: time="2025-08-13T01:07:05.115373697Z" level=info msg="RemoveContainer for \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\" returns successfully" Aug 13 01:07:05.115721 kubelet[1948]: I0813 01:07:05.115671 1948 scope.go:117] "RemoveContainer" containerID="ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864" Aug 13 01:07:05.116038 env[1207]: time="2025-08-13T01:07:05.115971358Z" level=error msg="ContainerStatus for \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\": not found" Aug 13 01:07:05.116181 kubelet[1948]: E0813 01:07:05.116151 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\": not found" containerID="ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864" Aug 13 01:07:05.116251 kubelet[1948]: I0813 01:07:05.116187 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864"} err="failed to get container status \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\": rpc error: code = NotFound desc = an error occurred when try to find container \"ab6e338e7278c7524db6f3aa99fa94c8462fe8a39aebfb947129e696421bc864\": not found" Aug 13 01:07:05.116251 kubelet[1948]: I0813 01:07:05.116207 1948 scope.go:117] "RemoveContainer" containerID="2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0" Aug 13 01:07:05.116429 env[1207]: time="2025-08-13T01:07:05.116379139Z" level=error msg="ContainerStatus for \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\": not found" Aug 13 01:07:05.116547 kubelet[1948]: E0813 01:07:05.116527 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\": not found" containerID="2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0" Aug 13 01:07:05.116595 kubelet[1948]: I0813 01:07:05.116549 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0"} err="failed to get container status \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\": rpc error: code = NotFound desc = an error occurred when try to find container \"2eefcc098db4093caafc28b123f2a6658313703e643e13dfb075c76c69ebc1b0\": not found" Aug 13 01:07:05.116595 kubelet[1948]: I0813 01:07:05.116563 1948 scope.go:117] "RemoveContainer" containerID="6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6" Aug 13 01:07:05.116764 env[1207]: time="2025-08-13T01:07:05.116719063Z" level=error msg="ContainerStatus for \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\": not found" Aug 13 01:07:05.116875 kubelet[1948]: E0813 01:07:05.116848 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\": not found" containerID="6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6" Aug 13 01:07:05.116950 kubelet[1948]: I0813 01:07:05.116876 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6"} err="failed to get container status \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\": rpc error: code = NotFound desc = an error occurred when try to find container \"6b40cb7df81505a4f5338c396ed3c8b681990ed7a7c23711915808eaf5f21ab6\": not found" Aug 13 01:07:05.116950 kubelet[1948]: I0813 01:07:05.116900 1948 scope.go:117] "RemoveContainer" containerID="570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053" Aug 13 01:07:05.117104 env[1207]: time="2025-08-13T01:07:05.117057824Z" level=error msg="ContainerStatus for \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\": not found" Aug 13 01:07:05.117207 kubelet[1948]: E0813 01:07:05.117183 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\": not found" containerID="570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053" Aug 13 01:07:05.117282 kubelet[1948]: I0813 01:07:05.117224 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053"} err="failed to get container status \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\": rpc error: code = NotFound desc = an error occurred when try to find container \"570b6f73e485952464bddf40e1fa814503286c1310f9f3c9d05e93217c5ef053\": not found" Aug 13 01:07:05.117282 kubelet[1948]: I0813 01:07:05.117245 1948 scope.go:117] "RemoveContainer" containerID="616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2" Aug 13 01:07:05.117481 env[1207]: time="2025-08-13T01:07:05.117431190Z" level=error msg="ContainerStatus for \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\": not found" Aug 13 01:07:05.117576 kubelet[1948]: E0813 01:07:05.117555 1948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\": not found" containerID="616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2" Aug 13 01:07:05.117625 kubelet[1948]: I0813 01:07:05.117578 1948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2"} err="failed to get container status \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\": rpc error: code = NotFound desc = an error occurred when try to find container \"616d2d56b82a26f2731a77bc541a9cf67df09952d306f26f065da7af52153bd2\": not found" Aug 13 01:07:05.591639 sshd[3594]: pam_unix(sshd:session): session closed for user core Aug 13 01:07:05.594607 systemd[1]: sshd@26-10.0.0.132:22-10.0.0.1:55232.service: Deactivated successfully. Aug 13 01:07:05.595204 systemd[1]: session-27.scope: Deactivated successfully. Aug 13 01:07:05.596076 systemd-logind[1195]: Session 27 logged out. Waiting for processes to exit. Aug 13 01:07:05.597578 systemd[1]: Started sshd@27-10.0.0.132:22-10.0.0.1:55234.service. Aug 13 01:07:05.598469 systemd-logind[1195]: Removed session 27. Aug 13 01:07:05.628065 sshd[3751]: Accepted publickey for core from 10.0.0.1 port 55234 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:07:05.629164 sshd[3751]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:07:05.632693 systemd-logind[1195]: New session 28 of user core. Aug 13 01:07:05.633572 systemd[1]: Started session-28.scope. Aug 13 01:07:06.095214 kubelet[1948]: E0813 01:07:06.095162 1948 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 01:07:07.054354 sshd[3751]: pam_unix(sshd:session): session closed for user core Aug 13 01:07:07.057352 systemd[1]: sshd@27-10.0.0.132:22-10.0.0.1:55234.service: Deactivated successfully. Aug 13 01:07:07.057962 systemd[1]: session-28.scope: Deactivated successfully. Aug 13 01:07:07.058597 systemd-logind[1195]: Session 28 logged out. Waiting for processes to exit. Aug 13 01:07:07.060606 systemd[1]: Started sshd@28-10.0.0.132:22-10.0.0.1:55242.service. Aug 13 01:07:07.063746 systemd-logind[1195]: Removed session 28. Aug 13 01:07:07.089328 sshd[3763]: Accepted publickey for core from 10.0.0.1 port 55242 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:07:07.090444 sshd[3763]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:07:07.094309 systemd-logind[1195]: New session 29 of user core. Aug 13 01:07:07.094982 systemd[1]: Started session-29.scope. Aug 13 01:07:07.434890 sshd[3763]: pam_unix(sshd:session): session closed for user core Aug 13 01:07:07.438100 systemd[1]: sshd@28-10.0.0.132:22-10.0.0.1:55242.service: Deactivated successfully. Aug 13 01:07:07.438649 systemd[1]: session-29.scope: Deactivated successfully. Aug 13 01:07:07.439416 systemd-logind[1195]: Session 29 logged out. Waiting for processes to exit. Aug 13 01:07:07.440652 systemd[1]: Started sshd@29-10.0.0.132:22-10.0.0.1:55256.service. Aug 13 01:07:07.441610 systemd-logind[1195]: Removed session 29. Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447110 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="clean-cilium-state" Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447142 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="cilium-agent" Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447149 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="mount-bpf-fs" Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447155 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="mount-cgroup" Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447161 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="apply-sysctl-overwrites" Aug 13 01:07:07.447156 kubelet[1948]: E0813 01:07:07.447166 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c0e2881b-250b-43b5-aa9b-437c0aa094c6" containerName="cilium-operator" Aug 13 01:07:07.447806 kubelet[1948]: I0813 01:07:07.447193 1948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a903bf21-d1fd-4b11-b47b-bf64f5bb48a6" containerName="cilium-agent" Aug 13 01:07:07.447806 kubelet[1948]: I0813 01:07:07.447202 1948 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0e2881b-250b-43b5-aa9b-437c0aa094c6" containerName="cilium-operator" Aug 13 01:07:07.454494 systemd[1]: Created slice kubepods-burstable-pod88b6eb70_f275_4f3b_8e62_cb57424e90b4.slice. Aug 13 01:07:07.469337 sshd[3779]: Accepted publickey for core from 10.0.0.1 port 55256 ssh2: RSA SHA256:qgan5rMjZ6sYv4kBQbHPXcuGXLcxEJ8myXWtyGqiw0s Aug 13 01:07:07.470733 sshd[3779]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 01:07:07.474478 systemd-logind[1195]: New session 30 of user core. Aug 13 01:07:07.475347 systemd[1]: Started session-30.scope. Aug 13 01:07:07.531594 kubelet[1948]: I0813 01:07:07.531546 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-lib-modules\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531594 kubelet[1948]: I0813 01:07:07.531590 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-clustermesh-secrets\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531594 kubelet[1948]: I0813 01:07:07.531609 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hostproc\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531864 kubelet[1948]: I0813 01:07:07.531637 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-config-path\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531864 kubelet[1948]: I0813 01:07:07.531652 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-ipsec-secrets\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531864 kubelet[1948]: I0813 01:07:07.531665 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-kernel\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531864 kubelet[1948]: I0813 01:07:07.531678 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g9sz\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-kube-api-access-5g9sz\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.531864 kubelet[1948]: I0813 01:07:07.531693 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-run\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531709 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-etc-cni-netd\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531725 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cni-path\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531738 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-xtables-lock\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531751 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-cgroup\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531767 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-net\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532084 kubelet[1948]: I0813 01:07:07.531778 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hubble-tls\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.532355 kubelet[1948]: I0813 01:07:07.531797 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-bpf-maps\") pod \"cilium-qd5zg\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " pod="kube-system/cilium-qd5zg" Aug 13 01:07:07.756765 kubelet[1948]: E0813 01:07:07.756701 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:07.757378 env[1207]: time="2025-08-13T01:07:07.757290409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qd5zg,Uid:88b6eb70-f275-4f3b-8e62-cb57424e90b4,Namespace:kube-system,Attempt:0,}" Aug 13 01:07:07.996049 env[1207]: time="2025-08-13T01:07:07.995930091Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:07:07.996049 env[1207]: time="2025-08-13T01:07:07.995987640Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:07:07.996049 env[1207]: time="2025-08-13T01:07:07.996003009Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:07:07.996303 env[1207]: time="2025-08-13T01:07:07.996215101Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1 pid=3802 runtime=io.containerd.runc.v2 Aug 13 01:07:08.008478 systemd[1]: Started cri-containerd-0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1.scope. Aug 13 01:07:08.028812 env[1207]: time="2025-08-13T01:07:08.028759136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qd5zg,Uid:88b6eb70-f275-4f3b-8e62-cb57424e90b4,Namespace:kube-system,Attempt:0,} returns sandbox id \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\"" Aug 13 01:07:08.029865 kubelet[1948]: E0813 01:07:08.029844 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:08.031899 env[1207]: time="2025-08-13T01:07:08.031860300Z" level=info msg="CreateContainer within sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 01:07:08.298010 env[1207]: time="2025-08-13T01:07:08.297846178Z" level=info msg="CreateContainer within sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\"" Aug 13 01:07:08.298623 env[1207]: time="2025-08-13T01:07:08.298559397Z" level=info msg="StartContainer for \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\"" Aug 13 01:07:08.312920 systemd[1]: Started cri-containerd-3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a.scope. Aug 13 01:07:08.323909 systemd[1]: cri-containerd-3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a.scope: Deactivated successfully. Aug 13 01:07:08.324164 systemd[1]: Stopped cri-containerd-3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a.scope. Aug 13 01:07:08.520516 env[1207]: time="2025-08-13T01:07:08.520455206Z" level=info msg="shim disconnected" id=3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a Aug 13 01:07:08.520516 env[1207]: time="2025-08-13T01:07:08.520508335Z" level=warning msg="cleaning up after shim disconnected" id=3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a namespace=k8s.io Aug 13 01:07:08.520516 env[1207]: time="2025-08-13T01:07:08.520516492Z" level=info msg="cleaning up dead shim" Aug 13 01:07:08.527748 env[1207]: time="2025-08-13T01:07:08.527684987Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3858 runtime=io.containerd.runc.v2\ntime=\"2025-08-13T01:07:08Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Aug 13 01:07:08.528141 env[1207]: time="2025-08-13T01:07:08.528003900Z" level=error msg="copy shim log" error="read /proc/self/fd/29: file already closed" Aug 13 01:07:08.528401 env[1207]: time="2025-08-13T01:07:08.528359724Z" level=error msg="Failed to pipe stdout of container \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\"" error="reading from a closed fifo" Aug 13 01:07:08.528461 env[1207]: time="2025-08-13T01:07:08.528429195Z" level=error msg="Failed to pipe stderr of container \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\"" error="reading from a closed fifo" Aug 13 01:07:08.602077 env[1207]: time="2025-08-13T01:07:08.601482246Z" level=error msg="StartContainer for \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Aug 13 01:07:08.602477 kubelet[1948]: E0813 01:07:08.602392 1948 log.go:32] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a" Aug 13 01:07:08.603463 kubelet[1948]: E0813 01:07:08.603434 1948 kuberuntime_manager.go:1274] "Unhandled Error" err=< Aug 13 01:07:08.603463 kubelet[1948]: init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Aug 13 01:07:08.603463 kubelet[1948]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Aug 13 01:07:08.603463 kubelet[1948]: rm /hostbin/cilium-mount Aug 13 01:07:08.603589 kubelet[1948]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g9sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:&AppArmorProfile{Type:Unconfined,LocalhostProfile:nil,},},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-qd5zg_kube-system(88b6eb70-f275-4f3b-8e62-cb57424e90b4): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Aug 13 01:07:08.603589 kubelet[1948]: > logger="UnhandledError" Aug 13 01:07:08.604611 kubelet[1948]: E0813 01:07:08.604558 1948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-qd5zg" podUID="88b6eb70-f275-4f3b-8e62-cb57424e90b4" Aug 13 01:07:09.606118 env[1207]: time="2025-08-13T01:07:09.606046693Z" level=info msg="StopPodSandbox for \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\"" Aug 13 01:07:09.606581 env[1207]: time="2025-08-13T01:07:09.606123499Z" level=info msg="Container to stop \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 01:07:09.608715 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1-shm.mount: Deactivated successfully. Aug 13 01:07:09.612675 systemd[1]: cri-containerd-0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1.scope: Deactivated successfully. Aug 13 01:07:09.629402 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1-rootfs.mount: Deactivated successfully. Aug 13 01:07:09.665231 env[1207]: time="2025-08-13T01:07:09.665159456Z" level=info msg="shim disconnected" id=0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1 Aug 13 01:07:09.666181 env[1207]: time="2025-08-13T01:07:09.666145409Z" level=warning msg="cleaning up after shim disconnected" id=0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1 namespace=k8s.io Aug 13 01:07:09.666181 env[1207]: time="2025-08-13T01:07:09.666163394Z" level=info msg="cleaning up dead shim" Aug 13 01:07:09.672806 env[1207]: time="2025-08-13T01:07:09.672744157Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:09Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3888 runtime=io.containerd.runc.v2\n" Aug 13 01:07:09.673213 env[1207]: time="2025-08-13T01:07:09.673180292Z" level=info msg="TearDown network for sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" successfully" Aug 13 01:07:09.673213 env[1207]: time="2025-08-13T01:07:09.673209867Z" level=info msg="StopPodSandbox for \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" returns successfully" Aug 13 01:07:09.746364 kubelet[1948]: I0813 01:07:09.746297 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-clustermesh-secrets\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746364 kubelet[1948]: I0813 01:07:09.746359 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-kernel\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746364 kubelet[1948]: I0813 01:07:09.746380 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g9sz\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-kube-api-access-5g9sz\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746396 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-cgroup\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746416 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hubble-tls\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746433 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cni-path\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746448 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hostproc\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746440 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746465 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-config-path\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746540 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-ipsec-secrets\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746568 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-run\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746588 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-etc-cni-netd\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746605 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-xtables-lock\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746624 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-bpf-maps\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746641 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-lib-modules\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746657 1948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-net\") pod \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\" (UID: \"88b6eb70-f275-4f3b-8e62-cb57424e90b4\") " Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746690 1948 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.746911 kubelet[1948]: I0813 01:07:09.746715 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747372 kubelet[1948]: I0813 01:07:09.747351 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747412 kubelet[1948]: I0813 01:07:09.747386 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747412 kubelet[1948]: I0813 01:07:09.747403 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747472 kubelet[1948]: I0813 01:07:09.747417 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747472 kubelet[1948]: I0813 01:07:09.747431 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747643 kubelet[1948]: I0813 01:07:09.747611 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cni-path" (OuterVolumeSpecName: "cni-path") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747701 kubelet[1948]: I0813 01:07:09.747662 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.747701 kubelet[1948]: I0813 01:07:09.747686 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hostproc" (OuterVolumeSpecName: "hostproc") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 01:07:09.748765 kubelet[1948]: I0813 01:07:09.748733 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 01:07:09.749653 kubelet[1948]: I0813 01:07:09.749618 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 01:07:09.750006 kubelet[1948]: I0813 01:07:09.749974 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-kube-api-access-5g9sz" (OuterVolumeSpecName: "kube-api-access-5g9sz") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "kube-api-access-5g9sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 01:07:09.750304 kubelet[1948]: I0813 01:07:09.750280 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 01:07:09.751420 systemd[1]: var-lib-kubelet-pods-88b6eb70\x2df275\x2d4f3b\x2d8e62\x2dcb57424e90b4-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d5g9sz.mount: Deactivated successfully. Aug 13 01:07:09.751528 systemd[1]: var-lib-kubelet-pods-88b6eb70\x2df275\x2d4f3b\x2d8e62\x2dcb57424e90b4-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Aug 13 01:07:09.751583 systemd[1]: var-lib-kubelet-pods-88b6eb70\x2df275\x2d4f3b\x2d8e62\x2dcb57424e90b4-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Aug 13 01:07:09.753797 systemd[1]: var-lib-kubelet-pods-88b6eb70\x2df275\x2d4f3b\x2d8e62\x2dcb57424e90b4-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Aug 13 01:07:09.755084 kubelet[1948]: I0813 01:07:09.755043 1948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "88b6eb70-f275-4f3b-8e62-cb57424e90b4" (UID: "88b6eb70-f275-4f3b-8e62-cb57424e90b4"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 01:07:09.847597 kubelet[1948]: I0813 01:07:09.847526 1948 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847597 kubelet[1948]: I0813 01:07:09.847578 1948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g9sz\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-kube-api-access-5g9sz\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847597 kubelet[1948]: I0813 01:07:09.847590 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847597 kubelet[1948]: I0813 01:07:09.847602 1948 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hubble-tls\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847597 kubelet[1948]: I0813 01:07:09.847614 1948 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cni-path\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847626 1948 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-hostproc\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847637 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847647 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-ipsec-secrets\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847657 1948 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-cilium-run\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847666 1948 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847676 1948 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-xtables-lock\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847688 1948 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-bpf-maps\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847719 1948 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-lib-modules\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:09.847875 kubelet[1948]: I0813 01:07:09.847730 1948 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/88b6eb70-f275-4f3b-8e62-cb57424e90b4-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Aug 13 01:07:10.608364 kubelet[1948]: I0813 01:07:10.608322 1948 scope.go:117] "RemoveContainer" containerID="3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a" Aug 13 01:07:10.611979 systemd[1]: Removed slice kubepods-burstable-pod88b6eb70_f275_4f3b_8e62_cb57424e90b4.slice. Aug 13 01:07:10.612581 env[1207]: time="2025-08-13T01:07:10.612535971Z" level=info msg="RemoveContainer for \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\"" Aug 13 01:07:10.679383 env[1207]: time="2025-08-13T01:07:10.679330749Z" level=info msg="RemoveContainer for \"3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a\" returns successfully" Aug 13 01:07:10.720760 kubelet[1948]: E0813 01:07:10.720722 1948 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="88b6eb70-f275-4f3b-8e62-cb57424e90b4" containerName="mount-cgroup" Aug 13 01:07:10.721047 kubelet[1948]: I0813 01:07:10.721014 1948 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b6eb70-f275-4f3b-8e62-cb57424e90b4" containerName="mount-cgroup" Aug 13 01:07:10.728020 systemd[1]: Created slice kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice. Aug 13 01:07:10.753356 kubelet[1948]: I0813 01:07:10.753320 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-host-proc-sys-kernel\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753356 kubelet[1948]: I0813 01:07:10.753355 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-cni-path\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753370 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2b22d242-0786-4cc6-90fc-ed25654ee072-hubble-tls\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753382 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-cilium-run\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753395 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-etc-cni-netd\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753407 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/2b22d242-0786-4cc6-90fc-ed25654ee072-cilium-ipsec-secrets\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753443 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-cilium-cgroup\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753468 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-lib-modules\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753481 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-xtables-lock\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753505 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-bpf-maps\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753585 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-hostproc\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753621 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2b22d242-0786-4cc6-90fc-ed25654ee072-host-proc-sys-net\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753645 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2b22d242-0786-4cc6-90fc-ed25654ee072-clustermesh-secrets\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753663 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2b22d242-0786-4cc6-90fc-ed25654ee072-cilium-config-path\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:10.753719 kubelet[1948]: I0813 01:07:10.753678 1948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggblh\" (UniqueName: \"kubernetes.io/projected/2b22d242-0786-4cc6-90fc-ed25654ee072-kube-api-access-ggblh\") pod \"cilium-p6sl8\" (UID: \"2b22d242-0786-4cc6-90fc-ed25654ee072\") " pod="kube-system/cilium-p6sl8" Aug 13 01:07:11.030003 kubelet[1948]: E0813 01:07:11.029962 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:11.030943 env[1207]: time="2025-08-13T01:07:11.030892477Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p6sl8,Uid:2b22d242-0786-4cc6-90fc-ed25654ee072,Namespace:kube-system,Attempt:0,}" Aug 13 01:07:11.043471 env[1207]: time="2025-08-13T01:07:11.043381285Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 01:07:11.043471 env[1207]: time="2025-08-13T01:07:11.043419126Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 01:07:11.043711 env[1207]: time="2025-08-13T01:07:11.043428944Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 01:07:11.043711 env[1207]: time="2025-08-13T01:07:11.043571755Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6 pid=3915 runtime=io.containerd.runc.v2 Aug 13 01:07:11.046792 kubelet[1948]: I0813 01:07:11.046765 1948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b6eb70-f275-4f3b-8e62-cb57424e90b4" path="/var/lib/kubelet/pods/88b6eb70-f275-4f3b-8e62-cb57424e90b4/volumes" Aug 13 01:07:11.055444 systemd[1]: Started cri-containerd-dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6.scope. Aug 13 01:07:11.076893 env[1207]: time="2025-08-13T01:07:11.076828419Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p6sl8,Uid:2b22d242-0786-4cc6-90fc-ed25654ee072,Namespace:kube-system,Attempt:0,} returns sandbox id \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\"" Aug 13 01:07:11.077934 kubelet[1948]: E0813 01:07:11.077752 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:11.079724 env[1207]: time="2025-08-13T01:07:11.079699056Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 01:07:11.091373 env[1207]: time="2025-08-13T01:07:11.091322537Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b\"" Aug 13 01:07:11.091782 env[1207]: time="2025-08-13T01:07:11.091742110Z" level=info msg="StartContainer for \"a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b\"" Aug 13 01:07:11.097385 kubelet[1948]: E0813 01:07:11.097339 1948 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 01:07:11.105591 systemd[1]: Started cri-containerd-a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b.scope. Aug 13 01:07:11.128469 env[1207]: time="2025-08-13T01:07:11.128419651Z" level=info msg="StartContainer for \"a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b\" returns successfully" Aug 13 01:07:11.137544 systemd[1]: cri-containerd-a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b.scope: Deactivated successfully. Aug 13 01:07:11.160437 env[1207]: time="2025-08-13T01:07:11.160376448Z" level=info msg="shim disconnected" id=a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b Aug 13 01:07:11.160437 env[1207]: time="2025-08-13T01:07:11.160432525Z" level=warning msg="cleaning up after shim disconnected" id=a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b namespace=k8s.io Aug 13 01:07:11.160638 env[1207]: time="2025-08-13T01:07:11.160442594Z" level=info msg="cleaning up dead shim" Aug 13 01:07:11.167009 env[1207]: time="2025-08-13T01:07:11.166968330Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:11Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3997 runtime=io.containerd.runc.v2\n" Aug 13 01:07:11.612626 kubelet[1948]: E0813 01:07:11.612455 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:11.614491 env[1207]: time="2025-08-13T01:07:11.614451992Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 01:07:11.625324 kubelet[1948]: W0813 01:07:11.625248 1948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88b6eb70_f275_4f3b_8e62_cb57424e90b4.slice/cri-containerd-3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a.scope WatchSource:0}: container "3071fd632be32f764fd137ae7c0c3ce54d7d2112d9ba547d6a5101a099834e3a" in namespace "k8s.io": not found Aug 13 01:07:11.626829 env[1207]: time="2025-08-13T01:07:11.626790265Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee\"" Aug 13 01:07:11.627135 env[1207]: time="2025-08-13T01:07:11.627096132Z" level=info msg="StartContainer for \"34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee\"" Aug 13 01:07:11.641870 systemd[1]: Started cri-containerd-34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee.scope. Aug 13 01:07:11.668499 systemd[1]: cri-containerd-34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee.scope: Deactivated successfully. Aug 13 01:07:11.870818 env[1207]: time="2025-08-13T01:07:11.870406730Z" level=info msg="StartContainer for \"34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee\" returns successfully" Aug 13 01:07:11.885065 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee-rootfs.mount: Deactivated successfully. Aug 13 01:07:12.283745 env[1207]: time="2025-08-13T01:07:12.283677690Z" level=info msg="shim disconnected" id=34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee Aug 13 01:07:12.283745 env[1207]: time="2025-08-13T01:07:12.283730028Z" level=warning msg="cleaning up after shim disconnected" id=34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee namespace=k8s.io Aug 13 01:07:12.283745 env[1207]: time="2025-08-13T01:07:12.283738074Z" level=info msg="cleaning up dead shim" Aug 13 01:07:12.290077 env[1207]: time="2025-08-13T01:07:12.290022601Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4058 runtime=io.containerd.runc.v2\n" Aug 13 01:07:12.618874 kubelet[1948]: E0813 01:07:12.618461 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:12.620104 env[1207]: time="2025-08-13T01:07:12.620063971Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 01:07:12.640077 env[1207]: time="2025-08-13T01:07:12.639970123Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798\"" Aug 13 01:07:12.640520 env[1207]: time="2025-08-13T01:07:12.640491338Z" level=info msg="StartContainer for \"7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798\"" Aug 13 01:07:12.658931 systemd[1]: Started cri-containerd-7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798.scope. Aug 13 01:07:12.687273 env[1207]: time="2025-08-13T01:07:12.687166466Z" level=info msg="StartContainer for \"7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798\" returns successfully" Aug 13 01:07:12.689712 systemd[1]: cri-containerd-7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798.scope: Deactivated successfully. Aug 13 01:07:12.712530 env[1207]: time="2025-08-13T01:07:12.712468247Z" level=info msg="shim disconnected" id=7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798 Aug 13 01:07:12.712530 env[1207]: time="2025-08-13T01:07:12.712517380Z" level=warning msg="cleaning up after shim disconnected" id=7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798 namespace=k8s.io Aug 13 01:07:12.712530 env[1207]: time="2025-08-13T01:07:12.712525365Z" level=info msg="cleaning up dead shim" Aug 13 01:07:12.719422 env[1207]: time="2025-08-13T01:07:12.719386462Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4114 runtime=io.containerd.runc.v2\n" Aug 13 01:07:12.859645 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798-rootfs.mount: Deactivated successfully. Aug 13 01:07:13.313696 kubelet[1948]: I0813 01:07:13.313622 1948 setters.go:600] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-08-13T01:07:13Z","lastTransitionTime":"2025-08-13T01:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Aug 13 01:07:13.622697 kubelet[1948]: E0813 01:07:13.622548 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:13.625608 env[1207]: time="2025-08-13T01:07:13.625547842Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 01:07:13.638565 env[1207]: time="2025-08-13T01:07:13.638506430Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d\"" Aug 13 01:07:13.639082 env[1207]: time="2025-08-13T01:07:13.639040961Z" level=info msg="StartContainer for \"91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d\"" Aug 13 01:07:13.656384 systemd[1]: Started cri-containerd-91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d.scope. Aug 13 01:07:13.675490 systemd[1]: cri-containerd-91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d.scope: Deactivated successfully. Aug 13 01:07:13.677489 env[1207]: time="2025-08-13T01:07:13.677397366Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice/cri-containerd-91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d.scope/memory.events\": no such file or directory" Aug 13 01:07:13.680220 env[1207]: time="2025-08-13T01:07:13.680165176Z" level=info msg="StartContainer for \"91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d\" returns successfully" Aug 13 01:07:13.701443 env[1207]: time="2025-08-13T01:07:13.701381822Z" level=info msg="shim disconnected" id=91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d Aug 13 01:07:13.701443 env[1207]: time="2025-08-13T01:07:13.701442176Z" level=warning msg="cleaning up after shim disconnected" id=91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d namespace=k8s.io Aug 13 01:07:13.701443 env[1207]: time="2025-08-13T01:07:13.701450893Z" level=info msg="cleaning up dead shim" Aug 13 01:07:13.708553 env[1207]: time="2025-08-13T01:07:13.708468314Z" level=warning msg="cleanup warnings time=\"2025-08-13T01:07:13Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4169 runtime=io.containerd.runc.v2\n" Aug 13 01:07:13.860739 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d-rootfs.mount: Deactivated successfully. Aug 13 01:07:14.626531 kubelet[1948]: E0813 01:07:14.626488 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:14.628074 env[1207]: time="2025-08-13T01:07:14.628027692Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 01:07:14.642799 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4265775466.mount: Deactivated successfully. Aug 13 01:07:14.646870 env[1207]: time="2025-08-13T01:07:14.646818257Z" level=info msg="CreateContainer within sandbox \"dbeca10d5f9c81166db6458fae889e4091b5bca444a776da4dc570580e0a2aa6\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"d25d784e9af6a3b779e6f2346ef5913022614a2540ee8b21d70658417cbf673f\"" Aug 13 01:07:14.647404 env[1207]: time="2025-08-13T01:07:14.647368537Z" level=info msg="StartContainer for \"d25d784e9af6a3b779e6f2346ef5913022614a2540ee8b21d70658417cbf673f\"" Aug 13 01:07:14.664141 systemd[1]: Started cri-containerd-d25d784e9af6a3b779e6f2346ef5913022614a2540ee8b21d70658417cbf673f.scope. Aug 13 01:07:14.696692 env[1207]: time="2025-08-13T01:07:14.696626182Z" level=info msg="StartContainer for \"d25d784e9af6a3b779e6f2346ef5913022614a2540ee8b21d70658417cbf673f\" returns successfully" Aug 13 01:07:14.736819 kubelet[1948]: W0813 01:07:14.734825 1948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice/cri-containerd-a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b.scope WatchSource:0}: task a92e2a6ea35446e62e23c8e45db434bb6408733d3f8a96860e7bf05b55d8753b not found: not found Aug 13 01:07:14.973337 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Aug 13 01:07:15.631149 kubelet[1948]: E0813 01:07:15.631112 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:17.031413 kubelet[1948]: E0813 01:07:17.031334 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:17.635031 systemd-networkd[1020]: lxc_health: Link UP Aug 13 01:07:17.646711 systemd-networkd[1020]: lxc_health: Gained carrier Aug 13 01:07:17.647423 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Aug 13 01:07:17.845551 kubelet[1948]: W0813 01:07:17.845475 1948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice/cri-containerd-34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee.scope WatchSource:0}: task 34f64f4ed739f39cc3d9dde0db48db7d981ded74e811082ab0f496789b17f1ee not found: not found Aug 13 01:07:19.010146 systemd-networkd[1020]: lxc_health: Gained IPv6LL Aug 13 01:07:19.031798 kubelet[1948]: E0813 01:07:19.031760 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:19.048309 kubelet[1948]: I0813 01:07:19.048027 1948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-p6sl8" podStartSLOduration=9.048008944 podStartE2EDuration="9.048008944s" podCreationTimestamp="2025-08-13 01:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 01:07:15.646237534 +0000 UTC m=+114.683797182" watchObservedRunningTime="2025-08-13 01:07:19.048008944 +0000 UTC m=+118.085568602" Aug 13 01:07:19.639180 kubelet[1948]: E0813 01:07:19.639144 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:20.640886 kubelet[1948]: E0813 01:07:20.640852 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:20.825958 systemd[1]: run-containerd-runc-k8s.io-d25d784e9af6a3b779e6f2346ef5913022614a2540ee8b21d70658417cbf673f-runc.RucOFr.mount: Deactivated successfully. Aug 13 01:07:20.958110 kubelet[1948]: W0813 01:07:20.958062 1948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice/cri-containerd-7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798.scope WatchSource:0}: task 7c00d8611d1188fbe1bf018326766832801ced855801afc7442e922360248798 not found: not found Aug 13 01:07:21.029942 env[1207]: time="2025-08-13T01:07:21.029891050Z" level=info msg="StopPodSandbox for \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\"" Aug 13 01:07:21.030341 env[1207]: time="2025-08-13T01:07:21.030037617Z" level=info msg="TearDown network for sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" successfully" Aug 13 01:07:21.030341 env[1207]: time="2025-08-13T01:07:21.030068685Z" level=info msg="StopPodSandbox for \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" returns successfully" Aug 13 01:07:21.030558 env[1207]: time="2025-08-13T01:07:21.030520378Z" level=info msg="RemovePodSandbox for \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\"" Aug 13 01:07:21.030619 env[1207]: time="2025-08-13T01:07:21.030565724Z" level=info msg="Forcibly stopping sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\"" Aug 13 01:07:21.030693 env[1207]: time="2025-08-13T01:07:21.030670161Z" level=info msg="TearDown network for sandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" successfully" Aug 13 01:07:21.038457 env[1207]: time="2025-08-13T01:07:21.038429832Z" level=info msg="RemovePodSandbox \"7627560148c71962129620d2b62688cac73cafdcd9bab602574b63487f0e32c8\" returns successfully" Aug 13 01:07:21.038909 env[1207]: time="2025-08-13T01:07:21.038882145Z" level=info msg="StopPodSandbox for \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\"" Aug 13 01:07:21.039009 env[1207]: time="2025-08-13T01:07:21.038963630Z" level=info msg="TearDown network for sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" successfully" Aug 13 01:07:21.039040 env[1207]: time="2025-08-13T01:07:21.039008033Z" level=info msg="StopPodSandbox for \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" returns successfully" Aug 13 01:07:21.039341 env[1207]: time="2025-08-13T01:07:21.039316736Z" level=info msg="RemovePodSandbox for \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\"" Aug 13 01:07:21.039405 env[1207]: time="2025-08-13T01:07:21.039346243Z" level=info msg="Forcibly stopping sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\"" Aug 13 01:07:21.039433 env[1207]: time="2025-08-13T01:07:21.039408109Z" level=info msg="TearDown network for sandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" successfully" Aug 13 01:07:21.042204 env[1207]: time="2025-08-13T01:07:21.042170905Z" level=info msg="RemovePodSandbox \"0ea72545bc48b7a02e20b1f4fd2c962aa47bc38630aafbbaeda7cf1b29736fa1\" returns successfully" Aug 13 01:07:21.042458 env[1207]: time="2025-08-13T01:07:21.042428001Z" level=info msg="StopPodSandbox for \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\"" Aug 13 01:07:21.042545 env[1207]: time="2025-08-13T01:07:21.042499556Z" level=info msg="TearDown network for sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" successfully" Aug 13 01:07:21.042545 env[1207]: time="2025-08-13T01:07:21.042536746Z" level=info msg="StopPodSandbox for \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" returns successfully" Aug 13 01:07:21.042844 env[1207]: time="2025-08-13T01:07:21.042802548Z" level=info msg="RemovePodSandbox for \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\"" Aug 13 01:07:21.042844 env[1207]: time="2025-08-13T01:07:21.042826723Z" level=info msg="Forcibly stopping sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\"" Aug 13 01:07:21.043058 env[1207]: time="2025-08-13T01:07:21.042876568Z" level=info msg="TearDown network for sandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" successfully" Aug 13 01:07:21.045701 env[1207]: time="2025-08-13T01:07:21.045680010Z" level=info msg="RemovePodSandbox \"be660848133d485f9871670d67d60ad49911fc332a9c4cf62371bbab27072d3f\" returns successfully" Aug 13 01:07:22.980733 sshd[3779]: pam_unix(sshd:session): session closed for user core Aug 13 01:07:22.983753 systemd[1]: sshd@29-10.0.0.132:22-10.0.0.1:55256.service: Deactivated successfully. Aug 13 01:07:22.984547 systemd[1]: session-30.scope: Deactivated successfully. Aug 13 01:07:22.985111 systemd-logind[1195]: Session 30 logged out. Waiting for processes to exit. Aug 13 01:07:22.986011 systemd-logind[1195]: Removed session 30. Aug 13 01:07:23.043896 kubelet[1948]: E0813 01:07:23.043865 1948 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Aug 13 01:07:24.063179 kubelet[1948]: W0813 01:07:24.063116 1948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b22d242_0786_4cc6_90fc_ed25654ee072.slice/cri-containerd-91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d.scope WatchSource:0}: task 91d57266f448e0a6629181135710a5437ff870fce8a237391498a222dcd3996d not found: not found