Nov 6 00:14:14.927460 kernel: Linux version 6.12.54-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed Nov 5 22:12:28 -00 2025 Nov 6 00:14:14.927493 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=59ca0b9e28689480cec05e5a7a50ffb2fd81e743a9e2986eb3bceb3b87f6702e Nov 6 00:14:14.927508 kernel: BIOS-provided physical RAM map: Nov 6 00:14:14.927516 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Nov 6 00:14:14.927525 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Nov 6 00:14:14.927533 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Nov 6 00:14:14.927544 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Nov 6 00:14:14.927553 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Nov 6 00:14:14.927565 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Nov 6 00:14:14.927583 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Nov 6 00:14:14.927595 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Nov 6 00:14:14.927610 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Nov 6 00:14:14.927630 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Nov 6 00:14:14.927644 kernel: NX (Execute Disable) protection: active Nov 6 00:14:14.927665 kernel: APIC: Static calls initialized Nov 6 00:14:14.927675 kernel: SMBIOS 2.8 present. Nov 6 00:14:14.927922 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Nov 6 00:14:14.927934 kernel: DMI: Memory slots populated: 1/1 Nov 6 00:14:14.927943 kernel: Hypervisor detected: KVM Nov 6 00:14:14.927953 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Nov 6 00:14:14.927962 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Nov 6 00:14:14.927971 kernel: kvm-clock: using sched offset of 4495293049 cycles Nov 6 00:14:14.927981 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Nov 6 00:14:14.927991 kernel: tsc: Detected 2794.748 MHz processor Nov 6 00:14:14.928001 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 6 00:14:14.928011 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 6 00:14:14.928024 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Nov 6 00:14:14.928033 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Nov 6 00:14:14.928043 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 6 00:14:14.928052 kernel: Using GB pages for direct mapping Nov 6 00:14:14.928062 kernel: ACPI: Early table checksum verification disabled Nov 6 00:14:14.928071 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Nov 6 00:14:14.928080 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928089 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928099 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928111 kernel: ACPI: FACS 0x000000009CFE0000 000040 Nov 6 00:14:14.928120 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928129 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928137 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928146 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 6 00:14:14.928159 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Nov 6 00:14:14.928172 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Nov 6 00:14:14.928202 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Nov 6 00:14:14.928214 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Nov 6 00:14:14.928224 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Nov 6 00:14:14.928234 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Nov 6 00:14:14.928244 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Nov 6 00:14:14.928264 kernel: No NUMA configuration found Nov 6 00:14:14.928275 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Nov 6 00:14:14.928312 kernel: NODE_DATA(0) allocated [mem 0x9cfd4dc0-0x9cfdbfff] Nov 6 00:14:14.928332 kernel: Zone ranges: Nov 6 00:14:14.928342 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 6 00:14:14.928351 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Nov 6 00:14:14.928359 kernel: Normal empty Nov 6 00:14:14.928368 kernel: Device empty Nov 6 00:14:14.928377 kernel: Movable zone start for each node Nov 6 00:14:14.928386 kernel: Early memory node ranges Nov 6 00:14:14.928394 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Nov 6 00:14:14.928412 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Nov 6 00:14:14.928668 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Nov 6 00:14:14.928682 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 6 00:14:14.928693 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Nov 6 00:14:14.928703 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Nov 6 00:14:14.928715 kernel: ACPI: PM-Timer IO Port: 0x608 Nov 6 00:14:14.928725 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Nov 6 00:14:14.928734 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Nov 6 00:14:14.928743 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 6 00:14:14.928755 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Nov 6 00:14:14.928769 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 6 00:14:14.928779 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Nov 6 00:14:14.928788 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Nov 6 00:14:14.928798 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 6 00:14:14.928807 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 6 00:14:14.928817 kernel: TSC deadline timer available Nov 6 00:14:14.928826 kernel: CPU topo: Max. logical packages: 1 Nov 6 00:14:14.928836 kernel: CPU topo: Max. logical dies: 1 Nov 6 00:14:14.928846 kernel: CPU topo: Max. dies per package: 1 Nov 6 00:14:14.928858 kernel: CPU topo: Max. threads per core: 1 Nov 6 00:14:14.928868 kernel: CPU topo: Num. cores per package: 4 Nov 6 00:14:14.928878 kernel: CPU topo: Num. threads per package: 4 Nov 6 00:14:14.928888 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Nov 6 00:14:14.928898 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Nov 6 00:14:14.928907 kernel: kvm-guest: KVM setup pv remote TLB flush Nov 6 00:14:14.928917 kernel: kvm-guest: setup PV sched yield Nov 6 00:14:14.928926 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Nov 6 00:14:14.928946 kernel: Booting paravirtualized kernel on KVM Nov 6 00:14:14.928961 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 6 00:14:14.928971 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Nov 6 00:14:14.928981 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Nov 6 00:14:14.928992 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Nov 6 00:14:14.929002 kernel: pcpu-alloc: [0] 0 1 2 3 Nov 6 00:14:14.929012 kernel: kvm-guest: PV spinlocks enabled Nov 6 00:14:14.929025 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Nov 6 00:14:14.929036 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=59ca0b9e28689480cec05e5a7a50ffb2fd81e743a9e2986eb3bceb3b87f6702e Nov 6 00:14:14.929047 kernel: random: crng init done Nov 6 00:14:14.929060 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Nov 6 00:14:14.929070 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 6 00:14:14.929080 kernel: Fallback order for Node 0: 0 Nov 6 00:14:14.929089 kernel: Built 1 zonelists, mobility grouping on. Total pages: 642938 Nov 6 00:14:14.929098 kernel: Policy zone: DMA32 Nov 6 00:14:14.929108 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 6 00:14:14.929117 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Nov 6 00:14:14.929126 kernel: ftrace: allocating 40021 entries in 157 pages Nov 6 00:14:14.929136 kernel: ftrace: allocated 157 pages with 5 groups Nov 6 00:14:14.929148 kernel: Dynamic Preempt: voluntary Nov 6 00:14:14.929158 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 6 00:14:14.929169 kernel: rcu: RCU event tracing is enabled. Nov 6 00:14:14.929179 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Nov 6 00:14:14.929204 kernel: Trampoline variant of Tasks RCU enabled. Nov 6 00:14:14.929218 kernel: Rude variant of Tasks RCU enabled. Nov 6 00:14:14.929228 kernel: Tracing variant of Tasks RCU enabled. Nov 6 00:14:14.929237 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 6 00:14:14.929247 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Nov 6 00:14:14.929259 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 6 00:14:14.929269 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 6 00:14:14.929279 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 6 00:14:14.929289 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Nov 6 00:14:14.929299 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 6 00:14:14.929318 kernel: Console: colour VGA+ 80x25 Nov 6 00:14:14.929330 kernel: printk: legacy console [ttyS0] enabled Nov 6 00:14:14.929341 kernel: ACPI: Core revision 20240827 Nov 6 00:14:14.929351 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Nov 6 00:14:14.929364 kernel: APIC: Switch to symmetric I/O mode setup Nov 6 00:14:14.929375 kernel: x2apic enabled Nov 6 00:14:14.929384 kernel: APIC: Switched APIC routing to: physical x2apic Nov 6 00:14:14.929402 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Nov 6 00:14:14.929413 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Nov 6 00:14:14.929423 kernel: kvm-guest: setup PV IPIs Nov 6 00:14:14.929433 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Nov 6 00:14:14.929444 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Nov 6 00:14:14.929458 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Nov 6 00:14:14.929469 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Nov 6 00:14:14.929479 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Nov 6 00:14:14.929489 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Nov 6 00:14:14.929500 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 6 00:14:14.929510 kernel: Spectre V2 : Mitigation: Retpolines Nov 6 00:14:14.929520 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Nov 6 00:14:14.929530 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Nov 6 00:14:14.929543 kernel: active return thunk: retbleed_return_thunk Nov 6 00:14:14.929553 kernel: RETBleed: Mitigation: untrained return thunk Nov 6 00:14:14.929563 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 6 00:14:14.929585 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 6 00:14:14.929596 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Nov 6 00:14:14.929607 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Nov 6 00:14:14.929617 kernel: active return thunk: srso_return_thunk Nov 6 00:14:14.929627 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Nov 6 00:14:14.929638 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 6 00:14:14.929652 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 6 00:14:14.929664 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 6 00:14:14.929677 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 6 00:14:14.929689 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Nov 6 00:14:14.929702 kernel: Freeing SMP alternatives memory: 32K Nov 6 00:14:14.929715 kernel: pid_max: default: 32768 minimum: 301 Nov 6 00:14:14.929727 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Nov 6 00:14:14.929739 kernel: landlock: Up and running. Nov 6 00:14:14.929751 kernel: SELinux: Initializing. Nov 6 00:14:14.929770 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 6 00:14:14.929784 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 6 00:14:14.929796 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Nov 6 00:14:14.929809 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Nov 6 00:14:14.929822 kernel: ... version: 0 Nov 6 00:14:14.929835 kernel: ... bit width: 48 Nov 6 00:14:14.929848 kernel: ... generic registers: 6 Nov 6 00:14:14.929861 kernel: ... value mask: 0000ffffffffffff Nov 6 00:14:14.929874 kernel: ... max period: 00007fffffffffff Nov 6 00:14:14.929886 kernel: ... fixed-purpose events: 0 Nov 6 00:14:14.929906 kernel: ... event mask: 000000000000003f Nov 6 00:14:14.929921 kernel: signal: max sigframe size: 1776 Nov 6 00:14:14.929942 kernel: rcu: Hierarchical SRCU implementation. Nov 6 00:14:14.929958 kernel: rcu: Max phase no-delay instances is 400. Nov 6 00:14:14.929969 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Nov 6 00:14:14.929979 kernel: smp: Bringing up secondary CPUs ... Nov 6 00:14:14.929988 kernel: smpboot: x86: Booting SMP configuration: Nov 6 00:14:14.930010 kernel: .... node #0, CPUs: #1 #2 #3 Nov 6 00:14:14.930041 kernel: smp: Brought up 1 node, 4 CPUs Nov 6 00:14:14.930075 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Nov 6 00:14:14.930291 kernel: Memory: 2422764K/2571752K available (14336K kernel code, 2436K rwdata, 26048K rodata, 45548K init, 1180K bss, 143048K reserved, 0K cma-reserved) Nov 6 00:14:14.930308 kernel: devtmpfs: initialized Nov 6 00:14:14.930319 kernel: x86/mm: Memory block size: 128MB Nov 6 00:14:14.930330 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 6 00:14:14.930340 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Nov 6 00:14:14.930351 kernel: pinctrl core: initialized pinctrl subsystem Nov 6 00:14:14.930365 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 6 00:14:14.930380 kernel: audit: initializing netlink subsys (disabled) Nov 6 00:14:14.930391 kernel: audit: type=2000 audit(1762388052.311:1): state=initialized audit_enabled=0 res=1 Nov 6 00:14:14.930401 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 6 00:14:14.930412 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 6 00:14:14.930422 kernel: cpuidle: using governor menu Nov 6 00:14:14.930433 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 6 00:14:14.930443 kernel: dca service started, version 1.12.1 Nov 6 00:14:14.930454 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Nov 6 00:14:14.930466 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Nov 6 00:14:14.930480 kernel: PCI: Using configuration type 1 for base access Nov 6 00:14:14.930491 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 6 00:14:14.930501 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 6 00:14:14.930513 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 6 00:14:14.930523 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 6 00:14:14.930534 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 6 00:14:14.930545 kernel: ACPI: Added _OSI(Module Device) Nov 6 00:14:14.930555 kernel: ACPI: Added _OSI(Processor Device) Nov 6 00:14:14.930566 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 6 00:14:14.930590 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 6 00:14:14.930603 kernel: ACPI: Interpreter enabled Nov 6 00:14:14.930617 kernel: ACPI: PM: (supports S0 S3 S5) Nov 6 00:14:14.930630 kernel: ACPI: Using IOAPIC for interrupt routing Nov 6 00:14:14.930644 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 6 00:14:14.930657 kernel: PCI: Using E820 reservations for host bridge windows Nov 6 00:14:14.930671 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Nov 6 00:14:14.930685 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 6 00:14:14.930963 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 6 00:14:14.931123 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Nov 6 00:14:14.931430 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Nov 6 00:14:14.931446 kernel: PCI host bridge to bus 0000:00 Nov 6 00:14:14.931620 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 6 00:14:14.931961 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 6 00:14:14.932103 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 6 00:14:14.932487 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Nov 6 00:14:14.932639 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Nov 6 00:14:14.932771 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Nov 6 00:14:14.932904 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 6 00:14:14.933103 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Nov 6 00:14:14.933285 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Nov 6 00:14:14.933441 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Nov 6 00:14:14.933605 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Nov 6 00:14:14.933755 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Nov 6 00:14:14.933914 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 6 00:14:14.934081 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Nov 6 00:14:14.934255 kernel: pci 0000:00:02.0: BAR 0 [io 0xc0c0-0xc0df] Nov 6 00:14:14.934410 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Nov 6 00:14:14.934582 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Nov 6 00:14:14.934747 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Nov 6 00:14:14.934898 kernel: pci 0000:00:03.0: BAR 0 [io 0xc000-0xc07f] Nov 6 00:14:14.935048 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Nov 6 00:14:14.935218 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Nov 6 00:14:14.935414 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Nov 6 00:14:14.935583 kernel: pci 0000:00:04.0: BAR 0 [io 0xc0e0-0xc0ff] Nov 6 00:14:14.935748 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebd3000-0xfebd3fff] Nov 6 00:14:14.935957 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe008000-0xfe00bfff 64bit pref] Nov 6 00:14:14.936398 kernel: pci 0000:00:04.0: ROM [mem 0xfeb80000-0xfebbffff pref] Nov 6 00:14:14.936630 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Nov 6 00:14:14.936876 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Nov 6 00:14:14.937213 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Nov 6 00:14:14.937387 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc100-0xc11f] Nov 6 00:14:14.937547 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd4000-0xfebd4fff] Nov 6 00:14:14.937809 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Nov 6 00:14:14.937964 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Nov 6 00:14:14.937979 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Nov 6 00:14:14.937991 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Nov 6 00:14:14.938002 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Nov 6 00:14:14.938013 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Nov 6 00:14:14.938029 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Nov 6 00:14:14.938039 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Nov 6 00:14:14.938050 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Nov 6 00:14:14.938061 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Nov 6 00:14:14.938072 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Nov 6 00:14:14.938082 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Nov 6 00:14:14.938093 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Nov 6 00:14:14.938104 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Nov 6 00:14:14.938115 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Nov 6 00:14:14.938129 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Nov 6 00:14:14.938140 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Nov 6 00:14:14.938151 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Nov 6 00:14:14.938161 kernel: iommu: Default domain type: Translated Nov 6 00:14:14.938172 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 6 00:14:14.938183 kernel: PCI: Using ACPI for IRQ routing Nov 6 00:14:14.938213 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 6 00:14:14.938225 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Nov 6 00:14:14.938236 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Nov 6 00:14:14.938395 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Nov 6 00:14:14.938548 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Nov 6 00:14:14.938713 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 6 00:14:14.938728 kernel: vgaarb: loaded Nov 6 00:14:14.938739 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Nov 6 00:14:14.938749 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Nov 6 00:14:14.938760 kernel: clocksource: Switched to clocksource kvm-clock Nov 6 00:14:14.938770 kernel: VFS: Disk quotas dquot_6.6.0 Nov 6 00:14:14.938781 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 6 00:14:14.938796 kernel: pnp: PnP ACPI init Nov 6 00:14:14.938965 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Nov 6 00:14:14.938981 kernel: pnp: PnP ACPI: found 6 devices Nov 6 00:14:14.938993 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 6 00:14:14.939004 kernel: NET: Registered PF_INET protocol family Nov 6 00:14:14.939015 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 6 00:14:14.939026 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Nov 6 00:14:14.939037 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 6 00:14:14.939052 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 6 00:14:14.939063 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Nov 6 00:14:14.939074 kernel: TCP: Hash tables configured (established 32768 bind 32768) Nov 6 00:14:14.939086 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 6 00:14:14.939097 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 6 00:14:14.939107 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 6 00:14:14.939118 kernel: NET: Registered PF_XDP protocol family Nov 6 00:14:14.939280 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 6 00:14:14.939418 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 6 00:14:14.939558 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 6 00:14:14.939705 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Nov 6 00:14:14.939838 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Nov 6 00:14:14.939969 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Nov 6 00:14:14.939983 kernel: PCI: CLS 0 bytes, default 64 Nov 6 00:14:14.939995 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns Nov 6 00:14:14.940006 kernel: Initialise system trusted keyrings Nov 6 00:14:14.940017 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Nov 6 00:14:14.940032 kernel: Key type asymmetric registered Nov 6 00:14:14.940042 kernel: Asymmetric key parser 'x509' registered Nov 6 00:14:14.940053 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Nov 6 00:14:14.940064 kernel: io scheduler mq-deadline registered Nov 6 00:14:14.940074 kernel: io scheduler kyber registered Nov 6 00:14:14.940085 kernel: io scheduler bfq registered Nov 6 00:14:14.940096 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 6 00:14:14.940108 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Nov 6 00:14:14.940119 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Nov 6 00:14:14.940132 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Nov 6 00:14:14.940143 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 6 00:14:14.940154 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 6 00:14:14.940165 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Nov 6 00:14:14.940175 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Nov 6 00:14:14.940201 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Nov 6 00:14:14.940358 kernel: rtc_cmos 00:04: RTC can wake from S4 Nov 6 00:14:14.940374 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Nov 6 00:14:14.940525 kernel: rtc_cmos 00:04: registered as rtc0 Nov 6 00:14:14.940671 kernel: rtc_cmos 00:04: setting system clock to 2025-11-06T00:14:14 UTC (1762388054) Nov 6 00:14:14.940812 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Nov 6 00:14:14.940828 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Nov 6 00:14:14.940840 kernel: NET: Registered PF_INET6 protocol family Nov 6 00:14:14.940850 kernel: Segment Routing with IPv6 Nov 6 00:14:14.940861 kernel: In-situ OAM (IOAM) with IPv6 Nov 6 00:14:14.940872 kernel: NET: Registered PF_PACKET protocol family Nov 6 00:14:14.940883 kernel: Key type dns_resolver registered Nov 6 00:14:14.940898 kernel: IPI shorthand broadcast: enabled Nov 6 00:14:14.940909 kernel: sched_clock: Marking stable (3125003304, 238529751)->(3430471258, -66938203) Nov 6 00:14:14.940920 kernel: registered taskstats version 1 Nov 6 00:14:14.940931 kernel: Loading compiled-in X.509 certificates Nov 6 00:14:14.940942 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.54-flatcar: f906521ec29cbf079ae365554bad8eb8ed6ecb31' Nov 6 00:14:14.940953 kernel: Demotion targets for Node 0: null Nov 6 00:14:14.940963 kernel: Key type .fscrypt registered Nov 6 00:14:14.940974 kernel: Key type fscrypt-provisioning registered Nov 6 00:14:14.940985 kernel: ima: No TPM chip found, activating TPM-bypass! Nov 6 00:14:14.940998 kernel: ima: Allocated hash algorithm: sha1 Nov 6 00:14:14.941009 kernel: ima: No architecture policies found Nov 6 00:14:14.941020 kernel: clk: Disabling unused clocks Nov 6 00:14:14.941031 kernel: Warning: unable to open an initial console. Nov 6 00:14:14.941042 kernel: Freeing unused kernel image (initmem) memory: 45548K Nov 6 00:14:14.941053 kernel: Write protecting the kernel read-only data: 40960k Nov 6 00:14:14.941064 kernel: Freeing unused kernel image (rodata/data gap) memory: 576K Nov 6 00:14:14.941075 kernel: Run /init as init process Nov 6 00:14:14.941086 kernel: with arguments: Nov 6 00:14:14.941099 kernel: /init Nov 6 00:14:14.941109 kernel: with environment: Nov 6 00:14:14.941120 kernel: HOME=/ Nov 6 00:14:14.941130 kernel: TERM=linux Nov 6 00:14:14.941142 systemd[1]: Successfully made /usr/ read-only. Nov 6 00:14:14.941159 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 6 00:14:14.941199 systemd[1]: Detected virtualization kvm. Nov 6 00:14:14.941212 systemd[1]: Detected architecture x86-64. Nov 6 00:14:14.941224 systemd[1]: Running in initrd. Nov 6 00:14:14.941236 systemd[1]: No hostname configured, using default hostname. Nov 6 00:14:14.941248 systemd[1]: Hostname set to . Nov 6 00:14:14.941260 systemd[1]: Initializing machine ID from VM UUID. Nov 6 00:14:14.941272 systemd[1]: Queued start job for default target initrd.target. Nov 6 00:14:14.941284 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 6 00:14:14.941299 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 6 00:14:14.941312 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 6 00:14:14.941326 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 6 00:14:14.941338 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 6 00:14:14.941351 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 6 00:14:14.941365 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 6 00:14:14.941379 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 6 00:14:14.941391 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 6 00:14:14.941403 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 6 00:14:14.941414 systemd[1]: Reached target paths.target - Path Units. Nov 6 00:14:14.941426 systemd[1]: Reached target slices.target - Slice Units. Nov 6 00:14:14.941438 systemd[1]: Reached target swap.target - Swaps. Nov 6 00:14:14.941450 systemd[1]: Reached target timers.target - Timer Units. Nov 6 00:14:14.941462 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 6 00:14:14.941474 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 6 00:14:14.941488 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 6 00:14:14.941500 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Nov 6 00:14:14.941512 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 6 00:14:14.941524 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 6 00:14:14.941536 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 6 00:14:14.941548 systemd[1]: Reached target sockets.target - Socket Units. Nov 6 00:14:14.941560 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 6 00:14:14.941582 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 6 00:14:14.941594 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 6 00:14:14.941607 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Nov 6 00:14:14.941619 systemd[1]: Starting systemd-fsck-usr.service... Nov 6 00:14:14.941631 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 6 00:14:14.941643 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 6 00:14:14.941655 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 6 00:14:14.941669 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 6 00:14:14.941682 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 6 00:14:14.941695 systemd[1]: Finished systemd-fsck-usr.service. Nov 6 00:14:14.941709 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 6 00:14:14.941761 systemd-journald[201]: Collecting audit messages is disabled. Nov 6 00:14:14.941803 systemd-journald[201]: Journal started Nov 6 00:14:14.941842 systemd-journald[201]: Runtime Journal (/run/log/journal/a9a6cc210c29401d835a48f88cff5f32) is 6M, max 48.3M, 42.2M free. Nov 6 00:14:14.943456 systemd-modules-load[203]: Inserted module 'overlay' Nov 6 00:14:14.998720 systemd[1]: Started systemd-journald.service - Journal Service. Nov 6 00:14:14.998748 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 6 00:14:14.998761 kernel: Bridge firewalling registered Nov 6 00:14:14.973588 systemd-modules-load[203]: Inserted module 'br_netfilter' Nov 6 00:14:15.001333 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 6 00:14:15.004028 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 6 00:14:15.004934 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 6 00:14:15.009820 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 6 00:14:15.012503 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 6 00:14:15.016775 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 6 00:14:15.055640 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 6 00:14:15.066706 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 6 00:14:15.070106 systemd-tmpfiles[222]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Nov 6 00:14:15.071296 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 6 00:14:15.075886 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 6 00:14:15.076950 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 6 00:14:15.086462 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 6 00:14:15.089604 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 6 00:14:15.124758 dracut-cmdline[242]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=59ca0b9e28689480cec05e5a7a50ffb2fd81e743a9e2986eb3bceb3b87f6702e Nov 6 00:14:15.142610 systemd-resolved[243]: Positive Trust Anchors: Nov 6 00:14:15.142626 systemd-resolved[243]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 6 00:14:15.142656 systemd-resolved[243]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 6 00:14:15.145233 systemd-resolved[243]: Defaulting to hostname 'linux'. Nov 6 00:14:15.146339 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 6 00:14:15.185610 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 6 00:14:15.283212 kernel: SCSI subsystem initialized Nov 6 00:14:15.293208 kernel: Loading iSCSI transport class v2.0-870. Nov 6 00:14:15.304213 kernel: iscsi: registered transport (tcp) Nov 6 00:14:15.327208 kernel: iscsi: registered transport (qla4xxx) Nov 6 00:14:15.327250 kernel: QLogic iSCSI HBA Driver Nov 6 00:14:15.348689 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 6 00:14:15.370300 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 6 00:14:15.377313 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 6 00:14:15.434317 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 6 00:14:15.439094 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 6 00:14:15.499218 kernel: raid6: avx2x4 gen() 29736 MB/s Nov 6 00:14:15.516213 kernel: raid6: avx2x2 gen() 29939 MB/s Nov 6 00:14:15.534027 kernel: raid6: avx2x1 gen() 24959 MB/s Nov 6 00:14:15.534043 kernel: raid6: using algorithm avx2x2 gen() 29939 MB/s Nov 6 00:14:15.552007 kernel: raid6: .... xor() 19313 MB/s, rmw enabled Nov 6 00:14:15.552023 kernel: raid6: using avx2x2 recovery algorithm Nov 6 00:14:15.621226 kernel: xor: automatically using best checksumming function avx Nov 6 00:14:15.820245 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 6 00:14:15.829465 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 6 00:14:15.833090 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 6 00:14:15.867712 systemd-udevd[454]: Using default interface naming scheme 'v255'. Nov 6 00:14:15.874886 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 6 00:14:15.881120 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 6 00:14:15.914894 dracut-pre-trigger[465]: rd.md=0: removing MD RAID activation Nov 6 00:14:15.944471 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 6 00:14:15.946632 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 6 00:14:16.029322 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 6 00:14:16.036218 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 6 00:14:16.080371 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Nov 6 00:14:16.123248 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Nov 6 00:14:16.130521 kernel: cryptd: max_cpu_qlen set to 1000 Nov 6 00:14:16.130558 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Nov 6 00:14:16.130573 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 6 00:14:16.132221 kernel: GPT:9289727 != 19775487 Nov 6 00:14:16.132247 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 6 00:14:16.134914 kernel: GPT:9289727 != 19775487 Nov 6 00:14:16.134942 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 6 00:14:16.134959 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 6 00:14:16.153205 kernel: libata version 3.00 loaded. Nov 6 00:14:16.155203 kernel: AES CTR mode by8 optimization enabled Nov 6 00:14:16.178495 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 6 00:14:16.182275 kernel: ahci 0000:00:1f.2: version 3.0 Nov 6 00:14:16.182466 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Nov 6 00:14:16.178642 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 6 00:14:16.196139 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Nov 6 00:14:16.196427 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Nov 6 00:14:16.196584 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Nov 6 00:14:16.189975 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 6 00:14:16.200569 kernel: scsi host0: ahci Nov 6 00:14:16.202114 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 6 00:14:16.209000 kernel: scsi host1: ahci Nov 6 00:14:16.209266 kernel: scsi host2: ahci Nov 6 00:14:16.209418 kernel: scsi host3: ahci Nov 6 00:14:16.206061 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Nov 6 00:14:16.228872 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Nov 6 00:14:16.245080 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Nov 6 00:14:16.264209 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 6 00:14:16.270358 kernel: scsi host4: ahci Nov 6 00:14:16.270815 kernel: scsi host5: ahci Nov 6 00:14:16.270974 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 lpm-pol 1 Nov 6 00:14:16.274023 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 lpm-pol 1 Nov 6 00:14:16.274046 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 lpm-pol 1 Nov 6 00:14:16.277254 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 lpm-pol 1 Nov 6 00:14:16.277273 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 lpm-pol 1 Nov 6 00:14:16.280899 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 lpm-pol 1 Nov 6 00:14:16.298073 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Nov 6 00:14:16.298741 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Nov 6 00:14:16.300742 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 6 00:14:16.397517 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 6 00:14:16.445144 disk-uuid[614]: Primary Header is updated. Nov 6 00:14:16.445144 disk-uuid[614]: Secondary Entries is updated. Nov 6 00:14:16.445144 disk-uuid[614]: Secondary Header is updated. Nov 6 00:14:16.450483 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 6 00:14:16.457221 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 6 00:14:16.588652 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 6 00:14:16.588729 kernel: ata2: SATA link down (SStatus 0 SControl 300) Nov 6 00:14:16.589239 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Nov 6 00:14:16.591501 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 6 00:14:16.592216 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 6 00:14:16.593220 kernel: ata1: SATA link down (SStatus 0 SControl 300) Nov 6 00:14:16.594223 kernel: ata3.00: LPM support broken, forcing max_power Nov 6 00:14:16.595993 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Nov 6 00:14:16.596021 kernel: ata3.00: applying bridge limits Nov 6 00:14:16.597919 kernel: ata3.00: LPM support broken, forcing max_power Nov 6 00:14:16.597943 kernel: ata3.00: configured for UDMA/100 Nov 6 00:14:16.599482 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Nov 6 00:14:16.648799 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Nov 6 00:14:16.649274 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Nov 6 00:14:16.667232 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Nov 6 00:14:17.075137 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 6 00:14:17.077552 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 6 00:14:17.081160 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 6 00:14:17.083457 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 6 00:14:17.086262 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 6 00:14:17.121201 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 6 00:14:17.474230 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 6 00:14:17.474356 disk-uuid[616]: The operation has completed successfully. Nov 6 00:14:17.505879 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 6 00:14:17.506003 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 6 00:14:17.540618 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 6 00:14:17.553271 sh[644]: Success Nov 6 00:14:17.575764 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 6 00:14:17.575821 kernel: device-mapper: uevent: version 1.0.3 Nov 6 00:14:17.577451 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Nov 6 00:14:17.587213 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Nov 6 00:14:17.616673 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 6 00:14:17.634246 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 6 00:14:17.651173 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 6 00:14:17.659288 kernel: BTRFS: device fsid 85d805c5-984c-4a6a-aaeb-49fff3689175 devid 1 transid 38 /dev/mapper/usr (253:0) scanned by mount (656) Nov 6 00:14:17.659311 kernel: BTRFS info (device dm-0): first mount of filesystem 85d805c5-984c-4a6a-aaeb-49fff3689175 Nov 6 00:14:17.659322 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 6 00:14:17.666372 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 6 00:14:17.666428 kernel: BTRFS info (device dm-0): enabling free space tree Nov 6 00:14:17.667600 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 6 00:14:17.668903 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Nov 6 00:14:17.670962 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 6 00:14:17.671779 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 6 00:14:17.687775 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 6 00:14:17.708229 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (687) Nov 6 00:14:17.708256 kernel: BTRFS info (device vda6): first mount of filesystem ca2bb832-66d5-4dca-a6d2-cbf7440d9381 Nov 6 00:14:17.711286 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 6 00:14:17.715758 kernel: BTRFS info (device vda6): turning on async discard Nov 6 00:14:17.715797 kernel: BTRFS info (device vda6): enabling free space tree Nov 6 00:14:17.722228 kernel: BTRFS info (device vda6): last unmount of filesystem ca2bb832-66d5-4dca-a6d2-cbf7440d9381 Nov 6 00:14:17.722323 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 6 00:14:17.727174 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 6 00:14:17.893806 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 6 00:14:17.900331 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 6 00:14:17.911679 ignition[735]: Ignition 2.22.0 Nov 6 00:14:17.911691 ignition[735]: Stage: fetch-offline Nov 6 00:14:17.911724 ignition[735]: no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:17.911733 ignition[735]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:17.911810 ignition[735]: parsed url from cmdline: "" Nov 6 00:14:17.911814 ignition[735]: no config URL provided Nov 6 00:14:17.911819 ignition[735]: reading system config file "/usr/lib/ignition/user.ign" Nov 6 00:14:17.911827 ignition[735]: no config at "/usr/lib/ignition/user.ign" Nov 6 00:14:17.911849 ignition[735]: op(1): [started] loading QEMU firmware config module Nov 6 00:14:17.911854 ignition[735]: op(1): executing: "modprobe" "qemu_fw_cfg" Nov 6 00:14:18.022379 ignition[735]: op(1): [finished] loading QEMU firmware config module Nov 6 00:14:18.053431 systemd-networkd[832]: lo: Link UP Nov 6 00:14:18.053440 systemd-networkd[832]: lo: Gained carrier Nov 6 00:14:18.057452 systemd-networkd[832]: Enumeration completed Nov 6 00:14:18.060049 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 6 00:14:18.065897 systemd[1]: Reached target network.target - Network. Nov 6 00:14:18.071682 systemd-networkd[832]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 6 00:14:18.071691 systemd-networkd[832]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 6 00:14:18.078640 systemd-networkd[832]: eth0: Link UP Nov 6 00:14:18.078834 systemd-networkd[832]: eth0: Gained carrier Nov 6 00:14:18.078843 systemd-networkd[832]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 6 00:14:18.106232 systemd-networkd[832]: eth0: DHCPv4 address 10.0.0.38/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 6 00:14:18.125552 ignition[735]: parsing config with SHA512: a67a82e8c2318d271eec7d37647467765f185025b4df0ec789d5f918d037ca6b18f6146ada61ecea95a4ef46dcb2f9ba978f651a08b9ad2cbf26fdae6c389cc5 Nov 6 00:14:18.129749 unknown[735]: fetched base config from "system" Nov 6 00:14:18.129763 unknown[735]: fetched user config from "qemu" Nov 6 00:14:18.130119 ignition[735]: fetch-offline: fetch-offline passed Nov 6 00:14:18.130180 ignition[735]: Ignition finished successfully Nov 6 00:14:18.135792 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 6 00:14:18.136725 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 6 00:14:18.139666 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 6 00:14:18.206531 ignition[840]: Ignition 2.22.0 Nov 6 00:14:18.206544 ignition[840]: Stage: kargs Nov 6 00:14:18.206678 ignition[840]: no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:18.206690 ignition[840]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:18.207436 ignition[840]: kargs: kargs passed Nov 6 00:14:18.207477 ignition[840]: Ignition finished successfully Nov 6 00:14:18.217878 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 6 00:14:18.221113 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 6 00:14:18.286354 ignition[848]: Ignition 2.22.0 Nov 6 00:14:18.286366 ignition[848]: Stage: disks Nov 6 00:14:18.286521 ignition[848]: no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:18.286533 ignition[848]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:18.287391 ignition[848]: disks: disks passed Nov 6 00:14:18.287435 ignition[848]: Ignition finished successfully Nov 6 00:14:18.296947 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 6 00:14:18.297868 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 6 00:14:18.298199 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 6 00:14:18.298741 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 6 00:14:18.299296 systemd[1]: Reached target sysinit.target - System Initialization. Nov 6 00:14:18.311670 systemd[1]: Reached target basic.target - Basic System. Nov 6 00:14:18.316363 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 6 00:14:18.352404 systemd-fsck[858]: ROOT: clean, 15/553520 files, 52789/553472 blocks Nov 6 00:14:18.360597 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 6 00:14:18.363013 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 6 00:14:18.570222 kernel: EXT4-fs (vda9): mounted filesystem 25ee01aa-0270-4de7-b5da-d8936d968d16 r/w with ordered data mode. Quota mode: none. Nov 6 00:14:18.571225 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 6 00:14:18.572935 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 6 00:14:18.576794 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 6 00:14:18.581514 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 6 00:14:18.586592 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Nov 6 00:14:18.586667 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 6 00:14:18.586706 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 6 00:14:18.606450 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 6 00:14:18.609532 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 6 00:14:18.619289 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (866) Nov 6 00:14:18.619321 kernel: BTRFS info (device vda6): first mount of filesystem ca2bb832-66d5-4dca-a6d2-cbf7440d9381 Nov 6 00:14:18.619333 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 6 00:14:18.619344 kernel: BTRFS info (device vda6): turning on async discard Nov 6 00:14:18.621365 kernel: BTRFS info (device vda6): enabling free space tree Nov 6 00:14:18.623399 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 6 00:14:18.652014 initrd-setup-root[890]: cut: /sysroot/etc/passwd: No such file or directory Nov 6 00:14:18.657817 initrd-setup-root[897]: cut: /sysroot/etc/group: No such file or directory Nov 6 00:14:18.663357 initrd-setup-root[904]: cut: /sysroot/etc/shadow: No such file or directory Nov 6 00:14:18.669043 initrd-setup-root[911]: cut: /sysroot/etc/gshadow: No such file or directory Nov 6 00:14:18.774466 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 6 00:14:18.777775 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 6 00:14:18.781697 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 6 00:14:18.804413 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 6 00:14:18.807316 kernel: BTRFS info (device vda6): last unmount of filesystem ca2bb832-66d5-4dca-a6d2-cbf7440d9381 Nov 6 00:14:18.826355 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 6 00:14:18.992262 ignition[980]: INFO : Ignition 2.22.0 Nov 6 00:14:18.992262 ignition[980]: INFO : Stage: mount Nov 6 00:14:18.995373 ignition[980]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:18.995373 ignition[980]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:18.995373 ignition[980]: INFO : mount: mount passed Nov 6 00:14:18.995373 ignition[980]: INFO : Ignition finished successfully Nov 6 00:14:18.996311 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 6 00:14:19.000054 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 6 00:14:19.036370 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 6 00:14:19.063786 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (992) Nov 6 00:14:19.063832 kernel: BTRFS info (device vda6): first mount of filesystem ca2bb832-66d5-4dca-a6d2-cbf7440d9381 Nov 6 00:14:19.063844 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 6 00:14:19.069346 kernel: BTRFS info (device vda6): turning on async discard Nov 6 00:14:19.069407 kernel: BTRFS info (device vda6): enabling free space tree Nov 6 00:14:19.071155 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 6 00:14:19.123612 ignition[1009]: INFO : Ignition 2.22.0 Nov 6 00:14:19.123612 ignition[1009]: INFO : Stage: files Nov 6 00:14:19.126163 ignition[1009]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:19.126163 ignition[1009]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:19.130178 ignition[1009]: DEBUG : files: compiled without relabeling support, skipping Nov 6 00:14:19.132332 ignition[1009]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 6 00:14:19.132332 ignition[1009]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 6 00:14:19.138564 ignition[1009]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 6 00:14:19.140819 ignition[1009]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 6 00:14:19.143469 unknown[1009]: wrote ssh authorized keys file for user: core Nov 6 00:14:19.145123 ignition[1009]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 6 00:14:19.147557 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 6 00:14:19.147557 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Nov 6 00:14:19.187324 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Nov 6 00:14:19.262938 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Nov 6 00:14:19.262938 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Nov 6 00:14:19.269093 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Nov 6 00:14:19.599864 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Nov 6 00:14:19.643392 systemd-networkd[832]: eth0: Gained IPv6LL Nov 6 00:14:20.003433 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Nov 6 00:14:20.003433 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 6 00:14:20.009517 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 6 00:14:20.029778 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Nov 6 00:14:20.430519 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Nov 6 00:14:21.190818 ignition[1009]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Nov 6 00:14:21.190818 ignition[1009]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Nov 6 00:14:21.196635 ignition[1009]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 6 00:14:21.203227 ignition[1009]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 6 00:14:21.203227 ignition[1009]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Nov 6 00:14:21.208611 ignition[1009]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Nov 6 00:14:21.208611 ignition[1009]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 6 00:14:21.208611 ignition[1009]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 6 00:14:21.208611 ignition[1009]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Nov 6 00:14:21.208611 ignition[1009]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Nov 6 00:14:21.229863 ignition[1009]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Nov 6 00:14:21.237511 ignition[1009]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 6 00:14:21.240069 ignition[1009]: INFO : files: files passed Nov 6 00:14:21.240069 ignition[1009]: INFO : Ignition finished successfully Nov 6 00:14:21.245978 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 6 00:14:21.255364 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 6 00:14:21.259549 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 6 00:14:21.276949 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 6 00:14:21.277827 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 6 00:14:21.283177 initrd-setup-root-after-ignition[1038]: grep: /sysroot/oem/oem-release: No such file or directory Nov 6 00:14:21.288074 initrd-setup-root-after-ignition[1040]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 6 00:14:21.288074 initrd-setup-root-after-ignition[1040]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 6 00:14:21.293546 initrd-setup-root-after-ignition[1044]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 6 00:14:21.297988 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 6 00:14:21.298887 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 6 00:14:21.305304 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 6 00:14:21.358976 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 6 00:14:21.359111 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 6 00:14:21.363099 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 6 00:14:21.363790 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 6 00:14:21.371157 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 6 00:14:21.373846 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 6 00:14:21.411546 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 6 00:14:21.418324 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 6 00:14:21.561983 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 6 00:14:21.562913 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 6 00:14:21.569910 systemd[1]: Stopped target timers.target - Timer Units. Nov 6 00:14:21.570734 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 6 00:14:21.570889 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 6 00:14:21.576700 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 6 00:14:21.577617 systemd[1]: Stopped target basic.target - Basic System. Nov 6 00:14:21.578208 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 6 00:14:21.585933 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 6 00:14:21.586924 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 6 00:14:21.592950 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Nov 6 00:14:21.593909 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 6 00:14:21.599653 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 6 00:14:21.603284 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 6 00:14:21.604238 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 6 00:14:21.610096 systemd[1]: Stopped target swap.target - Swaps. Nov 6 00:14:21.612979 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 6 00:14:21.613099 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 6 00:14:21.617465 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 6 00:14:21.621020 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 6 00:14:21.621938 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 6 00:14:21.626537 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 6 00:14:21.630115 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 6 00:14:21.630269 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 6 00:14:21.635575 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 6 00:14:21.635699 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 6 00:14:21.636931 systemd[1]: Stopped target paths.target - Path Units. Nov 6 00:14:21.641633 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 6 00:14:21.642567 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 6 00:14:21.645677 systemd[1]: Stopped target slices.target - Slice Units. Nov 6 00:14:21.646274 systemd[1]: Stopped target sockets.target - Socket Units. Nov 6 00:14:21.653443 systemd[1]: iscsid.socket: Deactivated successfully. Nov 6 00:14:21.653573 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 6 00:14:21.654947 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 6 00:14:21.655082 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 6 00:14:21.655892 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 6 00:14:21.656040 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 6 00:14:21.665522 systemd[1]: ignition-files.service: Deactivated successfully. Nov 6 00:14:21.665682 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 6 00:14:21.667846 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 6 00:14:21.670839 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 6 00:14:21.670988 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 6 00:14:21.674835 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 6 00:14:21.677880 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 6 00:14:21.678039 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 6 00:14:21.678932 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 6 00:14:21.679034 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 6 00:14:21.695809 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 6 00:14:21.695950 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 6 00:14:21.733480 ignition[1064]: INFO : Ignition 2.22.0 Nov 6 00:14:21.733480 ignition[1064]: INFO : Stage: umount Nov 6 00:14:21.735971 ignition[1064]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 6 00:14:21.735971 ignition[1064]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 6 00:14:21.735971 ignition[1064]: INFO : umount: umount passed Nov 6 00:14:21.735971 ignition[1064]: INFO : Ignition finished successfully Nov 6 00:14:21.742270 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 6 00:14:21.742440 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 6 00:14:21.743473 systemd[1]: Stopped target network.target - Network. Nov 6 00:14:21.763911 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 6 00:14:21.763986 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 6 00:14:21.766602 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 6 00:14:21.766655 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 6 00:14:21.769574 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 6 00:14:21.769635 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 6 00:14:21.772548 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 6 00:14:21.772602 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 6 00:14:21.775897 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 6 00:14:21.778686 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 6 00:14:21.790311 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 6 00:14:21.790489 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 6 00:14:21.795714 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Nov 6 00:14:21.796043 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 6 00:14:21.796100 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 6 00:14:21.806132 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Nov 6 00:14:21.806573 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 6 00:14:21.806729 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 6 00:14:21.814761 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Nov 6 00:14:21.817671 systemd[1]: Stopped target network-pre.target - Preparation for Network. Nov 6 00:14:21.818863 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 6 00:14:21.818957 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 6 00:14:21.821163 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 6 00:14:21.829697 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 6 00:14:21.829781 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 6 00:14:21.830936 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 6 00:14:21.831001 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 6 00:14:21.838627 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 6 00:14:21.838700 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 6 00:14:21.839578 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 6 00:14:21.850725 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 6 00:14:21.855552 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 6 00:14:21.859742 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 6 00:14:21.859907 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 6 00:14:21.862320 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 6 00:14:21.862396 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 6 00:14:21.869960 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 6 00:14:21.881485 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 6 00:14:21.883098 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 6 00:14:21.883165 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 6 00:14:21.889568 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 6 00:14:21.889640 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 6 00:14:21.890785 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 6 00:14:21.890876 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 6 00:14:21.892229 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 6 00:14:21.892298 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 6 00:14:21.893515 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 6 00:14:21.893585 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 6 00:14:21.896076 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 6 00:14:21.908700 systemd[1]: systemd-network-generator.service: Deactivated successfully. Nov 6 00:14:21.908839 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Nov 6 00:14:21.915389 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 6 00:14:21.915485 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 6 00:14:21.921241 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 6 00:14:21.921316 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 6 00:14:21.927469 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 6 00:14:21.927591 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 6 00:14:21.935432 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 6 00:14:21.935573 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 6 00:14:21.939486 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 6 00:14:21.944660 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 6 00:14:21.973882 systemd[1]: Switching root. Nov 6 00:14:22.018802 systemd-journald[201]: Journal stopped Nov 6 00:14:23.430091 systemd-journald[201]: Received SIGTERM from PID 1 (systemd). Nov 6 00:14:23.430176 kernel: SELinux: policy capability network_peer_controls=1 Nov 6 00:14:23.430207 kernel: SELinux: policy capability open_perms=1 Nov 6 00:14:23.430222 kernel: SELinux: policy capability extended_socket_class=1 Nov 6 00:14:23.430239 kernel: SELinux: policy capability always_check_network=0 Nov 6 00:14:23.430250 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 6 00:14:23.430266 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 6 00:14:23.430284 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 6 00:14:23.430304 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 6 00:14:23.430411 kernel: SELinux: policy capability userspace_initial_context=0 Nov 6 00:14:23.430424 kernel: audit: type=1403 audit(1762388062.552:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 6 00:14:23.430439 systemd[1]: Successfully loaded SELinux policy in 66.675ms. Nov 6 00:14:23.430466 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 8.670ms. Nov 6 00:14:23.430479 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Nov 6 00:14:23.430492 systemd[1]: Detected virtualization kvm. Nov 6 00:14:23.430504 systemd[1]: Detected architecture x86-64. Nov 6 00:14:23.430526 systemd[1]: Detected first boot. Nov 6 00:14:23.430547 systemd[1]: Initializing machine ID from VM UUID. Nov 6 00:14:23.430562 zram_generator::config[1109]: No configuration found. Nov 6 00:14:23.430576 kernel: Guest personality initialized and is inactive Nov 6 00:14:23.430587 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Nov 6 00:14:23.430604 kernel: Initialized host personality Nov 6 00:14:23.430615 kernel: NET: Registered PF_VSOCK protocol family Nov 6 00:14:23.430627 systemd[1]: Populated /etc with preset unit settings. Nov 6 00:14:23.430640 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Nov 6 00:14:23.430652 systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 6 00:14:23.430673 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Nov 6 00:14:23.430685 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 6 00:14:23.430698 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 6 00:14:23.430710 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 6 00:14:23.430722 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 6 00:14:23.430735 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 6 00:14:23.430748 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 6 00:14:23.430761 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 6 00:14:23.430781 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 6 00:14:23.430793 systemd[1]: Created slice user.slice - User and Session Slice. Nov 6 00:14:23.430806 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 6 00:14:23.430819 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 6 00:14:23.430839 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 6 00:14:23.430854 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 6 00:14:23.430868 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 6 00:14:23.430881 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 6 00:14:23.430903 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Nov 6 00:14:23.430916 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 6 00:14:23.430928 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 6 00:14:23.430940 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Nov 6 00:14:23.430952 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Nov 6 00:14:23.430965 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Nov 6 00:14:23.430977 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 6 00:14:23.430990 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 6 00:14:23.431003 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 6 00:14:23.431023 systemd[1]: Reached target slices.target - Slice Units. Nov 6 00:14:23.431035 systemd[1]: Reached target swap.target - Swaps. Nov 6 00:14:23.431047 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 6 00:14:23.431059 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 6 00:14:23.431071 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Nov 6 00:14:23.431084 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 6 00:14:23.431096 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 6 00:14:23.431108 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 6 00:14:23.431120 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 6 00:14:23.431141 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 6 00:14:23.431154 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 6 00:14:23.431166 systemd[1]: Mounting media.mount - External Media Directory... Nov 6 00:14:23.431178 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:23.431222 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 6 00:14:23.431236 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 6 00:14:23.431248 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 6 00:14:23.431261 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 6 00:14:23.431273 systemd[1]: Reached target machines.target - Containers. Nov 6 00:14:23.431432 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 6 00:14:23.431445 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 6 00:14:23.431457 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 6 00:14:23.431470 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 6 00:14:23.431482 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 6 00:14:23.431494 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 6 00:14:23.431506 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 6 00:14:23.431518 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 6 00:14:23.431825 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 6 00:14:23.431840 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 6 00:14:23.431853 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 6 00:14:23.431866 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Nov 6 00:14:23.431880 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Nov 6 00:14:23.431892 systemd[1]: Stopped systemd-fsck-usr.service. Nov 6 00:14:23.431905 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 6 00:14:23.431917 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 6 00:14:23.431929 kernel: fuse: init (API version 7.41) Nov 6 00:14:23.431950 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 6 00:14:23.431962 kernel: loop: module loaded Nov 6 00:14:23.431973 kernel: ACPI: bus type drm_connector registered Nov 6 00:14:23.431986 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 6 00:14:23.431998 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 6 00:14:23.432011 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Nov 6 00:14:23.432048 systemd-journald[1187]: Collecting audit messages is disabled. Nov 6 00:14:23.432083 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 6 00:14:23.432096 systemd[1]: verity-setup.service: Deactivated successfully. Nov 6 00:14:23.432109 systemd-journald[1187]: Journal started Nov 6 00:14:23.432132 systemd-journald[1187]: Runtime Journal (/run/log/journal/a9a6cc210c29401d835a48f88cff5f32) is 6M, max 48.3M, 42.2M free. Nov 6 00:14:23.118986 systemd[1]: Queued start job for default target multi-user.target. Nov 6 00:14:23.132589 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Nov 6 00:14:23.133127 systemd[1]: systemd-journald.service: Deactivated successfully. Nov 6 00:14:23.436097 systemd[1]: Stopped verity-setup.service. Nov 6 00:14:23.440218 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:23.446539 systemd[1]: Started systemd-journald.service - Journal Service. Nov 6 00:14:23.447665 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 6 00:14:23.449843 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 6 00:14:23.452199 systemd[1]: Mounted media.mount - External Media Directory. Nov 6 00:14:23.453974 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 6 00:14:23.455947 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 6 00:14:23.458512 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 6 00:14:23.460677 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 6 00:14:23.463005 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 6 00:14:23.465392 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 6 00:14:23.465623 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 6 00:14:23.468492 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 6 00:14:23.469616 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 6 00:14:23.471946 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 6 00:14:23.472214 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 6 00:14:23.474708 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 6 00:14:23.474945 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 6 00:14:23.477581 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 6 00:14:23.477808 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 6 00:14:23.479834 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 6 00:14:23.480061 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 6 00:14:23.482156 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 6 00:14:23.484429 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 6 00:14:23.486883 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 6 00:14:23.489346 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Nov 6 00:14:23.507431 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 6 00:14:23.510703 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 6 00:14:23.513599 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 6 00:14:23.515573 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 6 00:14:23.515603 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 6 00:14:23.518578 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Nov 6 00:14:23.524320 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 6 00:14:23.526515 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 6 00:14:23.528341 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 6 00:14:23.532745 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 6 00:14:23.534802 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 6 00:14:23.544457 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 6 00:14:23.547311 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 6 00:14:23.548709 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 6 00:14:23.552572 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 6 00:14:23.561278 systemd-journald[1187]: Time spent on flushing to /var/log/journal/a9a6cc210c29401d835a48f88cff5f32 is 14.884ms for 982 entries. Nov 6 00:14:23.561278 systemd-journald[1187]: System Journal (/var/log/journal/a9a6cc210c29401d835a48f88cff5f32) is 8M, max 195.6M, 187.6M free. Nov 6 00:14:23.593722 systemd-journald[1187]: Received client request to flush runtime journal. Nov 6 00:14:23.593784 kernel: loop0: detected capacity change from 0 to 224512 Nov 6 00:14:23.557305 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 6 00:14:23.568351 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 6 00:14:23.573105 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 6 00:14:23.577476 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 6 00:14:23.585851 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 6 00:14:23.593332 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 6 00:14:23.599133 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Nov 6 00:14:23.602040 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 6 00:14:23.608495 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 6 00:14:23.620179 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 6 00:14:23.625050 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 6 00:14:23.626211 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 6 00:14:23.641490 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Nov 6 00:14:23.648218 kernel: loop1: detected capacity change from 0 to 110984 Nov 6 00:14:23.659580 systemd-tmpfiles[1244]: ACLs are not supported, ignoring. Nov 6 00:14:23.659599 systemd-tmpfiles[1244]: ACLs are not supported, ignoring. Nov 6 00:14:23.665689 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 6 00:14:23.675217 kernel: loop2: detected capacity change from 0 to 128016 Nov 6 00:14:23.707222 kernel: loop3: detected capacity change from 0 to 224512 Nov 6 00:14:23.718230 kernel: loop4: detected capacity change from 0 to 110984 Nov 6 00:14:23.729221 kernel: loop5: detected capacity change from 0 to 128016 Nov 6 00:14:23.741435 (sd-merge)[1250]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Nov 6 00:14:23.742027 (sd-merge)[1250]: Merged extensions into '/usr'. Nov 6 00:14:23.748911 systemd[1]: Reload requested from client PID 1228 ('systemd-sysext') (unit systemd-sysext.service)... Nov 6 00:14:23.748937 systemd[1]: Reloading... Nov 6 00:14:23.938403 zram_generator::config[1273]: No configuration found. Nov 6 00:14:24.134754 ldconfig[1223]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 6 00:14:24.216781 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 6 00:14:24.217235 systemd[1]: Reloading finished in 467 ms. Nov 6 00:14:24.258068 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 6 00:14:24.326090 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 6 00:14:24.336706 systemd[1]: Starting ensure-sysext.service... Nov 6 00:14:24.338991 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 6 00:14:24.349393 systemd[1]: Reload requested from client PID 1313 ('systemctl') (unit ensure-sysext.service)... Nov 6 00:14:24.349408 systemd[1]: Reloading... Nov 6 00:14:24.364430 systemd-tmpfiles[1314]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Nov 6 00:14:24.364472 systemd-tmpfiles[1314]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Nov 6 00:14:24.364837 systemd-tmpfiles[1314]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 6 00:14:24.365315 systemd-tmpfiles[1314]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 6 00:14:24.366741 systemd-tmpfiles[1314]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 6 00:14:24.367071 systemd-tmpfiles[1314]: ACLs are not supported, ignoring. Nov 6 00:14:24.367173 systemd-tmpfiles[1314]: ACLs are not supported, ignoring. Nov 6 00:14:24.394824 systemd-tmpfiles[1314]: Detected autofs mount point /boot during canonicalization of boot. Nov 6 00:14:24.394841 systemd-tmpfiles[1314]: Skipping /boot Nov 6 00:14:24.402221 zram_generator::config[1341]: No configuration found. Nov 6 00:14:24.411597 systemd-tmpfiles[1314]: Detected autofs mount point /boot during canonicalization of boot. Nov 6 00:14:24.411615 systemd-tmpfiles[1314]: Skipping /boot Nov 6 00:14:24.604233 systemd[1]: Reloading finished in 254 ms. Nov 6 00:14:24.629831 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 6 00:14:24.647407 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 6 00:14:24.657924 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 6 00:14:24.661155 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 6 00:14:24.686987 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 6 00:14:24.694331 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 6 00:14:24.698935 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 6 00:14:24.702999 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 6 00:14:24.710270 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.710602 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 6 00:14:24.713358 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 6 00:14:24.717687 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 6 00:14:24.720839 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 6 00:14:24.724462 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 6 00:14:24.724579 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 6 00:14:24.726972 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 6 00:14:24.729818 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.731118 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 6 00:14:24.731667 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 6 00:14:24.736758 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 6 00:14:24.739397 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 6 00:14:24.739700 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 6 00:14:24.742343 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 6 00:14:24.742652 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 6 00:14:24.750780 systemd-udevd[1387]: Using default interface naming scheme 'v255'. Nov 6 00:14:24.756209 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.756586 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 6 00:14:24.758201 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 6 00:14:24.761173 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 6 00:14:24.778730 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 6 00:14:24.780644 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 6 00:14:24.780819 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 6 00:14:24.783400 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 6 00:14:24.804694 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.806801 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 6 00:14:24.809548 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 6 00:14:24.809770 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 6 00:14:24.812051 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 6 00:14:24.812291 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 6 00:14:24.814719 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 6 00:14:24.814949 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 6 00:14:24.826506 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 6 00:14:24.830457 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.830701 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 6 00:14:24.833435 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 6 00:14:24.838317 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 6 00:14:24.870008 augenrules[1427]: No rules Nov 6 00:14:24.882488 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 6 00:14:24.886181 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 6 00:14:24.889724 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 6 00:14:24.889768 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Nov 6 00:14:24.889846 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 6 00:14:24.890366 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 6 00:14:24.893072 systemd[1]: Finished ensure-sysext.service. Nov 6 00:14:24.895227 systemd[1]: audit-rules.service: Deactivated successfully. Nov 6 00:14:24.900063 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 6 00:14:24.909215 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 6 00:14:24.912276 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 6 00:14:24.914757 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 6 00:14:24.915044 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 6 00:14:24.917607 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 6 00:14:24.917903 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 6 00:14:24.920927 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 6 00:14:24.921160 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 6 00:14:24.924696 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 6 00:14:24.924928 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 6 00:14:24.947363 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 6 00:14:24.949141 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 6 00:14:24.949231 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 6 00:14:24.953340 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 6 00:14:24.955176 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 6 00:14:24.961759 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Nov 6 00:14:25.067916 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 6 00:14:25.075126 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 6 00:14:25.075259 kernel: mousedev: PS/2 mouse device common for all mice Nov 6 00:14:25.088451 systemd-resolved[1383]: Positive Trust Anchors: Nov 6 00:14:25.088459 systemd-resolved[1383]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 6 00:14:25.088488 systemd-resolved[1383]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 6 00:14:25.093067 systemd-resolved[1383]: Defaulting to hostname 'linux'. Nov 6 00:14:25.095130 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 6 00:14:25.096005 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 6 00:14:25.102469 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 6 00:14:25.108276 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Nov 6 00:14:25.116463 kernel: ACPI: button: Power Button [PWRF] Nov 6 00:14:25.126160 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Nov 6 00:14:25.126498 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Nov 6 00:14:25.201805 systemd-networkd[1470]: lo: Link UP Nov 6 00:14:25.201817 systemd-networkd[1470]: lo: Gained carrier Nov 6 00:14:25.203528 systemd-networkd[1470]: Enumeration completed Nov 6 00:14:25.203646 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 6 00:14:25.205028 systemd-networkd[1470]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 6 00:14:25.205041 systemd-networkd[1470]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 6 00:14:25.205717 systemd-networkd[1470]: eth0: Link UP Nov 6 00:14:25.205914 systemd-networkd[1470]: eth0: Gained carrier Nov 6 00:14:25.205928 systemd-networkd[1470]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 6 00:14:25.206060 systemd[1]: Reached target network.target - Network. Nov 6 00:14:25.211324 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Nov 6 00:14:25.214457 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 6 00:14:25.226655 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 6 00:14:25.228719 systemd[1]: Reached target sysinit.target - System Initialization. Nov 6 00:14:25.230555 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 6 00:14:25.232709 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 6 00:14:25.236605 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Nov 6 00:14:25.238493 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 6 00:14:25.240612 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 6 00:14:25.240647 systemd[1]: Reached target paths.target - Path Units. Nov 6 00:14:25.242172 systemd[1]: Reached target time-set.target - System Time Set. Nov 6 00:14:25.244075 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 6 00:14:25.246082 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 6 00:14:25.248296 systemd[1]: Reached target timers.target - Timer Units. Nov 6 00:14:25.251695 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 6 00:14:25.257477 systemd-networkd[1470]: eth0: DHCPv4 address 10.0.0.38/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 6 00:14:25.259542 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 6 00:14:25.261376 systemd-timesyncd[1471]: Network configuration changed, trying to establish connection. Nov 6 00:14:26.373063 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Nov 6 00:14:26.374130 systemd-resolved[1383]: Clock change detected. Flushing caches. Nov 6 00:14:26.374239 systemd-timesyncd[1471]: Contacted time server 10.0.0.1:123 (10.0.0.1). Nov 6 00:14:26.374308 systemd-timesyncd[1471]: Initial clock synchronization to Thu 2025-11-06 00:14:26.371808 UTC. Nov 6 00:14:26.375674 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Nov 6 00:14:26.377819 systemd[1]: Reached target ssh-access.target - SSH Access Available. Nov 6 00:14:26.383852 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 6 00:14:26.385836 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Nov 6 00:14:26.390790 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Nov 6 00:14:26.393210 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 6 00:14:26.407092 systemd[1]: Reached target sockets.target - Socket Units. Nov 6 00:14:26.409936 systemd[1]: Reached target basic.target - Basic System. Nov 6 00:14:26.411556 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 6 00:14:26.411661 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 6 00:14:26.415970 systemd[1]: Starting containerd.service - containerd container runtime... Nov 6 00:14:26.419977 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 6 00:14:26.423144 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 6 00:14:26.427006 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 6 00:14:26.430086 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 6 00:14:26.432830 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 6 00:14:26.439393 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Nov 6 00:14:26.443741 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 6 00:14:26.449079 jq[1510]: false Nov 6 00:14:26.451104 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 6 00:14:26.458915 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Refreshing passwd entry cache Nov 6 00:14:26.459195 oslogin_cache_refresh[1513]: Refreshing passwd entry cache Nov 6 00:14:26.471030 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Failure getting users, quitting Nov 6 00:14:26.471030 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 6 00:14:26.471030 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Refreshing group entry cache Nov 6 00:14:26.470543 oslogin_cache_refresh[1513]: Failure getting users, quitting Nov 6 00:14:26.470571 oslogin_cache_refresh[1513]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Nov 6 00:14:26.470633 oslogin_cache_refresh[1513]: Refreshing group entry cache Nov 6 00:14:26.477440 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Failure getting groups, quitting Nov 6 00:14:26.477485 oslogin_cache_refresh[1513]: Failure getting groups, quitting Nov 6 00:14:26.477556 google_oslogin_nss_cache[1513]: oslogin_cache_refresh[1513]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 6 00:14:26.477587 oslogin_cache_refresh[1513]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Nov 6 00:14:26.494498 extend-filesystems[1511]: Found /dev/vda6 Nov 6 00:14:26.502471 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 6 00:14:26.506372 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 6 00:14:26.508318 extend-filesystems[1511]: Found /dev/vda9 Nov 6 00:14:26.512789 extend-filesystems[1511]: Checking size of /dev/vda9 Nov 6 00:14:26.522586 kernel: kvm_amd: TSC scaling supported Nov 6 00:14:26.522635 kernel: kvm_amd: Nested Virtualization enabled Nov 6 00:14:26.522649 kernel: kvm_amd: Nested Paging enabled Nov 6 00:14:26.522662 kernel: kvm_amd: LBR virtualization supported Nov 6 00:14:26.522680 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Nov 6 00:14:26.523255 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 6 00:14:26.524046 kernel: kvm_amd: Virtual GIF supported Nov 6 00:14:26.527132 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Nov 6 00:14:26.527842 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 6 00:14:26.530246 systemd[1]: Starting update-engine.service - Update Engine... Nov 6 00:14:26.531131 extend-filesystems[1511]: Resized partition /dev/vda9 Nov 6 00:14:26.535090 extend-filesystems[1536]: resize2fs 1.47.3 (8-Jul-2025) Nov 6 00:14:26.542882 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Nov 6 00:14:26.543629 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 6 00:14:26.550391 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 6 00:14:26.555375 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 6 00:14:26.555647 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 6 00:14:26.556509 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Nov 6 00:14:26.556777 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Nov 6 00:14:26.557363 jq[1537]: true Nov 6 00:14:26.560577 systemd[1]: motdgen.service: Deactivated successfully. Nov 6 00:14:26.560867 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 6 00:14:26.570442 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 6 00:14:26.571834 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 6 00:14:26.574777 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Nov 6 00:14:26.587840 update_engine[1534]: I20251106 00:14:26.586395 1534 main.cc:92] Flatcar Update Engine starting Nov 6 00:14:26.602152 (ntainerd)[1545]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 6 00:14:26.607504 jq[1544]: true Nov 6 00:14:26.609961 extend-filesystems[1536]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Nov 6 00:14:26.609961 extend-filesystems[1536]: old_desc_blocks = 1, new_desc_blocks = 1 Nov 6 00:14:26.609961 extend-filesystems[1536]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Nov 6 00:14:26.620232 extend-filesystems[1511]: Resized filesystem in /dev/vda9 Nov 6 00:14:26.621311 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 6 00:14:26.622532 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 6 00:14:26.639175 kernel: EDAC MC: Ver: 3.0.0 Nov 6 00:14:26.641085 tar[1542]: linux-amd64/LICENSE Nov 6 00:14:26.642870 tar[1542]: linux-amd64/helm Nov 6 00:14:26.644903 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 6 00:14:26.663171 systemd-logind[1530]: Watching system buttons on /dev/input/event2 (Power Button) Nov 6 00:14:26.663196 systemd-logind[1530]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Nov 6 00:14:26.664930 dbus-daemon[1508]: [system] SELinux support is enabled Nov 6 00:14:26.665118 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 6 00:14:26.665667 systemd-logind[1530]: New seat seat0. Nov 6 00:14:26.667656 update_engine[1534]: I20251106 00:14:26.667512 1534 update_check_scheduler.cc:74] Next update check in 10m17s Nov 6 00:14:26.677099 systemd[1]: Started systemd-logind.service - User Login Management. Nov 6 00:14:26.678981 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 6 00:14:26.679009 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 6 00:14:26.681260 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 6 00:14:26.681289 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 6 00:14:26.683881 systemd[1]: Started update-engine.service - Update Engine. Nov 6 00:14:26.688548 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 6 00:14:26.754378 bash[1576]: Updated "/home/core/.ssh/authorized_keys" Nov 6 00:14:26.757006 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 6 00:14:26.760316 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Nov 6 00:14:26.848464 locksmithd[1578]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 6 00:14:26.939253 sshd_keygen[1538]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 6 00:14:26.940875 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 6 00:14:26.955136 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 6 00:14:26.958930 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 6 00:14:26.978488 systemd[1]: issuegen.service: Deactivated successfully. Nov 6 00:14:26.978739 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 6 00:14:26.984566 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 6 00:14:27.003588 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 6 00:14:27.011168 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 6 00:14:27.015713 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Nov 6 00:14:27.017932 systemd[1]: Reached target getty.target - Login Prompts. Nov 6 00:14:27.124251 containerd[1545]: time="2025-11-06T00:14:27Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Nov 6 00:14:27.125773 containerd[1545]: time="2025-11-06T00:14:27.125721879Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138353898Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="17.703µs" Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138405135Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138426735Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138640115Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138656947Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Nov 6 00:14:27.138780 containerd[1545]: time="2025-11-06T00:14:27.138685190Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 6 00:14:27.138966 containerd[1545]: time="2025-11-06T00:14:27.138944857Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139016 containerd[1545]: time="2025-11-06T00:14:27.139003747Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139419 containerd[1545]: time="2025-11-06T00:14:27.139398287Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139478 containerd[1545]: time="2025-11-06T00:14:27.139464822Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139534 containerd[1545]: time="2025-11-06T00:14:27.139521348Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139590 containerd[1545]: time="2025-11-06T00:14:27.139577153Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Nov 6 00:14:27.139752 containerd[1545]: time="2025-11-06T00:14:27.139734978Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Nov 6 00:14:27.140106 containerd[1545]: time="2025-11-06T00:14:27.140081869Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 6 00:14:27.140213 containerd[1545]: time="2025-11-06T00:14:27.140194470Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Nov 6 00:14:27.140283 containerd[1545]: time="2025-11-06T00:14:27.140269912Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Nov 6 00:14:27.140519 containerd[1545]: time="2025-11-06T00:14:27.140474024Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Nov 6 00:14:27.142266 containerd[1545]: time="2025-11-06T00:14:27.142114761Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Nov 6 00:14:27.142359 containerd[1545]: time="2025-11-06T00:14:27.142340284Z" level=info msg="metadata content store policy set" policy=shared Nov 6 00:14:27.148911 containerd[1545]: time="2025-11-06T00:14:27.148880389Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Nov 6 00:14:27.149002 containerd[1545]: time="2025-11-06T00:14:27.148988743Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Nov 6 00:14:27.149116 containerd[1545]: time="2025-11-06T00:14:27.149102045Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Nov 6 00:14:27.149175 containerd[1545]: time="2025-11-06T00:14:27.149162228Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Nov 6 00:14:27.149231 containerd[1545]: time="2025-11-06T00:14:27.149219115Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Nov 6 00:14:27.149297 containerd[1545]: time="2025-11-06T00:14:27.149283425Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149337717Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149356612Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149369016Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149378814Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149389144Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149419320Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149543703Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149565254Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149579871Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149598596Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149609126Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149619235Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149630486Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149645154Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Nov 6 00:14:27.150518 containerd[1545]: time="2025-11-06T00:14:27.149658799Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Nov 6 00:14:27.150824 tar[1542]: linux-amd64/README.md Nov 6 00:14:27.150855 containerd[1545]: time="2025-11-06T00:14:27.149670161Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Nov 6 00:14:27.150855 containerd[1545]: time="2025-11-06T00:14:27.149679999Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Nov 6 00:14:27.150855 containerd[1545]: time="2025-11-06T00:14:27.149783683Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Nov 6 00:14:27.150855 containerd[1545]: time="2025-11-06T00:14:27.149798702Z" level=info msg="Start snapshots syncer" Nov 6 00:14:27.150855 containerd[1545]: time="2025-11-06T00:14:27.149842554Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Nov 6 00:14:27.150997 containerd[1545]: time="2025-11-06T00:14:27.150107220Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Nov 6 00:14:27.150997 containerd[1545]: time="2025-11-06T00:14:27.150157454Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150233657Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150343323Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150362809Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150373510Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150386694Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150401222Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150415238Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150425387Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150448901Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150461434Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Nov 6 00:14:27.151164 containerd[1545]: time="2025-11-06T00:14:27.150471203Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Nov 6 00:14:27.151572 containerd[1545]: time="2025-11-06T00:14:27.151551549Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 6 00:14:27.151629 containerd[1545]: time="2025-11-06T00:14:27.151615138Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Nov 6 00:14:27.151675 containerd[1545]: time="2025-11-06T00:14:27.151663288Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 6 00:14:27.151723 containerd[1545]: time="2025-11-06T00:14:27.151710477Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Nov 6 00:14:27.151801 containerd[1545]: time="2025-11-06T00:14:27.151786690Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Nov 6 00:14:27.151852 containerd[1545]: time="2025-11-06T00:14:27.151840801Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Nov 6 00:14:27.151921 containerd[1545]: time="2025-11-06T00:14:27.151907657Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Nov 6 00:14:27.151983 containerd[1545]: time="2025-11-06T00:14:27.151972358Z" level=info msg="runtime interface created" Nov 6 00:14:27.152026 containerd[1545]: time="2025-11-06T00:14:27.152016130Z" level=info msg="created NRI interface" Nov 6 00:14:27.152072 containerd[1545]: time="2025-11-06T00:14:27.152060654Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Nov 6 00:14:27.152123 containerd[1545]: time="2025-11-06T00:14:27.152112210Z" level=info msg="Connect containerd service" Nov 6 00:14:27.152232 containerd[1545]: time="2025-11-06T00:14:27.152210224Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 6 00:14:27.153131 containerd[1545]: time="2025-11-06T00:14:27.153109170Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 6 00:14:27.175728 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 6 00:14:27.316515 containerd[1545]: time="2025-11-06T00:14:27.316454441Z" level=info msg="Start subscribing containerd event" Nov 6 00:14:27.316644 containerd[1545]: time="2025-11-06T00:14:27.316523159Z" level=info msg="Start recovering state" Nov 6 00:14:27.316668 containerd[1545]: time="2025-11-06T00:14:27.316627104Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 6 00:14:27.316711 containerd[1545]: time="2025-11-06T00:14:27.316687718Z" level=info msg="Start event monitor" Nov 6 00:14:27.316733 containerd[1545]: time="2025-11-06T00:14:27.316714188Z" level=info msg="Start cni network conf syncer for default" Nov 6 00:14:27.316733 containerd[1545]: time="2025-11-06T00:14:27.316726902Z" level=info msg="Start streaming server" Nov 6 00:14:27.316787 containerd[1545]: time="2025-11-06T00:14:27.316739555Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 6 00:14:27.316787 containerd[1545]: time="2025-11-06T00:14:27.316750997Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Nov 6 00:14:27.316787 containerd[1545]: time="2025-11-06T00:14:27.316773920Z" level=info msg="runtime interface starting up..." Nov 6 00:14:27.316787 containerd[1545]: time="2025-11-06T00:14:27.316781514Z" level=info msg="starting plugins..." Nov 6 00:14:27.316872 containerd[1545]: time="2025-11-06T00:14:27.316805429Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Nov 6 00:14:27.317027 containerd[1545]: time="2025-11-06T00:14:27.316998240Z" level=info msg="containerd successfully booted in 0.193403s" Nov 6 00:14:27.317115 systemd[1]: Started containerd.service - containerd container runtime. Nov 6 00:14:27.728039 systemd-networkd[1470]: eth0: Gained IPv6LL Nov 6 00:14:27.731123 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 6 00:14:27.733837 systemd[1]: Reached target network-online.target - Network is Online. Nov 6 00:14:27.737235 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Nov 6 00:14:27.740965 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:14:27.760065 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 6 00:14:27.786670 systemd[1]: coreos-metadata.service: Deactivated successfully. Nov 6 00:14:27.787065 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Nov 6 00:14:27.789712 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Nov 6 00:14:27.792095 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 6 00:14:29.412487 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:29.415146 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 6 00:14:29.417023 systemd[1]: Startup finished in 3.219s (kernel) + 7.875s (initrd) + 5.821s (userspace) = 16.917s. Nov 6 00:14:29.467200 (kubelet)[1651]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 6 00:14:29.978365 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 6 00:14:29.979835 systemd[1]: Started sshd@0-10.0.0.38:22-10.0.0.1:40088.service - OpenSSH per-connection server daemon (10.0.0.1:40088). Nov 6 00:14:30.081414 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 40088 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:30.084401 sshd-session[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:30.094543 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 6 00:14:30.096334 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 6 00:14:30.107092 systemd-logind[1530]: New session 1 of user core. Nov 6 00:14:30.131843 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 6 00:14:30.136071 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 6 00:14:30.156655 (systemd)[1669]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 6 00:14:30.157244 kubelet[1651]: E1106 00:14:30.157161 1651 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 6 00:14:30.159684 systemd-logind[1530]: New session c1 of user core. Nov 6 00:14:30.161680 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 6 00:14:30.161942 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 6 00:14:30.162397 systemd[1]: kubelet.service: Consumed 2.136s CPU time, 265.5M memory peak. Nov 6 00:14:30.319358 systemd[1669]: Queued start job for default target default.target. Nov 6 00:14:30.338307 systemd[1669]: Created slice app.slice - User Application Slice. Nov 6 00:14:30.338339 systemd[1669]: Reached target paths.target - Paths. Nov 6 00:14:30.338394 systemd[1669]: Reached target timers.target - Timers. Nov 6 00:14:30.340058 systemd[1669]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 6 00:14:30.389424 systemd[1669]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 6 00:14:30.389571 systemd[1669]: Reached target sockets.target - Sockets. Nov 6 00:14:30.389622 systemd[1669]: Reached target basic.target - Basic System. Nov 6 00:14:30.389665 systemd[1669]: Reached target default.target - Main User Target. Nov 6 00:14:30.389701 systemd[1669]: Startup finished in 218ms. Nov 6 00:14:30.389952 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 6 00:14:30.391703 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 6 00:14:30.453782 systemd[1]: Started sshd@1-10.0.0.38:22-10.0.0.1:40102.service - OpenSSH per-connection server daemon (10.0.0.1:40102). Nov 6 00:14:30.529228 sshd[1681]: Accepted publickey for core from 10.0.0.1 port 40102 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:30.530549 sshd-session[1681]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:30.534662 systemd-logind[1530]: New session 2 of user core. Nov 6 00:14:30.548883 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 6 00:14:30.600452 sshd[1684]: Connection closed by 10.0.0.1 port 40102 Nov 6 00:14:30.600695 sshd-session[1681]: pam_unix(sshd:session): session closed for user core Nov 6 00:14:30.614210 systemd[1]: sshd@1-10.0.0.38:22-10.0.0.1:40102.service: Deactivated successfully. Nov 6 00:14:30.615911 systemd[1]: session-2.scope: Deactivated successfully. Nov 6 00:14:30.616626 systemd-logind[1530]: Session 2 logged out. Waiting for processes to exit. Nov 6 00:14:30.619010 systemd[1]: Started sshd@2-10.0.0.38:22-10.0.0.1:40118.service - OpenSSH per-connection server daemon (10.0.0.1:40118). Nov 6 00:14:30.619825 systemd-logind[1530]: Removed session 2. Nov 6 00:14:30.672738 sshd[1690]: Accepted publickey for core from 10.0.0.1 port 40118 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:30.673982 sshd-session[1690]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:30.678130 systemd-logind[1530]: New session 3 of user core. Nov 6 00:14:30.689907 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 6 00:14:30.738032 sshd[1693]: Connection closed by 10.0.0.1 port 40118 Nov 6 00:14:30.738447 sshd-session[1690]: pam_unix(sshd:session): session closed for user core Nov 6 00:14:30.748430 systemd[1]: sshd@2-10.0.0.38:22-10.0.0.1:40118.service: Deactivated successfully. Nov 6 00:14:30.750400 systemd[1]: session-3.scope: Deactivated successfully. Nov 6 00:14:30.751263 systemd-logind[1530]: Session 3 logged out. Waiting for processes to exit. Nov 6 00:14:30.754067 systemd[1]: Started sshd@3-10.0.0.38:22-10.0.0.1:40130.service - OpenSSH per-connection server daemon (10.0.0.1:40130). Nov 6 00:14:30.754708 systemd-logind[1530]: Removed session 3. Nov 6 00:14:30.818447 sshd[1699]: Accepted publickey for core from 10.0.0.1 port 40130 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:30.819665 sshd-session[1699]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:30.823992 systemd-logind[1530]: New session 4 of user core. Nov 6 00:14:30.835915 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 6 00:14:30.888192 sshd[1704]: Connection closed by 10.0.0.1 port 40130 Nov 6 00:14:30.888708 sshd-session[1699]: pam_unix(sshd:session): session closed for user core Nov 6 00:14:30.900641 systemd[1]: sshd@3-10.0.0.38:22-10.0.0.1:40130.service: Deactivated successfully. Nov 6 00:14:30.902470 systemd[1]: session-4.scope: Deactivated successfully. Nov 6 00:14:30.903266 systemd-logind[1530]: Session 4 logged out. Waiting for processes to exit. Nov 6 00:14:30.905981 systemd[1]: Started sshd@4-10.0.0.38:22-10.0.0.1:40132.service - OpenSSH per-connection server daemon (10.0.0.1:40132). Nov 6 00:14:30.906705 systemd-logind[1530]: Removed session 4. Nov 6 00:14:30.953898 sshd[1710]: Accepted publickey for core from 10.0.0.1 port 40132 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:30.955189 sshd-session[1710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:30.959558 systemd-logind[1530]: New session 5 of user core. Nov 6 00:14:30.968906 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 6 00:14:31.028095 sudo[1714]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 6 00:14:31.028414 sudo[1714]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 6 00:14:31.047197 sudo[1714]: pam_unix(sudo:session): session closed for user root Nov 6 00:14:31.048721 sshd[1713]: Connection closed by 10.0.0.1 port 40132 Nov 6 00:14:31.049105 sshd-session[1710]: pam_unix(sshd:session): session closed for user core Nov 6 00:14:31.067405 systemd[1]: sshd@4-10.0.0.38:22-10.0.0.1:40132.service: Deactivated successfully. Nov 6 00:14:31.069316 systemd[1]: session-5.scope: Deactivated successfully. Nov 6 00:14:31.070072 systemd-logind[1530]: Session 5 logged out. Waiting for processes to exit. Nov 6 00:14:31.072963 systemd[1]: Started sshd@5-10.0.0.38:22-10.0.0.1:40142.service - OpenSSH per-connection server daemon (10.0.0.1:40142). Nov 6 00:14:31.073745 systemd-logind[1530]: Removed session 5. Nov 6 00:14:31.124140 sshd[1720]: Accepted publickey for core from 10.0.0.1 port 40142 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:31.125412 sshd-session[1720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:31.129839 systemd-logind[1530]: New session 6 of user core. Nov 6 00:14:31.151989 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 6 00:14:31.208222 sudo[1725]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 6 00:14:31.208528 sudo[1725]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 6 00:14:31.217591 sudo[1725]: pam_unix(sudo:session): session closed for user root Nov 6 00:14:31.227227 sudo[1724]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Nov 6 00:14:31.227731 sudo[1724]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 6 00:14:31.239170 systemd[1]: Starting audit-rules.service - Load Audit Rules... Nov 6 00:14:31.278896 augenrules[1747]: No rules Nov 6 00:14:31.280414 systemd[1]: audit-rules.service: Deactivated successfully. Nov 6 00:14:31.280696 systemd[1]: Finished audit-rules.service - Load Audit Rules. Nov 6 00:14:31.281991 sudo[1724]: pam_unix(sudo:session): session closed for user root Nov 6 00:14:31.283501 sshd[1723]: Connection closed by 10.0.0.1 port 40142 Nov 6 00:14:31.283866 sshd-session[1720]: pam_unix(sshd:session): session closed for user core Nov 6 00:14:31.300349 systemd[1]: sshd@5-10.0.0.38:22-10.0.0.1:40142.service: Deactivated successfully. Nov 6 00:14:31.302156 systemd[1]: session-6.scope: Deactivated successfully. Nov 6 00:14:31.302990 systemd-logind[1530]: Session 6 logged out. Waiting for processes to exit. Nov 6 00:14:31.305517 systemd[1]: Started sshd@6-10.0.0.38:22-10.0.0.1:40152.service - OpenSSH per-connection server daemon (10.0.0.1:40152). Nov 6 00:14:31.306248 systemd-logind[1530]: Removed session 6. Nov 6 00:14:31.362308 sshd[1756]: Accepted publickey for core from 10.0.0.1 port 40152 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:14:31.363610 sshd-session[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:14:31.367804 systemd-logind[1530]: New session 7 of user core. Nov 6 00:14:31.377915 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 6 00:14:31.431831 sudo[1760]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 6 00:14:31.432155 sudo[1760]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 6 00:14:32.326929 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 6 00:14:32.349097 (dockerd)[1780]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 6 00:14:32.712203 dockerd[1780]: time="2025-11-06T00:14:32.712026584Z" level=info msg="Starting up" Nov 6 00:14:32.713104 dockerd[1780]: time="2025-11-06T00:14:32.713081453Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Nov 6 00:14:32.740313 dockerd[1780]: time="2025-11-06T00:14:32.740238632Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Nov 6 00:14:33.303381 dockerd[1780]: time="2025-11-06T00:14:33.303310092Z" level=info msg="Loading containers: start." Nov 6 00:14:33.314792 kernel: Initializing XFRM netlink socket Nov 6 00:14:33.585970 systemd-networkd[1470]: docker0: Link UP Nov 6 00:14:33.595408 dockerd[1780]: time="2025-11-06T00:14:33.595345813Z" level=info msg="Loading containers: done." Nov 6 00:14:33.614612 dockerd[1780]: time="2025-11-06T00:14:33.614554987Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 6 00:14:33.614818 dockerd[1780]: time="2025-11-06T00:14:33.614640638Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Nov 6 00:14:33.614818 dockerd[1780]: time="2025-11-06T00:14:33.614790539Z" level=info msg="Initializing buildkit" Nov 6 00:14:33.651992 dockerd[1780]: time="2025-11-06T00:14:33.651935429Z" level=info msg="Completed buildkit initialization" Nov 6 00:14:33.656576 dockerd[1780]: time="2025-11-06T00:14:33.656543521Z" level=info msg="Daemon has completed initialization" Nov 6 00:14:33.656679 dockerd[1780]: time="2025-11-06T00:14:33.656595047Z" level=info msg="API listen on /run/docker.sock" Nov 6 00:14:33.656839 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 6 00:14:34.723015 containerd[1545]: time="2025-11-06T00:14:34.722962317Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Nov 6 00:14:35.327914 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2363666291.mount: Deactivated successfully. Nov 6 00:14:36.278138 containerd[1545]: time="2025-11-06T00:14:36.278067915Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:36.278919 containerd[1545]: time="2025-11-06T00:14:36.278842788Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Nov 6 00:14:36.280870 containerd[1545]: time="2025-11-06T00:14:36.280796081Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:36.285171 containerd[1545]: time="2025-11-06T00:14:36.285109731Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:36.286098 containerd[1545]: time="2025-11-06T00:14:36.286043382Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.563032774s" Nov 6 00:14:36.286098 containerd[1545]: time="2025-11-06T00:14:36.286096661Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Nov 6 00:14:36.287010 containerd[1545]: time="2025-11-06T00:14:36.286792176Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Nov 6 00:14:38.184267 containerd[1545]: time="2025-11-06T00:14:38.184187670Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:38.185260 containerd[1545]: time="2025-11-06T00:14:38.185224474Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Nov 6 00:14:38.186890 containerd[1545]: time="2025-11-06T00:14:38.186857016Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:38.189678 containerd[1545]: time="2025-11-06T00:14:38.189605861Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:38.190505 containerd[1545]: time="2025-11-06T00:14:38.190466735Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.903641357s" Nov 6 00:14:38.190505 containerd[1545]: time="2025-11-06T00:14:38.190496701Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Nov 6 00:14:38.191359 containerd[1545]: time="2025-11-06T00:14:38.191324414Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Nov 6 00:14:40.412480 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 6 00:14:40.414942 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:14:40.759917 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:40.764272 (kubelet)[2074]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 6 00:14:40.969977 containerd[1545]: time="2025-11-06T00:14:40.969897350Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:40.985168 containerd[1545]: time="2025-11-06T00:14:40.985087276Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Nov 6 00:14:40.997278 containerd[1545]: time="2025-11-06T00:14:40.997203468Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:41.003947 containerd[1545]: time="2025-11-06T00:14:41.003848630Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:41.005817 containerd[1545]: time="2025-11-06T00:14:41.005721703Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 2.814354139s" Nov 6 00:14:41.005817 containerd[1545]: time="2025-11-06T00:14:41.005805530Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Nov 6 00:14:41.006493 containerd[1545]: time="2025-11-06T00:14:41.006439670Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Nov 6 00:14:41.019218 kubelet[2074]: E1106 00:14:41.019048 2074 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 6 00:14:41.026561 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 6 00:14:41.026835 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 6 00:14:41.027333 systemd[1]: kubelet.service: Consumed 428ms CPU time, 111.5M memory peak. Nov 6 00:14:42.861017 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount309330725.mount: Deactivated successfully. Nov 6 00:14:43.846273 containerd[1545]: time="2025-11-06T00:14:43.846206957Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:43.847293 containerd[1545]: time="2025-11-06T00:14:43.847244302Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Nov 6 00:14:43.848615 containerd[1545]: time="2025-11-06T00:14:43.848582692Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:43.850665 containerd[1545]: time="2025-11-06T00:14:43.850629611Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:43.851404 containerd[1545]: time="2025-11-06T00:14:43.851369498Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 2.844883272s" Nov 6 00:14:43.851446 containerd[1545]: time="2025-11-06T00:14:43.851407590Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Nov 6 00:14:43.852081 containerd[1545]: time="2025-11-06T00:14:43.852057348Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Nov 6 00:14:44.544844 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1788182535.mount: Deactivated successfully. Nov 6 00:14:45.724848 containerd[1545]: time="2025-11-06T00:14:45.724750927Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:45.725553 containerd[1545]: time="2025-11-06T00:14:45.725494342Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Nov 6 00:14:45.726692 containerd[1545]: time="2025-11-06T00:14:45.726645981Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:45.729304 containerd[1545]: time="2025-11-06T00:14:45.729220930Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:45.730049 containerd[1545]: time="2025-11-06T00:14:45.730014959Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.87793035s" Nov 6 00:14:45.730102 containerd[1545]: time="2025-11-06T00:14:45.730055876Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Nov 6 00:14:45.730676 containerd[1545]: time="2025-11-06T00:14:45.730642666Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Nov 6 00:14:46.444372 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2406343360.mount: Deactivated successfully. Nov 6 00:14:46.569751 containerd[1545]: time="2025-11-06T00:14:46.569660845Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 6 00:14:46.572889 containerd[1545]: time="2025-11-06T00:14:46.572832002Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Nov 6 00:14:46.576136 containerd[1545]: time="2025-11-06T00:14:46.576084732Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 6 00:14:46.580948 containerd[1545]: time="2025-11-06T00:14:46.580892488Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 6 00:14:46.581542 containerd[1545]: time="2025-11-06T00:14:46.581497323Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 850.826624ms" Nov 6 00:14:46.581542 containerd[1545]: time="2025-11-06T00:14:46.581527309Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Nov 6 00:14:46.582096 containerd[1545]: time="2025-11-06T00:14:46.582067282Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Nov 6 00:14:48.787756 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1704707843.mount: Deactivated successfully. Nov 6 00:14:51.277333 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Nov 6 00:14:51.279109 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:14:51.675099 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:51.691228 (kubelet)[2210]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 6 00:14:51.851585 kubelet[2210]: E1106 00:14:51.851503 2210 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 6 00:14:51.855674 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 6 00:14:51.855950 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 6 00:14:51.856378 systemd[1]: kubelet.service: Consumed 258ms CPU time, 110.4M memory peak. Nov 6 00:14:52.565442 containerd[1545]: time="2025-11-06T00:14:52.565347297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:52.585525 containerd[1545]: time="2025-11-06T00:14:52.585428938Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Nov 6 00:14:52.588526 containerd[1545]: time="2025-11-06T00:14:52.588468558Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:52.594112 containerd[1545]: time="2025-11-06T00:14:52.594028184Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:14:52.595756 containerd[1545]: time="2025-11-06T00:14:52.595668050Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 6.013568137s" Nov 6 00:14:52.595756 containerd[1545]: time="2025-11-06T00:14:52.595749933Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Nov 6 00:14:55.738674 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:55.738929 systemd[1]: kubelet.service: Consumed 258ms CPU time, 110.4M memory peak. Nov 6 00:14:55.741290 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:14:55.770850 systemd[1]: Reload requested from client PID 2246 ('systemctl') (unit session-7.scope)... Nov 6 00:14:55.770864 systemd[1]: Reloading... Nov 6 00:14:55.863056 zram_generator::config[2290]: No configuration found. Nov 6 00:14:56.179624 systemd[1]: Reloading finished in 408 ms. Nov 6 00:14:56.258019 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Nov 6 00:14:56.258148 systemd[1]: kubelet.service: Failed with result 'signal'. Nov 6 00:14:56.258515 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:56.258591 systemd[1]: kubelet.service: Consumed 176ms CPU time, 98.2M memory peak. Nov 6 00:14:56.260483 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:14:56.456197 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:14:56.468331 (kubelet)[2336]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 6 00:14:56.548937 kubelet[2336]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 6 00:14:56.548937 kubelet[2336]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 6 00:14:56.548937 kubelet[2336]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 6 00:14:56.549567 kubelet[2336]: I1106 00:14:56.549029 2336 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 6 00:14:57.047848 kubelet[2336]: I1106 00:14:57.047793 2336 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 6 00:14:57.047848 kubelet[2336]: I1106 00:14:57.047828 2336 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 6 00:14:57.048299 kubelet[2336]: I1106 00:14:57.048146 2336 server.go:954] "Client rotation is on, will bootstrap in background" Nov 6 00:14:57.074814 kubelet[2336]: E1106 00:14:57.074755 2336 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.38:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:57.075964 kubelet[2336]: I1106 00:14:57.075943 2336 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 6 00:14:57.083201 kubelet[2336]: I1106 00:14:57.083168 2336 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 6 00:14:57.090099 kubelet[2336]: I1106 00:14:57.090059 2336 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 6 00:14:57.090384 kubelet[2336]: I1106 00:14:57.090335 2336 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 6 00:14:57.090586 kubelet[2336]: I1106 00:14:57.090372 2336 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 6 00:14:57.090854 kubelet[2336]: I1106 00:14:57.090602 2336 topology_manager.go:138] "Creating topology manager with none policy" Nov 6 00:14:57.090854 kubelet[2336]: I1106 00:14:57.090613 2336 container_manager_linux.go:304] "Creating device plugin manager" Nov 6 00:14:57.090854 kubelet[2336]: I1106 00:14:57.090804 2336 state_mem.go:36] "Initialized new in-memory state store" Nov 6 00:14:57.093987 kubelet[2336]: I1106 00:14:57.093960 2336 kubelet.go:446] "Attempting to sync node with API server" Nov 6 00:14:57.094045 kubelet[2336]: I1106 00:14:57.093993 2336 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 6 00:14:57.094045 kubelet[2336]: I1106 00:14:57.094028 2336 kubelet.go:352] "Adding apiserver pod source" Nov 6 00:14:57.094045 kubelet[2336]: I1106 00:14:57.094045 2336 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 6 00:14:57.097228 kubelet[2336]: I1106 00:14:57.097195 2336 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Nov 6 00:14:57.097586 kubelet[2336]: I1106 00:14:57.097556 2336 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 6 00:14:57.099992 kubelet[2336]: W1106 00:14:57.099962 2336 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 6 00:14:57.100884 kubelet[2336]: W1106 00:14:57.100816 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:57.100884 kubelet[2336]: W1106 00:14:57.100829 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:57.100959 kubelet[2336]: E1106 00:14:57.100908 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:57.100959 kubelet[2336]: E1106 00:14:57.100916 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:57.102515 kubelet[2336]: I1106 00:14:57.102486 2336 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 6 00:14:57.102567 kubelet[2336]: I1106 00:14:57.102530 2336 server.go:1287] "Started kubelet" Nov 6 00:14:57.103154 kubelet[2336]: I1106 00:14:57.103099 2336 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 6 00:14:57.103579 kubelet[2336]: I1106 00:14:57.103553 2336 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 6 00:14:57.103673 kubelet[2336]: I1106 00:14:57.103645 2336 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 6 00:14:57.103970 kubelet[2336]: I1106 00:14:57.103950 2336 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 6 00:14:57.105385 kubelet[2336]: I1106 00:14:57.105361 2336 server.go:479] "Adding debug handlers to kubelet server" Nov 6 00:14:57.106663 kubelet[2336]: I1106 00:14:57.106635 2336 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 6 00:14:57.110925 kubelet[2336]: I1106 00:14:57.110832 2336 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 6 00:14:57.111853 kubelet[2336]: E1106 00:14:57.111029 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.111853 kubelet[2336]: I1106 00:14:57.111369 2336 reconciler.go:26] "Reconciler: start to sync state" Nov 6 00:14:57.111853 kubelet[2336]: E1106 00:14:57.111377 2336 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.38:6443: connect: connection refused" interval="200ms" Nov 6 00:14:57.111853 kubelet[2336]: I1106 00:14:57.111402 2336 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 6 00:14:57.111853 kubelet[2336]: W1106 00:14:57.111641 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.38:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:57.111853 kubelet[2336]: E1106 00:14:57.111685 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.38:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:57.113101 kubelet[2336]: E1106 00:14:57.111432 2336 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.38:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.38:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.187542a3042beaeb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-11-06 00:14:57.102506731 +0000 UTC m=+0.624540668,LastTimestamp:2025-11-06 00:14:57.102506731 +0000 UTC m=+0.624540668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 6 00:14:57.113398 kubelet[2336]: I1106 00:14:57.113356 2336 factory.go:221] Registration of the systemd container factory successfully Nov 6 00:14:57.113504 kubelet[2336]: I1106 00:14:57.113472 2336 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 6 00:14:57.114585 kubelet[2336]: E1106 00:14:57.114545 2336 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 6 00:14:57.115141 kubelet[2336]: I1106 00:14:57.115120 2336 factory.go:221] Registration of the containerd container factory successfully Nov 6 00:14:57.131046 kubelet[2336]: I1106 00:14:57.131019 2336 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 6 00:14:57.131046 kubelet[2336]: I1106 00:14:57.131034 2336 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 6 00:14:57.131247 kubelet[2336]: I1106 00:14:57.131071 2336 state_mem.go:36] "Initialized new in-memory state store" Nov 6 00:14:57.140730 kubelet[2336]: I1106 00:14:57.140666 2336 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 6 00:14:57.142233 kubelet[2336]: I1106 00:14:57.142205 2336 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 6 00:14:57.142288 kubelet[2336]: I1106 00:14:57.142254 2336 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 6 00:14:57.142314 kubelet[2336]: I1106 00:14:57.142291 2336 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 6 00:14:57.142314 kubelet[2336]: I1106 00:14:57.142306 2336 kubelet.go:2382] "Starting kubelet main sync loop" Nov 6 00:14:57.142443 kubelet[2336]: E1106 00:14:57.142368 2336 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 6 00:14:57.142922 kubelet[2336]: W1106 00:14:57.142871 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:57.142922 kubelet[2336]: E1106 00:14:57.142905 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:57.211834 kubelet[2336]: E1106 00:14:57.211785 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.242996 kubelet[2336]: E1106 00:14:57.242962 2336 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 6 00:14:57.312296 kubelet[2336]: E1106 00:14:57.312203 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.312802 kubelet[2336]: E1106 00:14:57.312704 2336 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.38:6443: connect: connection refused" interval="400ms" Nov 6 00:14:57.413019 kubelet[2336]: E1106 00:14:57.412986 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.443197 kubelet[2336]: E1106 00:14:57.443144 2336 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 6 00:14:57.513960 kubelet[2336]: E1106 00:14:57.513940 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.614480 kubelet[2336]: E1106 00:14:57.614324 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.714496 kubelet[2336]: E1106 00:14:57.714410 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.714496 kubelet[2336]: E1106 00:14:57.714419 2336 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.38:6443: connect: connection refused" interval="800ms" Nov 6 00:14:57.814938 kubelet[2336]: E1106 00:14:57.814855 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:57.844237 kubelet[2336]: E1106 00:14:57.844140 2336 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 6 00:14:57.915892 kubelet[2336]: E1106 00:14:57.915684 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.016514 kubelet[2336]: E1106 00:14:58.016445 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.089714 kubelet[2336]: W1106 00:14:58.089615 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:58.089714 kubelet[2336]: E1106 00:14:58.089703 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:58.117175 kubelet[2336]: E1106 00:14:58.117081 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.172652 kubelet[2336]: W1106 00:14:58.172471 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:58.172652 kubelet[2336]: E1106 00:14:58.172545 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:58.217547 kubelet[2336]: E1106 00:14:58.217487 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.318336 kubelet[2336]: E1106 00:14:58.318264 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.411073 kubelet[2336]: W1106 00:14:58.410983 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:58.411073 kubelet[2336]: E1106 00:14:58.411069 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:58.419182 kubelet[2336]: E1106 00:14:58.419117 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.515206 kubelet[2336]: E1106 00:14:58.515042 2336 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.38:6443: connect: connection refused" interval="1.6s" Nov 6 00:14:58.520172 kubelet[2336]: E1106 00:14:58.520123 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.599568 kubelet[2336]: W1106 00:14:58.599452 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.38:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:58.599568 kubelet[2336]: E1106 00:14:58.599554 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.38:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:58.620967 kubelet[2336]: E1106 00:14:58.620886 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.645299 kubelet[2336]: E1106 00:14:58.645179 2336 kubelet.go:2406] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 6 00:14:58.721999 kubelet[2336]: E1106 00:14:58.721896 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.822741 kubelet[2336]: E1106 00:14:58.822522 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:58.923400 kubelet[2336]: E1106 00:14:58.923273 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:59.024015 kubelet[2336]: E1106 00:14:59.023926 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:59.104414 kubelet[2336]: I1106 00:14:59.104292 2336 policy_none.go:49] "None policy: Start" Nov 6 00:14:59.104414 kubelet[2336]: I1106 00:14:59.104346 2336 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 6 00:14:59.104414 kubelet[2336]: I1106 00:14:59.104370 2336 state_mem.go:35] "Initializing new in-memory state store" Nov 6 00:14:59.124928 kubelet[2336]: E1106 00:14:59.124869 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:14:59.146203 kubelet[2336]: E1106 00:14:59.146160 2336 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.38:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:59.176461 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Nov 6 00:14:59.195827 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Nov 6 00:14:59.208049 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Nov 6 00:14:59.209845 kubelet[2336]: I1106 00:14:59.209815 2336 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 6 00:14:59.210158 kubelet[2336]: I1106 00:14:59.210112 2336 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 6 00:14:59.210218 kubelet[2336]: I1106 00:14:59.210150 2336 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 6 00:14:59.210739 kubelet[2336]: I1106 00:14:59.210475 2336 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 6 00:14:59.211650 kubelet[2336]: E1106 00:14:59.211610 2336 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 6 00:14:59.211728 kubelet[2336]: E1106 00:14:59.211698 2336 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Nov 6 00:14:59.312641 kubelet[2336]: I1106 00:14:59.312579 2336 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:14:59.313073 kubelet[2336]: E1106 00:14:59.313042 2336 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.38:6443/api/v1/nodes\": dial tcp 10.0.0.38:6443: connect: connection refused" node="localhost" Nov 6 00:14:59.514466 kubelet[2336]: I1106 00:14:59.514335 2336 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:14:59.514742 kubelet[2336]: E1106 00:14:59.514704 2336 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.38:6443/api/v1/nodes\": dial tcp 10.0.0.38:6443: connect: connection refused" node="localhost" Nov 6 00:14:59.803504 kubelet[2336]: W1106 00:14:59.803337 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:14:59.803504 kubelet[2336]: E1106 00:14:59.803390 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.38:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:14:59.917506 kubelet[2336]: I1106 00:14:59.917203 2336 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:14:59.917946 kubelet[2336]: E1106 00:14:59.917671 2336 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.38:6443/api/v1/nodes\": dial tcp 10.0.0.38:6443: connect: connection refused" node="localhost" Nov 6 00:15:00.116035 kubelet[2336]: E1106 00:15:00.115854 2336 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.38:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.38:6443: connect: connection refused" interval="3.2s" Nov 6 00:15:00.256235 systemd[1]: Created slice kubepods-burstable-pod9dc8d7e3c5c84d6fc5fb5590729f1612.slice - libcontainer container kubepods-burstable-pod9dc8d7e3c5c84d6fc5fb5590729f1612.slice. Nov 6 00:15:00.267997 kubelet[2336]: E1106 00:15:00.267937 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:00.271334 systemd[1]: Created slice kubepods-burstable-poda1d51be1ff02022474f2598f6e43038f.slice - libcontainer container kubepods-burstable-poda1d51be1ff02022474f2598f6e43038f.slice. Nov 6 00:15:00.287443 kubelet[2336]: E1106 00:15:00.287399 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:00.290580 systemd[1]: Created slice kubepods-burstable-pod4654b122dbb389158fe3c0766e603624.slice - libcontainer container kubepods-burstable-pod4654b122dbb389158fe3c0766e603624.slice. Nov 6 00:15:00.293111 kubelet[2336]: E1106 00:15:00.293071 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:00.305900 kubelet[2336]: E1106 00:15:00.305724 2336 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.38:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.38:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.187542a3042beaeb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-11-06 00:14:57.102506731 +0000 UTC m=+0.624540668,LastTimestamp:2025-11-06 00:14:57.102506731 +0000 UTC m=+0.624540668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 6 00:15:00.332322 kubelet[2336]: I1106 00:15:00.332237 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:00.332322 kubelet[2336]: I1106 00:15:00.332284 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:00.332322 kubelet[2336]: I1106 00:15:00.332306 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:00.332322 kubelet[2336]: I1106 00:15:00.332332 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:00.332684 kubelet[2336]: I1106 00:15:00.332355 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:00.332684 kubelet[2336]: I1106 00:15:00.332378 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:00.332684 kubelet[2336]: I1106 00:15:00.332415 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:00.332684 kubelet[2336]: I1106 00:15:00.332451 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:00.332684 kubelet[2336]: I1106 00:15:00.332507 2336 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a1d51be1ff02022474f2598f6e43038f-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a1d51be1ff02022474f2598f6e43038f\") " pod="kube-system/kube-scheduler-localhost" Nov 6 00:15:00.568782 kubelet[2336]: E1106 00:15:00.568560 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.569782 containerd[1545]: time="2025-11-06T00:15:00.569705624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:9dc8d7e3c5c84d6fc5fb5590729f1612,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:00.589791 kubelet[2336]: E1106 00:15:00.587907 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.589973 containerd[1545]: time="2025-11-06T00:15:00.588485440Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a1d51be1ff02022474f2598f6e43038f,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:00.594517 kubelet[2336]: E1106 00:15:00.594487 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.594972 containerd[1545]: time="2025-11-06T00:15:00.594937414Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:4654b122dbb389158fe3c0766e603624,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:00.782171 kubelet[2336]: I1106 00:15:00.782122 2336 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:15:00.782380 containerd[1545]: time="2025-11-06T00:15:00.782107835Z" level=info msg="connecting to shim 4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781" address="unix:///run/containerd/s/bc0a06bcce5cd0cc56d33a603d7ab4da05f76c0d5cb4a474ee7c56cee6951d52" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:00.782612 kubelet[2336]: E1106 00:15:00.782588 2336 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.38:6443/api/v1/nodes\": dial tcp 10.0.0.38:6443: connect: connection refused" node="localhost" Nov 6 00:15:00.787303 containerd[1545]: time="2025-11-06T00:15:00.787252501Z" level=info msg="connecting to shim be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f" address="unix:///run/containerd/s/3ab06a12450247ad30717b2d1671748bb993a7eaba642951aae4b17cafa66501" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:00.789796 containerd[1545]: time="2025-11-06T00:15:00.789745793Z" level=info msg="connecting to shim 0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b" address="unix:///run/containerd/s/3e945f0044e3ce928f772eaae9d8ac237030f3c5c715e335d5f2f9ce4845be77" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:00.792803 kubelet[2336]: W1106 00:15:00.792144 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:15:00.792803 kubelet[2336]: E1106 00:15:00.792237 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.38:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:15:00.803129 kubelet[2336]: W1106 00:15:00.803099 2336 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.38:6443: connect: connection refused Nov 6 00:15:00.803205 kubelet[2336]: E1106 00:15:00.803161 2336 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.38:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.38:6443: connect: connection refused" logger="UnhandledError" Nov 6 00:15:00.836969 systemd[1]: Started cri-containerd-4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781.scope - libcontainer container 4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781. Nov 6 00:15:00.844403 systemd[1]: Started cri-containerd-0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b.scope - libcontainer container 0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b. Nov 6 00:15:00.846214 systemd[1]: Started cri-containerd-be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f.scope - libcontainer container be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f. Nov 6 00:15:00.916559 containerd[1545]: time="2025-11-06T00:15:00.916509825Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:a1d51be1ff02022474f2598f6e43038f,Namespace:kube-system,Attempt:0,} returns sandbox id \"0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b\"" Nov 6 00:15:00.919277 kubelet[2336]: E1106 00:15:00.919244 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.921862 containerd[1545]: time="2025-11-06T00:15:00.921828775Z" level=info msg="CreateContainer within sandbox \"0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 6 00:15:00.923792 containerd[1545]: time="2025-11-06T00:15:00.923720863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:9dc8d7e3c5c84d6fc5fb5590729f1612,Namespace:kube-system,Attempt:0,} returns sandbox id \"4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781\"" Nov 6 00:15:00.924498 kubelet[2336]: E1106 00:15:00.924473 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.926741 containerd[1545]: time="2025-11-06T00:15:00.926705117Z" level=info msg="CreateContainer within sandbox \"4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 6 00:15:00.934574 containerd[1545]: time="2025-11-06T00:15:00.934527098Z" level=info msg="Container 6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:00.938839 containerd[1545]: time="2025-11-06T00:15:00.938750488Z" level=info msg="Container 37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:00.942791 containerd[1545]: time="2025-11-06T00:15:00.942730350Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:4654b122dbb389158fe3c0766e603624,Namespace:kube-system,Attempt:0,} returns sandbox id \"be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f\"" Nov 6 00:15:00.943779 kubelet[2336]: E1106 00:15:00.943722 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:00.946043 containerd[1545]: time="2025-11-06T00:15:00.945377157Z" level=info msg="CreateContainer within sandbox \"be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 6 00:15:00.951692 containerd[1545]: time="2025-11-06T00:15:00.951659024Z" level=info msg="CreateContainer within sandbox \"4c57de9f412a74590467ea15aed11e92a835e396da531b031f18fd91f13c1781\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373\"" Nov 6 00:15:00.953484 containerd[1545]: time="2025-11-06T00:15:00.953424109Z" level=info msg="StartContainer for \"37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373\"" Nov 6 00:15:00.954999 containerd[1545]: time="2025-11-06T00:15:00.954960366Z" level=info msg="CreateContainer within sandbox \"0d83c7d4b10fcb6a0b814bbfde793ae94cda73da478a93cb61ebccded61ba57b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5\"" Nov 6 00:15:00.955366 containerd[1545]: time="2025-11-06T00:15:00.955331267Z" level=info msg="connecting to shim 37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373" address="unix:///run/containerd/s/bc0a06bcce5cd0cc56d33a603d7ab4da05f76c0d5cb4a474ee7c56cee6951d52" protocol=ttrpc version=3 Nov 6 00:15:00.955745 containerd[1545]: time="2025-11-06T00:15:00.955717067Z" level=info msg="StartContainer for \"6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5\"" Nov 6 00:15:00.956963 containerd[1545]: time="2025-11-06T00:15:00.956932859Z" level=info msg="connecting to shim 6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5" address="unix:///run/containerd/s/3e945f0044e3ce928f772eaae9d8ac237030f3c5c715e335d5f2f9ce4845be77" protocol=ttrpc version=3 Nov 6 00:15:00.958347 containerd[1545]: time="2025-11-06T00:15:00.958320009Z" level=info msg="Container ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:00.969476 containerd[1545]: time="2025-11-06T00:15:00.969437521Z" level=info msg="CreateContainer within sandbox \"be4993280fba3ccfb8c8ad1901a18bc69ab851a6747ef23ea7dbf696e7d9d33f\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d\"" Nov 6 00:15:00.970296 containerd[1545]: time="2025-11-06T00:15:00.970275477Z" level=info msg="StartContainer for \"ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d\"" Nov 6 00:15:00.971635 containerd[1545]: time="2025-11-06T00:15:00.971615216Z" level=info msg="connecting to shim ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d" address="unix:///run/containerd/s/3ab06a12450247ad30717b2d1671748bb993a7eaba642951aae4b17cafa66501" protocol=ttrpc version=3 Nov 6 00:15:00.976048 systemd[1]: Started cri-containerd-37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373.scope - libcontainer container 37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373. Nov 6 00:15:00.980577 systemd[1]: Started cri-containerd-6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5.scope - libcontainer container 6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5. Nov 6 00:15:01.021935 systemd[1]: Started cri-containerd-ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d.scope - libcontainer container ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d. Nov 6 00:15:01.050522 containerd[1545]: time="2025-11-06T00:15:01.050024574Z" level=info msg="StartContainer for \"6b774b06cfc61fbced78dde4e28a308b7b7df7568b6be6bf714303f6a6a29ea5\" returns successfully" Nov 6 00:15:01.079028 containerd[1545]: time="2025-11-06T00:15:01.078972685Z" level=info msg="StartContainer for \"37da1bf954d550aa2a42747c66e42d7b47cc59190a2ef340ce9f704c3ac3e373\" returns successfully" Nov 6 00:15:01.107115 containerd[1545]: time="2025-11-06T00:15:01.106931941Z" level=info msg="StartContainer for \"ebb8a6e67e23435ff33836fe23a7d54571676891d31a82b818ab281fd1b7125d\" returns successfully" Nov 6 00:15:01.159662 kubelet[2336]: E1106 00:15:01.159617 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:01.159887 kubelet[2336]: E1106 00:15:01.159744 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:01.163423 kubelet[2336]: E1106 00:15:01.163382 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:01.163945 kubelet[2336]: E1106 00:15:01.163910 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:01.164014 kubelet[2336]: E1106 00:15:01.163991 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:01.164052 kubelet[2336]: E1106 00:15:01.164041 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:02.169395 kubelet[2336]: E1106 00:15:02.169349 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:02.169940 kubelet[2336]: E1106 00:15:02.169482 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:02.169984 kubelet[2336]: E1106 00:15:02.169961 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:02.173092 kubelet[2336]: E1106 00:15:02.170047 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:02.384657 kubelet[2336]: I1106 00:15:02.384621 2336 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:15:02.542029 kubelet[2336]: I1106 00:15:02.541606 2336 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 6 00:15:02.542029 kubelet[2336]: E1106 00:15:02.541668 2336 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Nov 6 00:15:02.563086 kubelet[2336]: E1106 00:15:02.563014 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:02.663431 kubelet[2336]: E1106 00:15:02.663366 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:02.764104 kubelet[2336]: E1106 00:15:02.764038 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:02.864842 kubelet[2336]: E1106 00:15:02.864670 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:02.965355 kubelet[2336]: E1106 00:15:02.965282 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.066015 kubelet[2336]: E1106 00:15:03.065925 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.166294 kubelet[2336]: E1106 00:15:03.166091 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.267017 kubelet[2336]: E1106 00:15:03.266949 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.367323 kubelet[2336]: E1106 00:15:03.367214 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.468064 kubelet[2336]: E1106 00:15:03.467914 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.568533 kubelet[2336]: E1106 00:15:03.568462 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.628869 kubelet[2336]: E1106 00:15:03.628826 2336 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Nov 6 00:15:03.629049 kubelet[2336]: E1106 00:15:03.629002 2336 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:03.669153 kubelet[2336]: E1106 00:15:03.669099 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.770225 kubelet[2336]: E1106 00:15:03.770047 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.870712 kubelet[2336]: E1106 00:15:03.870641 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:03.971306 kubelet[2336]: E1106 00:15:03.971236 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.072119 kubelet[2336]: E1106 00:15:04.071954 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.172574 kubelet[2336]: E1106 00:15:04.172524 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.273104 kubelet[2336]: E1106 00:15:04.273043 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.373862 kubelet[2336]: E1106 00:15:04.373692 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.459529 systemd[1]: Reload requested from client PID 2625 ('systemctl') (unit session-7.scope)... Nov 6 00:15:04.459548 systemd[1]: Reloading... Nov 6 00:15:04.474325 kubelet[2336]: E1106 00:15:04.474287 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.543814 zram_generator::config[2667]: No configuration found. Nov 6 00:15:04.575448 kubelet[2336]: E1106 00:15:04.575396 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.675664 kubelet[2336]: E1106 00:15:04.675540 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.776185 kubelet[2336]: E1106 00:15:04.776128 2336 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Nov 6 00:15:04.827960 systemd[1]: Reloading finished in 368 ms. Nov 6 00:15:04.864652 kubelet[2336]: I1106 00:15:04.864564 2336 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 6 00:15:04.864735 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:15:04.886531 systemd[1]: kubelet.service: Deactivated successfully. Nov 6 00:15:04.887016 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:15:04.887093 systemd[1]: kubelet.service: Consumed 1.129s CPU time, 130.9M memory peak. Nov 6 00:15:04.890013 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 6 00:15:05.124678 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 6 00:15:05.130280 (kubelet)[2713]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 6 00:15:05.175842 kubelet[2713]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 6 00:15:05.175842 kubelet[2713]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Nov 6 00:15:05.175842 kubelet[2713]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 6 00:15:05.175842 kubelet[2713]: I1106 00:15:05.175677 2713 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 6 00:15:05.182552 kubelet[2713]: I1106 00:15:05.182509 2713 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Nov 6 00:15:05.182552 kubelet[2713]: I1106 00:15:05.182539 2713 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 6 00:15:05.182798 kubelet[2713]: I1106 00:15:05.182779 2713 server.go:954] "Client rotation is on, will bootstrap in background" Nov 6 00:15:05.183931 kubelet[2713]: I1106 00:15:05.183899 2713 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 6 00:15:05.186229 kubelet[2713]: I1106 00:15:05.186139 2713 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 6 00:15:05.192840 kubelet[2713]: I1106 00:15:05.192803 2713 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 6 00:15:05.197723 kubelet[2713]: I1106 00:15:05.197671 2713 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 6 00:15:05.198149 kubelet[2713]: I1106 00:15:05.197977 2713 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 6 00:15:05.198850 kubelet[2713]: I1106 00:15:05.198032 2713 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 6 00:15:05.198850 kubelet[2713]: I1106 00:15:05.198256 2713 topology_manager.go:138] "Creating topology manager with none policy" Nov 6 00:15:05.198850 kubelet[2713]: I1106 00:15:05.198268 2713 container_manager_linux.go:304] "Creating device plugin manager" Nov 6 00:15:05.198850 kubelet[2713]: I1106 00:15:05.198331 2713 state_mem.go:36] "Initialized new in-memory state store" Nov 6 00:15:05.198850 kubelet[2713]: I1106 00:15:05.198524 2713 kubelet.go:446] "Attempting to sync node with API server" Nov 6 00:15:05.200003 kubelet[2713]: I1106 00:15:05.198552 2713 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 6 00:15:05.200003 kubelet[2713]: I1106 00:15:05.198582 2713 kubelet.go:352] "Adding apiserver pod source" Nov 6 00:15:05.200003 kubelet[2713]: I1106 00:15:05.198594 2713 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 6 00:15:05.201623 kubelet[2713]: I1106 00:15:05.201598 2713 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Nov 6 00:15:05.202151 kubelet[2713]: I1106 00:15:05.202128 2713 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 6 00:15:05.202828 kubelet[2713]: I1106 00:15:05.202699 2713 watchdog_linux.go:99] "Systemd watchdog is not enabled" Nov 6 00:15:05.202828 kubelet[2713]: I1106 00:15:05.202745 2713 server.go:1287] "Started kubelet" Nov 6 00:15:05.203077 kubelet[2713]: I1106 00:15:05.202964 2713 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Nov 6 00:15:05.203294 kubelet[2713]: I1106 00:15:05.203143 2713 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 6 00:15:05.204007 kubelet[2713]: I1106 00:15:05.203975 2713 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 6 00:15:05.204544 kubelet[2713]: I1106 00:15:05.204478 2713 server.go:479] "Adding debug handlers to kubelet server" Nov 6 00:15:05.211422 kubelet[2713]: E1106 00:15:05.211332 2713 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 6 00:15:05.211951 kubelet[2713]: I1106 00:15:05.211744 2713 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 6 00:15:05.213298 kubelet[2713]: I1106 00:15:05.213256 2713 volume_manager.go:297] "Starting Kubelet Volume Manager" Nov 6 00:15:05.213703 kubelet[2713]: I1106 00:15:05.213656 2713 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Nov 6 00:15:05.214298 kubelet[2713]: I1106 00:15:05.214279 2713 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Nov 6 00:15:05.214480 kubelet[2713]: I1106 00:15:05.214468 2713 reconciler.go:26] "Reconciler: start to sync state" Nov 6 00:15:05.217092 kubelet[2713]: I1106 00:15:05.217045 2713 factory.go:221] Registration of the systemd container factory successfully Nov 6 00:15:05.217279 kubelet[2713]: I1106 00:15:05.217245 2713 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 6 00:15:05.221529 kubelet[2713]: I1106 00:15:05.221480 2713 factory.go:221] Registration of the containerd container factory successfully Nov 6 00:15:05.230003 kubelet[2713]: I1106 00:15:05.229956 2713 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 6 00:15:05.232256 kubelet[2713]: I1106 00:15:05.231852 2713 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 6 00:15:05.232256 kubelet[2713]: I1106 00:15:05.231885 2713 status_manager.go:227] "Starting to sync pod status with apiserver" Nov 6 00:15:05.232256 kubelet[2713]: I1106 00:15:05.231915 2713 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Nov 6 00:15:05.232256 kubelet[2713]: I1106 00:15:05.231930 2713 kubelet.go:2382] "Starting kubelet main sync loop" Nov 6 00:15:05.232256 kubelet[2713]: E1106 00:15:05.231995 2713 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 6 00:15:05.291739 kubelet[2713]: I1106 00:15:05.291702 2713 cpu_manager.go:221] "Starting CPU manager" policy="none" Nov 6 00:15:05.291955 kubelet[2713]: I1106 00:15:05.291924 2713 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Nov 6 00:15:05.291955 kubelet[2713]: I1106 00:15:05.291950 2713 state_mem.go:36] "Initialized new in-memory state store" Nov 6 00:15:05.292135 kubelet[2713]: I1106 00:15:05.292118 2713 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 6 00:15:05.292163 kubelet[2713]: I1106 00:15:05.292129 2713 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 6 00:15:05.292163 kubelet[2713]: I1106 00:15:05.292160 2713 policy_none.go:49] "None policy: Start" Nov 6 00:15:05.292225 kubelet[2713]: I1106 00:15:05.292170 2713 memory_manager.go:186] "Starting memorymanager" policy="None" Nov 6 00:15:05.292225 kubelet[2713]: I1106 00:15:05.292181 2713 state_mem.go:35] "Initializing new in-memory state store" Nov 6 00:15:05.292351 kubelet[2713]: I1106 00:15:05.292294 2713 state_mem.go:75] "Updated machine memory state" Nov 6 00:15:05.297281 kubelet[2713]: I1106 00:15:05.297243 2713 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 6 00:15:05.297474 kubelet[2713]: I1106 00:15:05.297451 2713 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 6 00:15:05.297514 kubelet[2713]: I1106 00:15:05.297467 2713 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 6 00:15:05.297957 kubelet[2713]: I1106 00:15:05.297928 2713 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 6 00:15:05.300630 kubelet[2713]: E1106 00:15:05.299678 2713 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Nov 6 00:15:05.332686 kubelet[2713]: I1106 00:15:05.332620 2713 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:05.332874 kubelet[2713]: I1106 00:15:05.332835 2713 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Nov 6 00:15:05.332974 kubelet[2713]: I1106 00:15:05.332949 2713 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.405089 kubelet[2713]: I1106 00:15:05.404965 2713 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Nov 6 00:15:05.516414 kubelet[2713]: I1106 00:15:05.516365 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:05.516414 kubelet[2713]: I1106 00:15:05.516410 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:05.516631 kubelet[2713]: I1106 00:15:05.516441 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.516631 kubelet[2713]: I1106 00:15:05.516460 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.516631 kubelet[2713]: I1106 00:15:05.516475 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9dc8d7e3c5c84d6fc5fb5590729f1612-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"9dc8d7e3c5c84d6fc5fb5590729f1612\") " pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:05.516631 kubelet[2713]: I1106 00:15:05.516574 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.516730 kubelet[2713]: I1106 00:15:05.516637 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.516730 kubelet[2713]: I1106 00:15:05.516666 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a1d51be1ff02022474f2598f6e43038f-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"a1d51be1ff02022474f2598f6e43038f\") " pod="kube-system/kube-scheduler-localhost" Nov 6 00:15:05.516730 kubelet[2713]: I1106 00:15:05.516685 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4654b122dbb389158fe3c0766e603624-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"4654b122dbb389158fe3c0766e603624\") " pod="kube-system/kube-controller-manager-localhost" Nov 6 00:15:05.650159 kubelet[2713]: E1106 00:15:05.649870 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:05.650159 kubelet[2713]: E1106 00:15:05.649989 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:05.650944 kubelet[2713]: E1106 00:15:05.650234 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:05.662552 kubelet[2713]: I1106 00:15:05.662386 2713 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Nov 6 00:15:05.662552 kubelet[2713]: I1106 00:15:05.662492 2713 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Nov 6 00:15:06.044884 sudo[2749]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Nov 6 00:15:06.045262 sudo[2749]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Nov 6 00:15:06.199154 kubelet[2713]: I1106 00:15:06.199067 2713 apiserver.go:52] "Watching apiserver" Nov 6 00:15:06.214494 kubelet[2713]: I1106 00:15:06.214445 2713 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Nov 6 00:15:06.246790 kubelet[2713]: I1106 00:15:06.246608 2713 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:06.246790 kubelet[2713]: E1106 00:15:06.246726 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:06.248142 kubelet[2713]: E1106 00:15:06.248099 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:06.340112 kubelet[2713]: E1106 00:15:06.339863 2713 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Nov 6 00:15:06.340112 kubelet[2713]: E1106 00:15:06.340112 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:06.420021 sudo[2749]: pam_unix(sudo:session): session closed for user root Nov 6 00:15:06.493142 kubelet[2713]: I1106 00:15:06.493063 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.493043245 podStartE2EDuration="1.493043245s" podCreationTimestamp="2025-11-06 00:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:06.49297768 +0000 UTC m=+1.357887855" watchObservedRunningTime="2025-11-06 00:15:06.493043245 +0000 UTC m=+1.357953390" Nov 6 00:15:06.744425 kubelet[2713]: I1106 00:15:06.741554 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.741504403 podStartE2EDuration="1.741504403s" podCreationTimestamp="2025-11-06 00:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:06.540791467 +0000 UTC m=+1.405701612" watchObservedRunningTime="2025-11-06 00:15:06.741504403 +0000 UTC m=+1.606414548" Nov 6 00:15:07.247726 kubelet[2713]: E1106 00:15:07.247672 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:07.248241 kubelet[2713]: E1106 00:15:07.247804 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:08.202715 sudo[1760]: pam_unix(sudo:session): session closed for user root Nov 6 00:15:08.204280 sshd[1759]: Connection closed by 10.0.0.1 port 40152 Nov 6 00:15:08.204810 sshd-session[1756]: pam_unix(sshd:session): session closed for user core Nov 6 00:15:08.209441 systemd[1]: sshd@6-10.0.0.38:22-10.0.0.1:40152.service: Deactivated successfully. Nov 6 00:15:08.211919 systemd[1]: session-7.scope: Deactivated successfully. Nov 6 00:15:08.212151 systemd[1]: session-7.scope: Consumed 5.673s CPU time, 259.4M memory peak. Nov 6 00:15:08.213933 systemd-logind[1530]: Session 7 logged out. Waiting for processes to exit. Nov 6 00:15:08.215170 systemd-logind[1530]: Removed session 7. Nov 6 00:15:08.250819 kubelet[2713]: E1106 00:15:08.250783 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:09.853729 kubelet[2713]: E1106 00:15:09.853664 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:10.487693 kubelet[2713]: I1106 00:15:10.487644 2713 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 6 00:15:10.490943 containerd[1545]: time="2025-11-06T00:15:10.490750769Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 6 00:15:10.491553 kubelet[2713]: I1106 00:15:10.491071 2713 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 6 00:15:11.226179 kubelet[2713]: I1106 00:15:11.225438 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=6.225386754 podStartE2EDuration="6.225386754s" podCreationTimestamp="2025-11-06 00:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:06.741827528 +0000 UTC m=+1.606737673" watchObservedRunningTime="2025-11-06 00:15:11.225386754 +0000 UTC m=+6.090296899" Nov 6 00:15:11.257877 systemd[1]: Created slice kubepods-besteffort-poddc9bf445_947a_4f34_b266_5fb4171e597f.slice - libcontainer container kubepods-besteffort-poddc9bf445_947a_4f34_b266_5fb4171e597f.slice. Nov 6 00:15:11.260788 kubelet[2713]: I1106 00:15:11.260552 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dc9bf445-947a-4f34-b266-5fb4171e597f-lib-modules\") pod \"kube-proxy-njbs5\" (UID: \"dc9bf445-947a-4f34-b266-5fb4171e597f\") " pod="kube-system/kube-proxy-njbs5" Nov 6 00:15:11.260788 kubelet[2713]: I1106 00:15:11.260618 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/dc9bf445-947a-4f34-b266-5fb4171e597f-kube-proxy\") pod \"kube-proxy-njbs5\" (UID: \"dc9bf445-947a-4f34-b266-5fb4171e597f\") " pod="kube-system/kube-proxy-njbs5" Nov 6 00:15:11.260788 kubelet[2713]: I1106 00:15:11.260650 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/dc9bf445-947a-4f34-b266-5fb4171e597f-xtables-lock\") pod \"kube-proxy-njbs5\" (UID: \"dc9bf445-947a-4f34-b266-5fb4171e597f\") " pod="kube-system/kube-proxy-njbs5" Nov 6 00:15:11.260788 kubelet[2713]: I1106 00:15:11.260685 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7x7z\" (UniqueName: \"kubernetes.io/projected/dc9bf445-947a-4f34-b266-5fb4171e597f-kube-api-access-f7x7z\") pod \"kube-proxy-njbs5\" (UID: \"dc9bf445-947a-4f34-b266-5fb4171e597f\") " pod="kube-system/kube-proxy-njbs5" Nov 6 00:15:11.503515 update_engine[1534]: I20251106 00:15:11.503191 1534 update_attempter.cc:509] Updating boot flags... Nov 6 00:15:11.780940 kubelet[2713]: E1106 00:15:11.780486 2713 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Nov 6 00:15:11.780940 kubelet[2713]: E1106 00:15:11.780538 2713 projected.go:194] Error preparing data for projected volume kube-api-access-f7x7z for pod kube-system/kube-proxy-njbs5: configmap "kube-root-ca.crt" not found Nov 6 00:15:11.781375 kubelet[2713]: E1106 00:15:11.781329 2713 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/dc9bf445-947a-4f34-b266-5fb4171e597f-kube-api-access-f7x7z podName:dc9bf445-947a-4f34-b266-5fb4171e597f nodeName:}" failed. No retries permitted until 2025-11-06 00:15:12.280897496 +0000 UTC m=+7.145807641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-f7x7z" (UniqueName: "kubernetes.io/projected/dc9bf445-947a-4f34-b266-5fb4171e597f-kube-api-access-f7x7z") pod "kube-proxy-njbs5" (UID: "dc9bf445-947a-4f34-b266-5fb4171e597f") : configmap "kube-root-ca.crt" not found Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867278 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-bpf-maps\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867347 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-kernel\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867383 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-hubble-tls\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867410 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-etc-cni-netd\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867437 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-hostproc\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.868798 kubelet[2713]: I1106 00:15:11.867463 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cni-path\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867485 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-lib-modules\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867514 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-net\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867542 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-cgroup\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867573 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-xtables-lock\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867598 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-config-path\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869180 kubelet[2713]: I1106 00:15:11.867647 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-run\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869405 kubelet[2713]: I1106 00:15:11.867673 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9b0aebf9-f411-4839-bf6e-13273982517b-clustermesh-secrets\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.869405 kubelet[2713]: I1106 00:15:11.867694 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk25r\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-kube-api-access-wk25r\") pod \"cilium-mnkhl\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " pod="kube-system/cilium-mnkhl" Nov 6 00:15:11.894101 systemd[1]: Created slice kubepods-burstable-pod9b0aebf9_f411_4839_bf6e_13273982517b.slice - libcontainer container kubepods-burstable-pod9b0aebf9_f411_4839_bf6e_13273982517b.slice. Nov 6 00:15:11.968231 kubelet[2713]: I1106 00:15:11.968162 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzxp\" (UniqueName: \"kubernetes.io/projected/2633eddd-03ad-432e-944c-01548ec50159-kube-api-access-qxzxp\") pod \"cilium-operator-6c4d7847fc-h77r8\" (UID: \"2633eddd-03ad-432e-944c-01548ec50159\") " pod="kube-system/cilium-operator-6c4d7847fc-h77r8" Nov 6 00:15:11.968231 kubelet[2713]: I1106 00:15:11.968229 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2633eddd-03ad-432e-944c-01548ec50159-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-h77r8\" (UID: \"2633eddd-03ad-432e-944c-01548ec50159\") " pod="kube-system/cilium-operator-6c4d7847fc-h77r8" Nov 6 00:15:12.013859 systemd[1]: Created slice kubepods-besteffort-pod2633eddd_03ad_432e_944c_01548ec50159.slice - libcontainer container kubepods-besteffort-pod2633eddd_03ad_432e_944c_01548ec50159.slice. Nov 6 00:15:12.226268 kubelet[2713]: E1106 00:15:12.226193 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.227644 containerd[1545]: time="2025-11-06T00:15:12.227588283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mnkhl,Uid:9b0aebf9-f411-4839-bf6e-13273982517b,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:12.249386 containerd[1545]: time="2025-11-06T00:15:12.249320914Z" level=info msg="connecting to shim e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:12.289083 systemd[1]: Started cri-containerd-e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11.scope - libcontainer container e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11. Nov 6 00:15:12.317169 kubelet[2713]: E1106 00:15:12.317125 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.318490 containerd[1545]: time="2025-11-06T00:15:12.318110035Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-h77r8,Uid:2633eddd-03ad-432e-944c-01548ec50159,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:12.322841 containerd[1545]: time="2025-11-06T00:15:12.322807967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-mnkhl,Uid:9b0aebf9-f411-4839-bf6e-13273982517b,Namespace:kube-system,Attempt:0,} returns sandbox id \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\"" Nov 6 00:15:12.323460 kubelet[2713]: E1106 00:15:12.323436 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.324902 containerd[1545]: time="2025-11-06T00:15:12.324859585Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Nov 6 00:15:12.367788 containerd[1545]: time="2025-11-06T00:15:12.367719121Z" level=info msg="connecting to shim e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b" address="unix:///run/containerd/s/f27a997d6b10ad5144eae6b8b01cb9e3174bf0be6b018a61043031b9f913fc6e" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:12.402923 systemd[1]: Started cri-containerd-e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b.scope - libcontainer container e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b. Nov 6 00:15:12.450036 containerd[1545]: time="2025-11-06T00:15:12.449971994Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-h77r8,Uid:2633eddd-03ad-432e-944c-01548ec50159,Namespace:kube-system,Attempt:0,} returns sandbox id \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\"" Nov 6 00:15:12.453025 kubelet[2713]: E1106 00:15:12.452988 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.478828 kubelet[2713]: E1106 00:15:12.478628 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.480193 containerd[1545]: time="2025-11-06T00:15:12.479648099Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-njbs5,Uid:dc9bf445-947a-4f34-b266-5fb4171e597f,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:12.507803 containerd[1545]: time="2025-11-06T00:15:12.507715214Z" level=info msg="connecting to shim 068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d" address="unix:///run/containerd/s/a5c55bba911269f203c0b87a50c94513e9b90c6e37c2858adade7a81a728c37e" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:12.564943 systemd[1]: Started cri-containerd-068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d.scope - libcontainer container 068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d. Nov 6 00:15:12.615282 containerd[1545]: time="2025-11-06T00:15:12.615223916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-njbs5,Uid:dc9bf445-947a-4f34-b266-5fb4171e597f,Namespace:kube-system,Attempt:0,} returns sandbox id \"068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d\"" Nov 6 00:15:12.615855 kubelet[2713]: E1106 00:15:12.615829 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:12.618106 containerd[1545]: time="2025-11-06T00:15:12.618056974Z" level=info msg="CreateContainer within sandbox \"068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 6 00:15:12.637257 containerd[1545]: time="2025-11-06T00:15:12.637170924Z" level=info msg="Container 7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:12.647693 containerd[1545]: time="2025-11-06T00:15:12.647624954Z" level=info msg="CreateContainer within sandbox \"068e2377aa83484e93c174918407f61d27dd10027c362ed5fa643c0545ff6f3d\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8\"" Nov 6 00:15:12.648547 containerd[1545]: time="2025-11-06T00:15:12.648376148Z" level=info msg="StartContainer for \"7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8\"" Nov 6 00:15:12.650828 containerd[1545]: time="2025-11-06T00:15:12.650756948Z" level=info msg="connecting to shim 7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8" address="unix:///run/containerd/s/a5c55bba911269f203c0b87a50c94513e9b90c6e37c2858adade7a81a728c37e" protocol=ttrpc version=3 Nov 6 00:15:12.675913 systemd[1]: Started cri-containerd-7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8.scope - libcontainer container 7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8. Nov 6 00:15:12.729085 containerd[1545]: time="2025-11-06T00:15:12.728939790Z" level=info msg="StartContainer for \"7316933422660a64ccf9e2d850f4731d49ad870d91d442ce436cd7caf99043a8\" returns successfully" Nov 6 00:15:13.271467 kubelet[2713]: E1106 00:15:13.271421 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:13.281314 kubelet[2713]: I1106 00:15:13.281237 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-njbs5" podStartSLOduration=2.281202037 podStartE2EDuration="2.281202037s" podCreationTimestamp="2025-11-06 00:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:13.281173965 +0000 UTC m=+8.146084110" watchObservedRunningTime="2025-11-06 00:15:13.281202037 +0000 UTC m=+8.146112183" Nov 6 00:15:15.503165 kubelet[2713]: E1106 00:15:15.503110 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:16.278907 kubelet[2713]: E1106 00:15:16.278865 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:16.882622 kubelet[2713]: E1106 00:15:16.882572 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:17.280454 kubelet[2713]: E1106 00:15:17.280243 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:19.859191 kubelet[2713]: E1106 00:15:19.859121 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:22.872586 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3349756824.mount: Deactivated successfully. Nov 6 00:15:28.969171 containerd[1545]: time="2025-11-06T00:15:28.969081392Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:15:28.971234 containerd[1545]: time="2025-11-06T00:15:28.971198166Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Nov 6 00:15:28.973354 containerd[1545]: time="2025-11-06T00:15:28.973299001Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:15:28.974872 containerd[1545]: time="2025-11-06T00:15:28.974838058Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 16.649928659s" Nov 6 00:15:28.974934 containerd[1545]: time="2025-11-06T00:15:28.974874467Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Nov 6 00:15:28.982143 containerd[1545]: time="2025-11-06T00:15:28.982094126Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Nov 6 00:15:28.983254 containerd[1545]: time="2025-11-06T00:15:28.983213964Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Nov 6 00:15:29.062743 containerd[1545]: time="2025-11-06T00:15:29.062669748Z" level=info msg="Container c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:29.067617 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1579740069.mount: Deactivated successfully. Nov 6 00:15:29.187185 containerd[1545]: time="2025-11-06T00:15:29.187131953Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\"" Nov 6 00:15:29.187676 containerd[1545]: time="2025-11-06T00:15:29.187636633Z" level=info msg="StartContainer for \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\"" Nov 6 00:15:29.188573 containerd[1545]: time="2025-11-06T00:15:29.188544572Z" level=info msg="connecting to shim c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" protocol=ttrpc version=3 Nov 6 00:15:29.217110 systemd[1]: Started cri-containerd-c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54.scope - libcontainer container c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54. Nov 6 00:15:29.305299 systemd[1]: cri-containerd-c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54.scope: Deactivated successfully. Nov 6 00:15:29.305832 systemd[1]: cri-containerd-c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54.scope: Consumed 27ms CPU time, 6.9M memory peak, 3.2M written to disk. Nov 6 00:15:29.308266 containerd[1545]: time="2025-11-06T00:15:29.308231108Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" id:\"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" pid:3154 exited_at:{seconds:1762388129 nanos:307307059}" Nov 6 00:15:29.362258 containerd[1545]: time="2025-11-06T00:15:29.362196466Z" level=info msg="received exit event container_id:\"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" id:\"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" pid:3154 exited_at:{seconds:1762388129 nanos:307307059}" Nov 6 00:15:29.372109 containerd[1545]: time="2025-11-06T00:15:29.372063093Z" level=info msg="StartContainer for \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" returns successfully" Nov 6 00:15:29.386117 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54-rootfs.mount: Deactivated successfully. Nov 6 00:15:30.372840 kubelet[2713]: E1106 00:15:30.372798 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:30.374785 containerd[1545]: time="2025-11-06T00:15:30.374726939Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Nov 6 00:15:30.386287 containerd[1545]: time="2025-11-06T00:15:30.386226956Z" level=info msg="Container fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:30.398331 containerd[1545]: time="2025-11-06T00:15:30.398275515Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\"" Nov 6 00:15:30.399891 containerd[1545]: time="2025-11-06T00:15:30.399854756Z" level=info msg="StartContainer for \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\"" Nov 6 00:15:30.400923 containerd[1545]: time="2025-11-06T00:15:30.400890997Z" level=info msg="connecting to shim fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" protocol=ttrpc version=3 Nov 6 00:15:30.424929 systemd[1]: Started cri-containerd-fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa.scope - libcontainer container fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa. Nov 6 00:15:30.457270 containerd[1545]: time="2025-11-06T00:15:30.457223955Z" level=info msg="StartContainer for \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" returns successfully" Nov 6 00:15:30.473117 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 6 00:15:30.473656 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 6 00:15:30.474948 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Nov 6 00:15:30.478124 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 6 00:15:30.481109 containerd[1545]: time="2025-11-06T00:15:30.480980383Z" level=info msg="received exit event container_id:\"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" id:\"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" pid:3202 exited_at:{seconds:1762388130 nanos:480544793}" Nov 6 00:15:30.481109 containerd[1545]: time="2025-11-06T00:15:30.481073108Z" level=info msg="TaskExit event in podsandbox handler container_id:\"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" id:\"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" pid:3202 exited_at:{seconds:1762388130 nanos:480544793}" Nov 6 00:15:30.482510 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Nov 6 00:15:30.483909 systemd[1]: cri-containerd-fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa.scope: Deactivated successfully. Nov 6 00:15:30.506049 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa-rootfs.mount: Deactivated successfully. Nov 6 00:15:30.518445 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 6 00:15:31.378646 kubelet[2713]: E1106 00:15:31.378589 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:31.382536 containerd[1545]: time="2025-11-06T00:15:31.382462761Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Nov 6 00:15:31.407788 containerd[1545]: time="2025-11-06T00:15:31.407124993Z" level=info msg="Container 7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:31.423192 containerd[1545]: time="2025-11-06T00:15:31.423121164Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\"" Nov 6 00:15:31.424071 containerd[1545]: time="2025-11-06T00:15:31.424013302Z" level=info msg="StartContainer for \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\"" Nov 6 00:15:31.425426 containerd[1545]: time="2025-11-06T00:15:31.425377087Z" level=info msg="connecting to shim 7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" protocol=ttrpc version=3 Nov 6 00:15:31.455918 systemd[1]: Started cri-containerd-7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b.scope - libcontainer container 7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b. Nov 6 00:15:31.485301 containerd[1545]: time="2025-11-06T00:15:31.485238493Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:15:31.488707 containerd[1545]: time="2025-11-06T00:15:31.486474238Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Nov 6 00:15:31.490035 containerd[1545]: time="2025-11-06T00:15:31.489993006Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 6 00:15:31.492414 containerd[1545]: time="2025-11-06T00:15:31.492369718Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.510241427s" Nov 6 00:15:31.492414 containerd[1545]: time="2025-11-06T00:15:31.492406246Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Nov 6 00:15:31.495756 containerd[1545]: time="2025-11-06T00:15:31.495712315Z" level=info msg="CreateContainer within sandbox \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Nov 6 00:15:31.508363 containerd[1545]: time="2025-11-06T00:15:31.507579879Z" level=info msg="Container f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:31.517787 systemd[1]: cri-containerd-7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b.scope: Deactivated successfully. Nov 6 00:15:31.519173 containerd[1545]: time="2025-11-06T00:15:31.519125368Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" id:\"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" pid:3262 exited_at:{seconds:1762388131 nanos:518500893}" Nov 6 00:15:31.742130 containerd[1545]: time="2025-11-06T00:15:31.742021809Z" level=info msg="received exit event container_id:\"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" id:\"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" pid:3262 exited_at:{seconds:1762388131 nanos:518500893}" Nov 6 00:15:31.751753 containerd[1545]: time="2025-11-06T00:15:31.751691618Z" level=info msg="StartContainer for \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" returns successfully" Nov 6 00:15:32.037717 containerd[1545]: time="2025-11-06T00:15:32.037573559Z" level=info msg="CreateContainer within sandbox \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\"" Nov 6 00:15:32.038315 containerd[1545]: time="2025-11-06T00:15:32.038288884Z" level=info msg="StartContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\"" Nov 6 00:15:32.039242 containerd[1545]: time="2025-11-06T00:15:32.039175392Z" level=info msg="connecting to shim f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82" address="unix:///run/containerd/s/f27a997d6b10ad5144eae6b8b01cb9e3174bf0be6b018a61043031b9f913fc6e" protocol=ttrpc version=3 Nov 6 00:15:32.061943 systemd[1]: Started cri-containerd-f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82.scope - libcontainer container f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82. Nov 6 00:15:32.096044 containerd[1545]: time="2025-11-06T00:15:32.095991968Z" level=info msg="StartContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" returns successfully" Nov 6 00:15:32.383134 kubelet[2713]: E1106 00:15:32.382945 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:32.390087 kubelet[2713]: E1106 00:15:32.390040 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:32.392480 containerd[1545]: time="2025-11-06T00:15:32.392430463Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Nov 6 00:15:32.394810 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b-rootfs.mount: Deactivated successfully. Nov 6 00:15:32.416343 containerd[1545]: time="2025-11-06T00:15:32.416208363Z" level=info msg="Container 0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:32.419420 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1584969455.mount: Deactivated successfully. Nov 6 00:15:32.439245 containerd[1545]: time="2025-11-06T00:15:32.439161322Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\"" Nov 6 00:15:32.441216 containerd[1545]: time="2025-11-06T00:15:32.441149661Z" level=info msg="StartContainer for \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\"" Nov 6 00:15:32.442886 containerd[1545]: time="2025-11-06T00:15:32.442779727Z" level=info msg="connecting to shim 0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" protocol=ttrpc version=3 Nov 6 00:15:32.449508 kubelet[2713]: I1106 00:15:32.449291 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-h77r8" podStartSLOduration=2.409186583 podStartE2EDuration="21.449264152s" podCreationTimestamp="2025-11-06 00:15:11 +0000 UTC" firstStartedPulling="2025-11-06 00:15:12.453530648 +0000 UTC m=+7.318440793" lastFinishedPulling="2025-11-06 00:15:31.493608217 +0000 UTC m=+26.358518362" observedRunningTime="2025-11-06 00:15:32.409909517 +0000 UTC m=+27.274819662" watchObservedRunningTime="2025-11-06 00:15:32.449264152 +0000 UTC m=+27.314174297" Nov 6 00:15:32.482011 systemd[1]: Started cri-containerd-0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc.scope - libcontainer container 0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc. Nov 6 00:15:32.530605 containerd[1545]: time="2025-11-06T00:15:32.530555791Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" id:\"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" pid:3339 exited_at:{seconds:1762388132 nanos:530305249}" Nov 6 00:15:32.530946 systemd[1]: cri-containerd-0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc.scope: Deactivated successfully. Nov 6 00:15:32.563430 containerd[1545]: time="2025-11-06T00:15:32.562904280Z" level=info msg="received exit event container_id:\"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" id:\"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" pid:3339 exited_at:{seconds:1762388132 nanos:530305249}" Nov 6 00:15:32.565141 containerd[1545]: time="2025-11-06T00:15:32.565105039Z" level=info msg="StartContainer for \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" returns successfully" Nov 6 00:15:33.390739 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc-rootfs.mount: Deactivated successfully. Nov 6 00:15:33.395446 kubelet[2713]: E1106 00:15:33.395398 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:33.395909 kubelet[2713]: E1106 00:15:33.395567 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:33.397703 containerd[1545]: time="2025-11-06T00:15:33.397653595Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Nov 6 00:15:33.414542 containerd[1545]: time="2025-11-06T00:15:33.414477280Z" level=info msg="Container 376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:33.417612 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1495402397.mount: Deactivated successfully. Nov 6 00:15:33.423486 containerd[1545]: time="2025-11-06T00:15:33.423439612Z" level=info msg="CreateContainer within sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\"" Nov 6 00:15:33.424076 containerd[1545]: time="2025-11-06T00:15:33.424025684Z" level=info msg="StartContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\"" Nov 6 00:15:33.425237 containerd[1545]: time="2025-11-06T00:15:33.425207717Z" level=info msg="connecting to shim 376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b" address="unix:///run/containerd/s/0706f078a51af6c0d259a11c1e5e4be44eca4c6643177e1d83241ef0caeb1187" protocol=ttrpc version=3 Nov 6 00:15:33.453942 systemd[1]: Started cri-containerd-376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b.scope - libcontainer container 376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b. Nov 6 00:15:33.498067 containerd[1545]: time="2025-11-06T00:15:33.498011027Z" level=info msg="StartContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" returns successfully" Nov 6 00:15:33.595975 containerd[1545]: time="2025-11-06T00:15:33.595868268Z" level=info msg="TaskExit event in podsandbox handler container_id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" id:\"75e9e235ad4f0f1bd0c6638d78f5f2fffc3e88474071b7a239e06ff08a7fa18b\" pid:3407 exited_at:{seconds:1762388133 nanos:595321169}" Nov 6 00:15:33.674530 kubelet[2713]: I1106 00:15:33.674123 2713 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Nov 6 00:15:33.714296 systemd[1]: Created slice kubepods-burstable-pod2eba401b_d685_4ac0_8b15_ce2ed245f2c2.slice - libcontainer container kubepods-burstable-pod2eba401b_d685_4ac0_8b15_ce2ed245f2c2.slice. Nov 6 00:15:33.719639 systemd[1]: Created slice kubepods-burstable-pod08bfb451_533a_4bbf_a609_9ec22752b223.slice - libcontainer container kubepods-burstable-pod08bfb451_533a_4bbf_a609_9ec22752b223.slice. Nov 6 00:15:33.816780 kubelet[2713]: I1106 00:15:33.816719 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eba401b-d685-4ac0-8b15-ce2ed245f2c2-config-volume\") pod \"coredns-668d6bf9bc-9xwsq\" (UID: \"2eba401b-d685-4ac0-8b15-ce2ed245f2c2\") " pod="kube-system/coredns-668d6bf9bc-9xwsq" Nov 6 00:15:33.816780 kubelet[2713]: I1106 00:15:33.816773 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7jfv\" (UniqueName: \"kubernetes.io/projected/2eba401b-d685-4ac0-8b15-ce2ed245f2c2-kube-api-access-v7jfv\") pod \"coredns-668d6bf9bc-9xwsq\" (UID: \"2eba401b-d685-4ac0-8b15-ce2ed245f2c2\") " pod="kube-system/coredns-668d6bf9bc-9xwsq" Nov 6 00:15:33.816990 kubelet[2713]: I1106 00:15:33.816799 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08bfb451-533a-4bbf-a609-9ec22752b223-config-volume\") pod \"coredns-668d6bf9bc-bgspn\" (UID: \"08bfb451-533a-4bbf-a609-9ec22752b223\") " pod="kube-system/coredns-668d6bf9bc-bgspn" Nov 6 00:15:33.816990 kubelet[2713]: I1106 00:15:33.816814 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdrrw\" (UniqueName: \"kubernetes.io/projected/08bfb451-533a-4bbf-a609-9ec22752b223-kube-api-access-kdrrw\") pod \"coredns-668d6bf9bc-bgspn\" (UID: \"08bfb451-533a-4bbf-a609-9ec22752b223\") " pod="kube-system/coredns-668d6bf9bc-bgspn" Nov 6 00:15:34.018068 kubelet[2713]: E1106 00:15:34.018022 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:34.019075 containerd[1545]: time="2025-11-06T00:15:34.019036631Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-9xwsq,Uid:2eba401b-d685-4ac0-8b15-ce2ed245f2c2,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:34.023321 kubelet[2713]: E1106 00:15:34.023290 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:34.024013 containerd[1545]: time="2025-11-06T00:15:34.023973753Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-bgspn,Uid:08bfb451-533a-4bbf-a609-9ec22752b223,Namespace:kube-system,Attempt:0,}" Nov 6 00:15:34.435170 kubelet[2713]: E1106 00:15:34.435135 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:34.482144 kubelet[2713]: I1106 00:15:34.482061 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-mnkhl" podStartSLOduration=6.82428044 podStartE2EDuration="23.482036762s" podCreationTimestamp="2025-11-06 00:15:11 +0000 UTC" firstStartedPulling="2025-11-06 00:15:12.324130443 +0000 UTC m=+7.189040588" lastFinishedPulling="2025-11-06 00:15:28.981886765 +0000 UTC m=+23.846796910" observedRunningTime="2025-11-06 00:15:34.480992629 +0000 UTC m=+29.345902774" watchObservedRunningTime="2025-11-06 00:15:34.482036762 +0000 UTC m=+29.346946907" Nov 6 00:15:35.437757 kubelet[2713]: E1106 00:15:35.437714 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:35.789741 systemd-networkd[1470]: cilium_host: Link UP Nov 6 00:15:35.790787 systemd-networkd[1470]: cilium_net: Link UP Nov 6 00:15:35.791321 systemd-networkd[1470]: cilium_net: Gained carrier Nov 6 00:15:35.791622 systemd-networkd[1470]: cilium_host: Gained carrier Nov 6 00:15:35.851941 systemd-networkd[1470]: cilium_host: Gained IPv6LL Nov 6 00:15:35.926873 systemd-networkd[1470]: cilium_vxlan: Link UP Nov 6 00:15:35.926897 systemd-networkd[1470]: cilium_vxlan: Gained carrier Nov 6 00:15:35.968015 systemd-networkd[1470]: cilium_net: Gained IPv6LL Nov 6 00:15:36.179787 kernel: NET: Registered PF_ALG protocol family Nov 6 00:15:36.439626 kubelet[2713]: E1106 00:15:36.439482 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:36.950356 systemd-networkd[1470]: lxc_health: Link UP Nov 6 00:15:36.951121 systemd-networkd[1470]: lxc_health: Gained carrier Nov 6 00:15:37.120983 systemd-networkd[1470]: lxcec7a85cd355c: Link UP Nov 6 00:15:37.127814 kernel: eth0: renamed from tmpa852d Nov 6 00:15:37.134820 systemd-networkd[1470]: lxcc66ceb779013: Link UP Nov 6 00:15:37.144555 kernel: eth0: renamed from tmpa234c Nov 6 00:15:37.145580 systemd-networkd[1470]: lxcec7a85cd355c: Gained carrier Nov 6 00:15:37.147849 systemd-networkd[1470]: lxcc66ceb779013: Gained carrier Nov 6 00:15:37.359997 systemd-networkd[1470]: cilium_vxlan: Gained IPv6LL Nov 6 00:15:38.228657 kubelet[2713]: E1106 00:15:38.228591 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:38.256040 systemd-networkd[1470]: lxcc66ceb779013: Gained IPv6LL Nov 6 00:15:38.446872 kubelet[2713]: E1106 00:15:38.446799 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:38.832102 systemd-networkd[1470]: lxcec7a85cd355c: Gained IPv6LL Nov 6 00:15:38.959940 systemd-networkd[1470]: lxc_health: Gained IPv6LL Nov 6 00:15:39.448791 kubelet[2713]: E1106 00:15:39.448692 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:39.513917 systemd[1]: Started sshd@7-10.0.0.38:22-10.0.0.1:34638.service - OpenSSH per-connection server daemon (10.0.0.1:34638). Nov 6 00:15:39.589961 sshd[3878]: Accepted publickey for core from 10.0.0.1 port 34638 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:15:39.591459 sshd-session[3878]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:15:39.602475 systemd-logind[1530]: New session 8 of user core. Nov 6 00:15:39.608047 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 6 00:15:39.790870 sshd[3882]: Connection closed by 10.0.0.1 port 34638 Nov 6 00:15:39.793027 sshd-session[3878]: pam_unix(sshd:session): session closed for user core Nov 6 00:15:39.797594 systemd-logind[1530]: Session 8 logged out. Waiting for processes to exit. Nov 6 00:15:39.798204 systemd[1]: sshd@7-10.0.0.38:22-10.0.0.1:34638.service: Deactivated successfully. Nov 6 00:15:39.800519 systemd[1]: session-8.scope: Deactivated successfully. Nov 6 00:15:39.803133 systemd-logind[1530]: Removed session 8. Nov 6 00:15:41.064856 containerd[1545]: time="2025-11-06T00:15:41.062477092Z" level=info msg="connecting to shim a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2" address="unix:///run/containerd/s/490c9885b8df55fad95cc83132b2ffcc3f8d2d74322e86a980ed9a0b947f7b55" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:41.069626 containerd[1545]: time="2025-11-06T00:15:41.069419803Z" level=info msg="connecting to shim a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844" address="unix:///run/containerd/s/60e3cb09a1fcd9549c4e8649a4264149ec452f2693fafa9eee3f4ba33b5c6832" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:15:41.108262 systemd[1]: Started cri-containerd-a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2.scope - libcontainer container a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2. Nov 6 00:15:41.113805 systemd[1]: Started cri-containerd-a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844.scope - libcontainer container a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844. Nov 6 00:15:41.126378 systemd-resolved[1383]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 6 00:15:41.131403 systemd-resolved[1383]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 6 00:15:41.159887 containerd[1545]: time="2025-11-06T00:15:41.159625021Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-9xwsq,Uid:2eba401b-d685-4ac0-8b15-ce2ed245f2c2,Namespace:kube-system,Attempt:0,} returns sandbox id \"a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2\"" Nov 6 00:15:41.161811 kubelet[2713]: E1106 00:15:41.161459 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:41.168247 containerd[1545]: time="2025-11-06T00:15:41.168202734Z" level=info msg="CreateContainer within sandbox \"a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 6 00:15:41.173724 containerd[1545]: time="2025-11-06T00:15:41.173681295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-bgspn,Uid:08bfb451-533a-4bbf-a609-9ec22752b223,Namespace:kube-system,Attempt:0,} returns sandbox id \"a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844\"" Nov 6 00:15:41.174379 kubelet[2713]: E1106 00:15:41.174341 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:41.175785 containerd[1545]: time="2025-11-06T00:15:41.175726456Z" level=info msg="CreateContainer within sandbox \"a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 6 00:15:41.275616 containerd[1545]: time="2025-11-06T00:15:41.275552716Z" level=info msg="Container 9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:41.277828 containerd[1545]: time="2025-11-06T00:15:41.277741819Z" level=info msg="Container a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:15:41.284338 containerd[1545]: time="2025-11-06T00:15:41.284284559Z" level=info msg="CreateContainer within sandbox \"a852daa06e1ebf13177d5b03b17b34b1229c3e9649aa764be9833ba0f9f6b6a2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53\"" Nov 6 00:15:41.285510 containerd[1545]: time="2025-11-06T00:15:41.284829642Z" level=info msg="StartContainer for \"9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53\"" Nov 6 00:15:41.285842 containerd[1545]: time="2025-11-06T00:15:41.285818830Z" level=info msg="connecting to shim 9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53" address="unix:///run/containerd/s/490c9885b8df55fad95cc83132b2ffcc3f8d2d74322e86a980ed9a0b947f7b55" protocol=ttrpc version=3 Nov 6 00:15:41.287853 containerd[1545]: time="2025-11-06T00:15:41.287814268Z" level=info msg="CreateContainer within sandbox \"a234c1bc527a04dd345a8a39cf6f11e72dc6ed9f9d29ca578f7c1f18daf7c844\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2\"" Nov 6 00:15:41.288309 containerd[1545]: time="2025-11-06T00:15:41.288287828Z" level=info msg="StartContainer for \"a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2\"" Nov 6 00:15:41.289245 containerd[1545]: time="2025-11-06T00:15:41.289211593Z" level=info msg="connecting to shim a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2" address="unix:///run/containerd/s/60e3cb09a1fcd9549c4e8649a4264149ec452f2693fafa9eee3f4ba33b5c6832" protocol=ttrpc version=3 Nov 6 00:15:41.313934 systemd[1]: Started cri-containerd-9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53.scope - libcontainer container 9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53. Nov 6 00:15:41.318205 systemd[1]: Started cri-containerd-a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2.scope - libcontainer container a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2. Nov 6 00:15:41.358579 containerd[1545]: time="2025-11-06T00:15:41.358521323Z" level=info msg="StartContainer for \"9592f12a50496e28fe181f03c34f015505c9baeabf07ebe320fb4d2f89d42d53\" returns successfully" Nov 6 00:15:41.374840 containerd[1545]: time="2025-11-06T00:15:41.374752040Z" level=info msg="StartContainer for \"a57f33ef79bf6f1986e6c925056e2ab7b57b0c3d224d0455088288237b73bbf2\" returns successfully" Nov 6 00:15:41.454316 kubelet[2713]: E1106 00:15:41.454277 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:41.462555 kubelet[2713]: E1106 00:15:41.462476 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:41.490250 kubelet[2713]: I1106 00:15:41.490174 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-bgspn" podStartSLOduration=30.490147124 podStartE2EDuration="30.490147124s" podCreationTimestamp="2025-11-06 00:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:41.489640112 +0000 UTC m=+36.354550267" watchObservedRunningTime="2025-11-06 00:15:41.490147124 +0000 UTC m=+36.355057269" Nov 6 00:15:41.490460 kubelet[2713]: I1106 00:15:41.490291 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-9xwsq" podStartSLOduration=30.490285003 podStartE2EDuration="30.490285003s" podCreationTimestamp="2025-11-06 00:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:15:41.475126099 +0000 UTC m=+36.340036284" watchObservedRunningTime="2025-11-06 00:15:41.490285003 +0000 UTC m=+36.355195148" Nov 6 00:15:42.463239 kubelet[2713]: E1106 00:15:42.463101 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:42.463239 kubelet[2713]: E1106 00:15:42.463101 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:43.465667 kubelet[2713]: E1106 00:15:43.465606 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:43.466211 kubelet[2713]: E1106 00:15:43.465700 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:15:44.806656 systemd[1]: Started sshd@8-10.0.0.38:22-10.0.0.1:34648.service - OpenSSH per-connection server daemon (10.0.0.1:34648). Nov 6 00:15:44.861587 sshd[4075]: Accepted publickey for core from 10.0.0.1 port 34648 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:15:44.863662 sshd-session[4075]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:15:44.869399 systemd-logind[1530]: New session 9 of user core. Nov 6 00:15:44.878927 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 6 00:15:45.029734 sshd[4078]: Connection closed by 10.0.0.1 port 34648 Nov 6 00:15:45.030135 sshd-session[4075]: pam_unix(sshd:session): session closed for user core Nov 6 00:15:45.033746 systemd[1]: sshd@8-10.0.0.38:22-10.0.0.1:34648.service: Deactivated successfully. Nov 6 00:15:45.036162 systemd[1]: session-9.scope: Deactivated successfully. Nov 6 00:15:45.037795 systemd-logind[1530]: Session 9 logged out. Waiting for processes to exit. Nov 6 00:15:45.039379 systemd-logind[1530]: Removed session 9. Nov 6 00:15:50.047333 systemd[1]: Started sshd@9-10.0.0.38:22-10.0.0.1:40920.service - OpenSSH per-connection server daemon (10.0.0.1:40920). Nov 6 00:15:50.125808 sshd[4094]: Accepted publickey for core from 10.0.0.1 port 40920 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:15:50.128299 sshd-session[4094]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:15:50.134817 systemd-logind[1530]: New session 10 of user core. Nov 6 00:15:50.141938 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 6 00:15:50.280946 sshd[4097]: Connection closed by 10.0.0.1 port 40920 Nov 6 00:15:50.281328 sshd-session[4094]: pam_unix(sshd:session): session closed for user core Nov 6 00:15:50.285532 systemd[1]: sshd@9-10.0.0.38:22-10.0.0.1:40920.service: Deactivated successfully. Nov 6 00:15:50.287897 systemd[1]: session-10.scope: Deactivated successfully. Nov 6 00:15:50.289621 systemd-logind[1530]: Session 10 logged out. Waiting for processes to exit. Nov 6 00:15:50.291444 systemd-logind[1530]: Removed session 10. Nov 6 00:15:55.301284 systemd[1]: Started sshd@10-10.0.0.38:22-10.0.0.1:40932.service - OpenSSH per-connection server daemon (10.0.0.1:40932). Nov 6 00:15:55.366032 sshd[4111]: Accepted publickey for core from 10.0.0.1 port 40932 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:15:55.368408 sshd-session[4111]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:15:55.377060 systemd-logind[1530]: New session 11 of user core. Nov 6 00:15:55.390060 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 6 00:15:55.539724 sshd[4114]: Connection closed by 10.0.0.1 port 40932 Nov 6 00:15:55.540191 sshd-session[4111]: pam_unix(sshd:session): session closed for user core Nov 6 00:15:55.544568 systemd[1]: sshd@10-10.0.0.38:22-10.0.0.1:40932.service: Deactivated successfully. Nov 6 00:15:55.547606 systemd[1]: session-11.scope: Deactivated successfully. Nov 6 00:15:55.549901 systemd-logind[1530]: Session 11 logged out. Waiting for processes to exit. Nov 6 00:15:55.552300 systemd-logind[1530]: Removed session 11. Nov 6 00:16:00.557258 systemd[1]: Started sshd@11-10.0.0.38:22-10.0.0.1:56698.service - OpenSSH per-connection server daemon (10.0.0.1:56698). Nov 6 00:16:00.624934 sshd[4129]: Accepted publickey for core from 10.0.0.1 port 56698 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:00.627139 sshd-session[4129]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:00.633276 systemd-logind[1530]: New session 12 of user core. Nov 6 00:16:00.642115 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 6 00:16:00.795953 sshd[4132]: Connection closed by 10.0.0.1 port 56698 Nov 6 00:16:00.796353 sshd-session[4129]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:00.802585 systemd[1]: sshd@11-10.0.0.38:22-10.0.0.1:56698.service: Deactivated successfully. Nov 6 00:16:00.804734 systemd[1]: session-12.scope: Deactivated successfully. Nov 6 00:16:00.805957 systemd-logind[1530]: Session 12 logged out. Waiting for processes to exit. Nov 6 00:16:00.807749 systemd-logind[1530]: Removed session 12. Nov 6 00:16:05.816225 systemd[1]: Started sshd@12-10.0.0.38:22-10.0.0.1:56704.service - OpenSSH per-connection server daemon (10.0.0.1:56704). Nov 6 00:16:05.864276 sshd[4148]: Accepted publickey for core from 10.0.0.1 port 56704 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:05.866246 sshd-session[4148]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:05.871691 systemd-logind[1530]: New session 13 of user core. Nov 6 00:16:05.886088 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 6 00:16:06.022153 sshd[4151]: Connection closed by 10.0.0.1 port 56704 Nov 6 00:16:06.022684 sshd-session[4148]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:06.036325 systemd[1]: sshd@12-10.0.0.38:22-10.0.0.1:56704.service: Deactivated successfully. Nov 6 00:16:06.039311 systemd[1]: session-13.scope: Deactivated successfully. Nov 6 00:16:06.042112 systemd-logind[1530]: Session 13 logged out. Waiting for processes to exit. Nov 6 00:16:06.045990 systemd[1]: Started sshd@13-10.0.0.38:22-10.0.0.1:57072.service - OpenSSH per-connection server daemon (10.0.0.1:57072). Nov 6 00:16:06.046754 systemd-logind[1530]: Removed session 13. Nov 6 00:16:06.126160 sshd[4165]: Accepted publickey for core from 10.0.0.1 port 57072 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:06.128697 sshd-session[4165]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:06.134690 systemd-logind[1530]: New session 14 of user core. Nov 6 00:16:06.150091 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 6 00:16:06.420189 sshd[4168]: Connection closed by 10.0.0.1 port 57072 Nov 6 00:16:06.420489 sshd-session[4165]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:06.439403 systemd[1]: sshd@13-10.0.0.38:22-10.0.0.1:57072.service: Deactivated successfully. Nov 6 00:16:06.441915 systemd[1]: session-14.scope: Deactivated successfully. Nov 6 00:16:06.443036 systemd-logind[1530]: Session 14 logged out. Waiting for processes to exit. Nov 6 00:16:06.446499 systemd[1]: Started sshd@14-10.0.0.38:22-10.0.0.1:57086.service - OpenSSH per-connection server daemon (10.0.0.1:57086). Nov 6 00:16:06.447251 systemd-logind[1530]: Removed session 14. Nov 6 00:16:06.509594 sshd[4180]: Accepted publickey for core from 10.0.0.1 port 57086 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:06.511425 sshd-session[4180]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:06.518258 systemd-logind[1530]: New session 15 of user core. Nov 6 00:16:06.529070 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 6 00:16:06.754888 sshd[4183]: Connection closed by 10.0.0.1 port 57086 Nov 6 00:16:06.755385 sshd-session[4180]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:06.760827 systemd[1]: sshd@14-10.0.0.38:22-10.0.0.1:57086.service: Deactivated successfully. Nov 6 00:16:06.763936 systemd[1]: session-15.scope: Deactivated successfully. Nov 6 00:16:06.766652 systemd-logind[1530]: Session 15 logged out. Waiting for processes to exit. Nov 6 00:16:06.768484 systemd-logind[1530]: Removed session 15. Nov 6 00:16:11.797309 systemd[1]: Started sshd@15-10.0.0.38:22-10.0.0.1:57094.service - OpenSSH per-connection server daemon (10.0.0.1:57094). Nov 6 00:16:11.986491 sshd[4198]: Accepted publickey for core from 10.0.0.1 port 57094 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:11.990647 sshd-session[4198]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:12.007367 systemd-logind[1530]: New session 16 of user core. Nov 6 00:16:12.016117 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 6 00:16:12.306219 sshd[4201]: Connection closed by 10.0.0.1 port 57094 Nov 6 00:16:12.306984 sshd-session[4198]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:12.320145 systemd[1]: sshd@15-10.0.0.38:22-10.0.0.1:57094.service: Deactivated successfully. Nov 6 00:16:12.327990 systemd[1]: session-16.scope: Deactivated successfully. Nov 6 00:16:12.335684 systemd-logind[1530]: Session 16 logged out. Waiting for processes to exit. Nov 6 00:16:12.342097 systemd-logind[1530]: Removed session 16. Nov 6 00:16:17.343526 systemd[1]: Started sshd@16-10.0.0.38:22-10.0.0.1:49750.service - OpenSSH per-connection server daemon (10.0.0.1:49750). Nov 6 00:16:17.505081 sshd[4216]: Accepted publickey for core from 10.0.0.1 port 49750 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:17.515785 sshd-session[4216]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:17.526813 systemd-logind[1530]: New session 17 of user core. Nov 6 00:16:17.539155 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 6 00:16:17.826399 sshd[4219]: Connection closed by 10.0.0.1 port 49750 Nov 6 00:16:17.828969 sshd-session[4216]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:17.846749 systemd[1]: sshd@16-10.0.0.38:22-10.0.0.1:49750.service: Deactivated successfully. Nov 6 00:16:17.854617 systemd[1]: session-17.scope: Deactivated successfully. Nov 6 00:16:17.865589 systemd-logind[1530]: Session 17 logged out. Waiting for processes to exit. Nov 6 00:16:17.871478 systemd-logind[1530]: Removed session 17. Nov 6 00:16:19.249012 kubelet[2713]: E1106 00:16:19.246687 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:22.240545 kubelet[2713]: E1106 00:16:22.238287 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:22.875320 systemd[1]: Started sshd@17-10.0.0.38:22-10.0.0.1:49764.service - OpenSSH per-connection server daemon (10.0.0.1:49764). Nov 6 00:16:23.111803 sshd[4232]: Accepted publickey for core from 10.0.0.1 port 49764 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:23.126752 sshd-session[4232]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:23.181830 systemd-logind[1530]: New session 18 of user core. Nov 6 00:16:23.210360 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 6 00:16:23.634105 sshd[4235]: Connection closed by 10.0.0.1 port 49764 Nov 6 00:16:23.644240 sshd-session[4232]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:23.671338 systemd[1]: sshd@17-10.0.0.38:22-10.0.0.1:49764.service: Deactivated successfully. Nov 6 00:16:23.683880 systemd[1]: session-18.scope: Deactivated successfully. Nov 6 00:16:23.688297 systemd-logind[1530]: Session 18 logged out. Waiting for processes to exit. Nov 6 00:16:23.689576 systemd-logind[1530]: Removed session 18. Nov 6 00:16:27.175797 kernel: hrtimer: interrupt took 3811607 ns Nov 6 00:16:28.667242 systemd[1]: Started sshd@18-10.0.0.38:22-10.0.0.1:56946.service - OpenSSH per-connection server daemon (10.0.0.1:56946). Nov 6 00:16:28.809546 sshd[4248]: Accepted publickey for core from 10.0.0.1 port 56946 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:28.815071 sshd-session[4248]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:28.844494 systemd-logind[1530]: New session 19 of user core. Nov 6 00:16:28.870452 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 6 00:16:29.130858 sshd[4251]: Connection closed by 10.0.0.1 port 56946 Nov 6 00:16:29.129965 sshd-session[4248]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:29.148494 systemd[1]: sshd@18-10.0.0.38:22-10.0.0.1:56946.service: Deactivated successfully. Nov 6 00:16:29.155145 systemd[1]: session-19.scope: Deactivated successfully. Nov 6 00:16:29.162468 systemd-logind[1530]: Session 19 logged out. Waiting for processes to exit. Nov 6 00:16:29.166404 systemd-logind[1530]: Removed session 19. Nov 6 00:16:34.171071 systemd[1]: Started sshd@19-10.0.0.38:22-10.0.0.1:56950.service - OpenSSH per-connection server daemon (10.0.0.1:56950). Nov 6 00:16:34.341440 sshd[4265]: Accepted publickey for core from 10.0.0.1 port 56950 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:34.341173 sshd-session[4265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:34.369098 systemd-logind[1530]: New session 20 of user core. Nov 6 00:16:34.396323 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 6 00:16:34.719640 sshd[4268]: Connection closed by 10.0.0.1 port 56950 Nov 6 00:16:34.722435 sshd-session[4265]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:34.743451 systemd[1]: sshd@19-10.0.0.38:22-10.0.0.1:56950.service: Deactivated successfully. Nov 6 00:16:34.753660 systemd[1]: session-20.scope: Deactivated successfully. Nov 6 00:16:34.759872 systemd-logind[1530]: Session 20 logged out. Waiting for processes to exit. Nov 6 00:16:34.774354 systemd[1]: Started sshd@20-10.0.0.38:22-10.0.0.1:56956.service - OpenSSH per-connection server daemon (10.0.0.1:56956). Nov 6 00:16:34.783259 systemd-logind[1530]: Removed session 20. Nov 6 00:16:34.934415 sshd[4281]: Accepted publickey for core from 10.0.0.1 port 56956 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:34.942038 sshd-session[4281]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:34.970170 systemd-logind[1530]: New session 21 of user core. Nov 6 00:16:34.979187 systemd[1]: Started session-21.scope - Session 21 of User core. Nov 6 00:16:35.956850 sshd[4284]: Connection closed by 10.0.0.1 port 56956 Nov 6 00:16:35.966693 sshd-session[4281]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:36.004153 systemd[1]: sshd@20-10.0.0.38:22-10.0.0.1:56956.service: Deactivated successfully. Nov 6 00:16:36.014007 systemd[1]: session-21.scope: Deactivated successfully. Nov 6 00:16:36.039613 systemd-logind[1530]: Session 21 logged out. Waiting for processes to exit. Nov 6 00:16:36.064954 systemd[1]: Started sshd@21-10.0.0.38:22-10.0.0.1:56958.service - OpenSSH per-connection server daemon (10.0.0.1:56958). Nov 6 00:16:36.067127 systemd-logind[1530]: Removed session 21. Nov 6 00:16:36.258671 sshd[4296]: Accepted publickey for core from 10.0.0.1 port 56958 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:36.264837 sshd-session[4296]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:36.288868 systemd-logind[1530]: New session 22 of user core. Nov 6 00:16:36.307714 systemd[1]: Started session-22.scope - Session 22 of User core. Nov 6 00:16:38.158313 sshd[4299]: Connection closed by 10.0.0.1 port 56958 Nov 6 00:16:38.158607 sshd-session[4296]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:38.187569 systemd[1]: sshd@21-10.0.0.38:22-10.0.0.1:56958.service: Deactivated successfully. Nov 6 00:16:38.196503 systemd[1]: session-22.scope: Deactivated successfully. Nov 6 00:16:38.207021 systemd-logind[1530]: Session 22 logged out. Waiting for processes to exit. Nov 6 00:16:38.232036 systemd[1]: Started sshd@22-10.0.0.38:22-10.0.0.1:60668.service - OpenSSH per-connection server daemon (10.0.0.1:60668). Nov 6 00:16:38.241502 systemd-logind[1530]: Removed session 22. Nov 6 00:16:38.257964 kubelet[2713]: E1106 00:16:38.254120 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:38.400533 sshd[4318]: Accepted publickey for core from 10.0.0.1 port 60668 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:38.407938 sshd-session[4318]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:38.437667 systemd-logind[1530]: New session 23 of user core. Nov 6 00:16:38.450746 systemd[1]: Started session-23.scope - Session 23 of User core. Nov 6 00:16:39.205546 sshd[4321]: Connection closed by 10.0.0.1 port 60668 Nov 6 00:16:39.205940 sshd-session[4318]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:39.237737 systemd[1]: sshd@22-10.0.0.38:22-10.0.0.1:60668.service: Deactivated successfully. Nov 6 00:16:39.252987 systemd[1]: session-23.scope: Deactivated successfully. Nov 6 00:16:39.263710 systemd-logind[1530]: Session 23 logged out. Waiting for processes to exit. Nov 6 00:16:39.268207 systemd[1]: Started sshd@23-10.0.0.38:22-10.0.0.1:60672.service - OpenSSH per-connection server daemon (10.0.0.1:60672). Nov 6 00:16:39.273065 systemd-logind[1530]: Removed session 23. Nov 6 00:16:39.394753 sshd[4332]: Accepted publickey for core from 10.0.0.1 port 60672 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:39.397429 sshd-session[4332]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:39.422232 systemd-logind[1530]: New session 24 of user core. Nov 6 00:16:39.441737 systemd[1]: Started session-24.scope - Session 24 of User core. Nov 6 00:16:39.768775 sshd[4335]: Connection closed by 10.0.0.1 port 60672 Nov 6 00:16:39.769870 sshd-session[4332]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:39.792392 systemd[1]: sshd@23-10.0.0.38:22-10.0.0.1:60672.service: Deactivated successfully. Nov 6 00:16:39.798117 systemd[1]: session-24.scope: Deactivated successfully. Nov 6 00:16:39.809421 systemd-logind[1530]: Session 24 logged out. Waiting for processes to exit. Nov 6 00:16:39.826605 systemd-logind[1530]: Removed session 24. Nov 6 00:16:44.831111 systemd[1]: Started sshd@24-10.0.0.38:22-10.0.0.1:60674.service - OpenSSH per-connection server daemon (10.0.0.1:60674). Nov 6 00:16:44.972721 sshd[4350]: Accepted publickey for core from 10.0.0.1 port 60674 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:44.984125 sshd-session[4350]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:45.006429 systemd-logind[1530]: New session 25 of user core. Nov 6 00:16:45.022160 systemd[1]: Started session-25.scope - Session 25 of User core. Nov 6 00:16:45.234814 kubelet[2713]: E1106 00:16:45.234660 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:45.310983 sshd[4353]: Connection closed by 10.0.0.1 port 60674 Nov 6 00:16:45.313415 sshd-session[4350]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:45.321273 systemd[1]: sshd@24-10.0.0.38:22-10.0.0.1:60674.service: Deactivated successfully. Nov 6 00:16:45.325591 systemd[1]: session-25.scope: Deactivated successfully. Nov 6 00:16:45.329044 systemd-logind[1530]: Session 25 logged out. Waiting for processes to exit. Nov 6 00:16:45.333465 systemd-logind[1530]: Removed session 25. Nov 6 00:16:46.233364 kubelet[2713]: E1106 00:16:46.232729 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:50.382048 systemd[1]: Started sshd@25-10.0.0.38:22-10.0.0.1:44628.service - OpenSSH per-connection server daemon (10.0.0.1:44628). Nov 6 00:16:50.568614 sshd[4367]: Accepted publickey for core from 10.0.0.1 port 44628 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:50.574816 sshd-session[4367]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:50.603541 systemd-logind[1530]: New session 26 of user core. Nov 6 00:16:50.616589 systemd[1]: Started session-26.scope - Session 26 of User core. Nov 6 00:16:51.101446 sshd[4370]: Connection closed by 10.0.0.1 port 44628 Nov 6 00:16:51.101144 sshd-session[4367]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:51.130156 systemd[1]: sshd@25-10.0.0.38:22-10.0.0.1:44628.service: Deactivated successfully. Nov 6 00:16:51.142388 systemd[1]: session-26.scope: Deactivated successfully. Nov 6 00:16:51.148288 systemd-logind[1530]: Session 26 logged out. Waiting for processes to exit. Nov 6 00:16:51.167370 systemd-logind[1530]: Removed session 26. Nov 6 00:16:56.128271 systemd[1]: Started sshd@26-10.0.0.38:22-10.0.0.1:33316.service - OpenSSH per-connection server daemon (10.0.0.1:33316). Nov 6 00:16:56.237794 kubelet[2713]: E1106 00:16:56.237032 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:16:56.311634 sshd[4384]: Accepted publickey for core from 10.0.0.1 port 33316 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:16:56.314705 sshd-session[4384]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:16:56.330827 systemd-logind[1530]: New session 27 of user core. Nov 6 00:16:56.363198 systemd[1]: Started session-27.scope - Session 27 of User core. Nov 6 00:16:56.592645 sshd[4388]: Connection closed by 10.0.0.1 port 33316 Nov 6 00:16:56.593076 sshd-session[4384]: pam_unix(sshd:session): session closed for user core Nov 6 00:16:56.604747 systemd[1]: sshd@26-10.0.0.38:22-10.0.0.1:33316.service: Deactivated successfully. Nov 6 00:16:56.608954 systemd[1]: session-27.scope: Deactivated successfully. Nov 6 00:16:56.623610 systemd-logind[1530]: Session 27 logged out. Waiting for processes to exit. Nov 6 00:16:56.625127 systemd-logind[1530]: Removed session 27. Nov 6 00:17:01.235726 kubelet[2713]: E1106 00:17:01.234218 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:01.632158 systemd[1]: Started sshd@27-10.0.0.38:22-10.0.0.1:33322.service - OpenSSH per-connection server daemon (10.0.0.1:33322). Nov 6 00:17:01.789526 sshd[4404]: Accepted publickey for core from 10.0.0.1 port 33322 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:01.793999 sshd-session[4404]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:01.820348 systemd-logind[1530]: New session 28 of user core. Nov 6 00:17:01.834137 systemd[1]: Started session-28.scope - Session 28 of User core. Nov 6 00:17:02.099867 sshd[4407]: Connection closed by 10.0.0.1 port 33322 Nov 6 00:17:02.100106 sshd-session[4404]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:02.114919 systemd[1]: sshd@27-10.0.0.38:22-10.0.0.1:33322.service: Deactivated successfully. Nov 6 00:17:02.130370 systemd[1]: session-28.scope: Deactivated successfully. Nov 6 00:17:02.140016 systemd-logind[1530]: Session 28 logged out. Waiting for processes to exit. Nov 6 00:17:02.142333 systemd-logind[1530]: Removed session 28. Nov 6 00:17:07.141655 systemd[1]: Started sshd@28-10.0.0.38:22-10.0.0.1:54510.service - OpenSSH per-connection server daemon (10.0.0.1:54510). Nov 6 00:17:07.234449 kubelet[2713]: E1106 00:17:07.234375 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:07.255233 sshd[4422]: Accepted publickey for core from 10.0.0.1 port 54510 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:07.260406 sshd-session[4422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:07.284570 systemd-logind[1530]: New session 29 of user core. Nov 6 00:17:07.292201 systemd[1]: Started session-29.scope - Session 29 of User core. Nov 6 00:17:07.586809 sshd[4427]: Connection closed by 10.0.0.1 port 54510 Nov 6 00:17:07.588389 sshd-session[4422]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:07.604629 systemd[1]: sshd@28-10.0.0.38:22-10.0.0.1:54510.service: Deactivated successfully. Nov 6 00:17:07.608482 systemd[1]: session-29.scope: Deactivated successfully. Nov 6 00:17:07.611179 systemd-logind[1530]: Session 29 logged out. Waiting for processes to exit. Nov 6 00:17:07.613711 systemd-logind[1530]: Removed session 29. Nov 6 00:17:12.608594 systemd[1]: Started sshd@29-10.0.0.38:22-10.0.0.1:54518.service - OpenSSH per-connection server daemon (10.0.0.1:54518). Nov 6 00:17:12.758227 sshd[4441]: Accepted publickey for core from 10.0.0.1 port 54518 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:12.759987 sshd-session[4441]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:12.782615 systemd-logind[1530]: New session 30 of user core. Nov 6 00:17:12.793116 systemd[1]: Started session-30.scope - Session 30 of User core. Nov 6 00:17:13.187499 sshd[4444]: Connection closed by 10.0.0.1 port 54518 Nov 6 00:17:13.189265 sshd-session[4441]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:13.214030 systemd[1]: sshd@29-10.0.0.38:22-10.0.0.1:54518.service: Deactivated successfully. Nov 6 00:17:13.217325 systemd[1]: session-30.scope: Deactivated successfully. Nov 6 00:17:13.228912 systemd-logind[1530]: Session 30 logged out. Waiting for processes to exit. Nov 6 00:17:13.237189 systemd[1]: Started sshd@30-10.0.0.38:22-10.0.0.1:54520.service - OpenSSH per-connection server daemon (10.0.0.1:54520). Nov 6 00:17:13.253084 systemd-logind[1530]: Removed session 30. Nov 6 00:17:13.360553 sshd[4459]: Accepted publickey for core from 10.0.0.1 port 54520 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:13.360250 sshd-session[4459]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:13.386809 systemd-logind[1530]: New session 31 of user core. Nov 6 00:17:13.411294 systemd[1]: Started session-31.scope - Session 31 of User core. Nov 6 00:17:15.574794 containerd[1545]: time="2025-11-06T00:17:15.574692015Z" level=info msg="StopContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" with timeout 30 (s)" Nov 6 00:17:15.597300 containerd[1545]: time="2025-11-06T00:17:15.597155829Z" level=info msg="Stop container \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" with signal terminated" Nov 6 00:17:15.636907 systemd[1]: cri-containerd-f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82.scope: Deactivated successfully. Nov 6 00:17:15.647104 containerd[1545]: time="2025-11-06T00:17:15.643071071Z" level=info msg="received exit event container_id:\"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" id:\"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" pid:3305 exited_at:{seconds:1762388235 nanos:642581207}" Nov 6 00:17:15.647104 containerd[1545]: time="2025-11-06T00:17:15.643222887Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" id:\"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" pid:3305 exited_at:{seconds:1762388235 nanos:642581207}" Nov 6 00:17:15.672396 containerd[1545]: time="2025-11-06T00:17:15.672181544Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 6 00:17:15.687337 containerd[1545]: time="2025-11-06T00:17:15.687278061Z" level=info msg="TaskExit event in podsandbox handler container_id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" id:\"9e2831a35957bcfd02d56d454fd7e6b9a0e2bf8fb008d75baf80f0d02abdd6f2\" pid:4483 exited_at:{seconds:1762388235 nanos:686601986}" Nov 6 00:17:15.697143 containerd[1545]: time="2025-11-06T00:17:15.696906674Z" level=info msg="StopContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" with timeout 2 (s)" Nov 6 00:17:15.698105 containerd[1545]: time="2025-11-06T00:17:15.698081099Z" level=info msg="Stop container \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" with signal terminated" Nov 6 00:17:15.723752 systemd-networkd[1470]: lxc_health: Link DOWN Nov 6 00:17:15.723796 systemd-networkd[1470]: lxc_health: Lost carrier Nov 6 00:17:15.753232 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82-rootfs.mount: Deactivated successfully. Nov 6 00:17:15.804740 systemd[1]: cri-containerd-376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b.scope: Deactivated successfully. Nov 6 00:17:15.806860 systemd[1]: cri-containerd-376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b.scope: Consumed 8.324s CPU time, 125M memory peak, 216K read from disk, 13.3M written to disk. Nov 6 00:17:15.817712 containerd[1545]: time="2025-11-06T00:17:15.812030738Z" level=info msg="TaskExit event in podsandbox handler container_id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" pid:3377 exited_at:{seconds:1762388235 nanos:806296801}" Nov 6 00:17:15.817712 containerd[1545]: time="2025-11-06T00:17:15.812203694Z" level=info msg="received exit event container_id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" id:\"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" pid:3377 exited_at:{seconds:1762388235 nanos:806296801}" Nov 6 00:17:15.866038 containerd[1545]: time="2025-11-06T00:17:15.865414387Z" level=info msg="StopContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" returns successfully" Nov 6 00:17:15.880443 containerd[1545]: time="2025-11-06T00:17:15.880346755Z" level=info msg="StopPodSandbox for \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\"" Nov 6 00:17:15.908143 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b-rootfs.mount: Deactivated successfully. Nov 6 00:17:15.910964 containerd[1545]: time="2025-11-06T00:17:15.910643705Z" level=info msg="Container to stop \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:15.945928 systemd[1]: cri-containerd-e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b.scope: Deactivated successfully. Nov 6 00:17:15.957710 containerd[1545]: time="2025-11-06T00:17:15.946493561Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" id:\"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" pid:2893 exit_status:137 exited_at:{seconds:1762388235 nanos:946086984}" Nov 6 00:17:16.075524 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b-rootfs.mount: Deactivated successfully. Nov 6 00:17:16.171434 containerd[1545]: time="2025-11-06T00:17:16.169753971Z" level=info msg="shim disconnected" id=e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b namespace=k8s.io Nov 6 00:17:16.171434 containerd[1545]: time="2025-11-06T00:17:16.169835575Z" level=warning msg="cleaning up after shim disconnected" id=e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b namespace=k8s.io Nov 6 00:17:16.294384 containerd[1545]: time="2025-11-06T00:17:16.169846085Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 6 00:17:16.294597 containerd[1545]: time="2025-11-06T00:17:16.247028256Z" level=info msg="StopContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" returns successfully" Nov 6 00:17:16.295159 containerd[1545]: time="2025-11-06T00:17:16.295132740Z" level=info msg="StopPodSandbox for \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\"" Nov 6 00:17:16.296485 containerd[1545]: time="2025-11-06T00:17:16.295328499Z" level=info msg="Container to stop \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:16.297488 containerd[1545]: time="2025-11-06T00:17:16.295354868Z" level=info msg="Container to stop \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:16.297488 containerd[1545]: time="2025-11-06T00:17:16.297434271Z" level=info msg="Container to stop \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:16.297488 containerd[1545]: time="2025-11-06T00:17:16.297455972Z" level=info msg="Container to stop \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:16.297488 containerd[1545]: time="2025-11-06T00:17:16.297469156Z" level=info msg="Container to stop \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 6 00:17:16.338882 systemd[1]: cri-containerd-e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11.scope: Deactivated successfully. Nov 6 00:17:16.432818 containerd[1545]: time="2025-11-06T00:17:16.429025727Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" pid:2846 exit_status:137 exited_at:{seconds:1762388236 nanos:340816117}" Nov 6 00:17:16.444905 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b-shm.mount: Deactivated successfully. Nov 6 00:17:16.445130 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11-rootfs.mount: Deactivated successfully. Nov 6 00:17:16.463722 containerd[1545]: time="2025-11-06T00:17:16.463183796Z" level=info msg="received exit event sandbox_id:\"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" exit_status:137 exited_at:{seconds:1762388235 nanos:946086984}" Nov 6 00:17:16.472805 containerd[1545]: time="2025-11-06T00:17:16.472718690Z" level=info msg="TearDown network for sandbox \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" successfully" Nov 6 00:17:16.473018 containerd[1545]: time="2025-11-06T00:17:16.472991625Z" level=info msg="StopPodSandbox for \"e7a68f9216c6ba2d21cf6b794e7680bdbd0c64aaff1d86beb57b573f01de149b\" returns successfully" Nov 6 00:17:16.477426 containerd[1545]: time="2025-11-06T00:17:16.476955431Z" level=info msg="shim disconnected" id=e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11 namespace=k8s.io Nov 6 00:17:16.477426 containerd[1545]: time="2025-11-06T00:17:16.477003773Z" level=warning msg="cleaning up after shim disconnected" id=e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11 namespace=k8s.io Nov 6 00:17:16.477426 containerd[1545]: time="2025-11-06T00:17:16.477021035Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 6 00:17:16.477426 containerd[1545]: time="2025-11-06T00:17:16.477203299Z" level=error msg="Failed to handle event container_id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" pid:2846 exit_status:137 exited_at:{seconds:1762388236 nanos:340816117} for e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11" error="failed to handle container TaskExit event: failed to stop sandbox: ttrpc: closed" Nov 6 00:17:16.535594 containerd[1545]: time="2025-11-06T00:17:16.535449292Z" level=info msg="received exit event sandbox_id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" exit_status:137 exited_at:{seconds:1762388236 nanos:340816117}" Nov 6 00:17:16.539531 containerd[1545]: time="2025-11-06T00:17:16.539450739Z" level=info msg="TearDown network for sandbox \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" successfully" Nov 6 00:17:16.539531 containerd[1545]: time="2025-11-06T00:17:16.539493440Z" level=info msg="StopPodSandbox for \"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" returns successfully" Nov 6 00:17:16.542290 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11-shm.mount: Deactivated successfully. Nov 6 00:17:16.593498 kubelet[2713]: I1106 00:17:16.593313 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2633eddd-03ad-432e-944c-01548ec50159-cilium-config-path\") pod \"2633eddd-03ad-432e-944c-01548ec50159\" (UID: \"2633eddd-03ad-432e-944c-01548ec50159\") " Nov 6 00:17:16.593498 kubelet[2713]: I1106 00:17:16.593379 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxzxp\" (UniqueName: \"kubernetes.io/projected/2633eddd-03ad-432e-944c-01548ec50159-kube-api-access-qxzxp\") pod \"2633eddd-03ad-432e-944c-01548ec50159\" (UID: \"2633eddd-03ad-432e-944c-01548ec50159\") " Nov 6 00:17:16.600061 kubelet[2713]: I1106 00:17:16.599998 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2633eddd-03ad-432e-944c-01548ec50159-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2633eddd-03ad-432e-944c-01548ec50159" (UID: "2633eddd-03ad-432e-944c-01548ec50159"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 6 00:17:16.607144 kubelet[2713]: I1106 00:17:16.607055 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2633eddd-03ad-432e-944c-01548ec50159-kube-api-access-qxzxp" (OuterVolumeSpecName: "kube-api-access-qxzxp") pod "2633eddd-03ad-432e-944c-01548ec50159" (UID: "2633eddd-03ad-432e-944c-01548ec50159"). InnerVolumeSpecName "kube-api-access-qxzxp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 6 00:17:16.694470 kubelet[2713]: I1106 00:17:16.694306 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-etc-cni-netd\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.695012 kubelet[2713]: I1106 00:17:16.694990 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-hostproc\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.695260 kubelet[2713]: I1106 00:17:16.695238 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-net\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.696035 kubelet[2713]: I1106 00:17:16.696002 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-hostproc" (OuterVolumeSpecName: "hostproc") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.696231 kubelet[2713]: I1106 00:17:16.696159 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.696389 kubelet[2713]: I1106 00:17:16.696321 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.702921 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-hubble-tls\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.703003 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-cgroup\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.703065 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk25r\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-kube-api-access-wk25r\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.703083 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cni-path\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.703127 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-config-path\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705041 kubelet[2713]: I1106 00:17:16.703145 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-run\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703166 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-kernel\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703213 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9b0aebf9-f411-4839-bf6e-13273982517b-clustermesh-secrets\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703233 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-xtables-lock\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703283 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-bpf-maps\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703302 2713 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-lib-modules\") pod \"9b0aebf9-f411-4839-bf6e-13273982517b\" (UID: \"9b0aebf9-f411-4839-bf6e-13273982517b\") " Nov 6 00:17:16.705625 kubelet[2713]: I1106 00:17:16.703431 2713 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2633eddd-03ad-432e-944c-01548ec50159-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.703492 2713 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qxzxp\" (UniqueName: \"kubernetes.io/projected/2633eddd-03ad-432e-944c-01548ec50159-kube-api-access-qxzxp\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.703509 2713 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.703522 2713 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.703566 2713 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-hostproc\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.703649 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.708304 kubelet[2713]: I1106 00:17:16.707275 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.708654 kubelet[2713]: I1106 00:17:16.707330 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.708832 kubelet[2713]: I1106 00:17:16.708802 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cni-path" (OuterVolumeSpecName: "cni-path") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.709662 kubelet[2713]: I1106 00:17:16.709559 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.714447 kubelet[2713]: I1106 00:17:16.709867 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.714447 kubelet[2713]: I1106 00:17:16.711199 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Nov 6 00:17:16.718935 kubelet[2713]: I1106 00:17:16.718884 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Nov 6 00:17:16.732281 kubelet[2713]: I1106 00:17:16.732200 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 6 00:17:16.751890 systemd[1]: var-lib-kubelet-pods-2633eddd\x2d03ad\x2d432e\x2d944c\x2d01548ec50159-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dqxzxp.mount: Deactivated successfully. Nov 6 00:17:16.756145 kubelet[2713]: I1106 00:17:16.755629 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b0aebf9-f411-4839-bf6e-13273982517b-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Nov 6 00:17:16.752044 systemd[1]: var-lib-kubelet-pods-9b0aebf9\x2df411\x2d4839\x2dbf6e\x2d13273982517b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dwk25r.mount: Deactivated successfully. Nov 6 00:17:16.752135 systemd[1]: var-lib-kubelet-pods-9b0aebf9\x2df411\x2d4839\x2dbf6e\x2d13273982517b-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Nov 6 00:17:16.752241 systemd[1]: var-lib-kubelet-pods-9b0aebf9\x2df411\x2d4839\x2dbf6e\x2d13273982517b-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Nov 6 00:17:16.765329 kubelet[2713]: I1106 00:17:16.763973 2713 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-kube-api-access-wk25r" (OuterVolumeSpecName: "kube-api-access-wk25r") pod "9b0aebf9-f411-4839-bf6e-13273982517b" (UID: "9b0aebf9-f411-4839-bf6e-13273982517b"). InnerVolumeSpecName "kube-api-access-wk25r". PluginName "kubernetes.io/projected", VolumeGIDValue "" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804213 2713 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804270 2713 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-hubble-tls\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804284 2713 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wk25r\" (UniqueName: \"kubernetes.io/projected/9b0aebf9-f411-4839-bf6e-13273982517b-kube-api-access-wk25r\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804302 2713 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cni-path\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804317 2713 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804329 2713 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-cilium-run\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.804449 kubelet[2713]: I1106 00:17:16.804339 2713 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9b0aebf9-f411-4839-bf6e-13273982517b-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.809582 kubelet[2713]: I1106 00:17:16.804351 2713 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.809690 kubelet[2713]: I1106 00:17:16.809589 2713 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-bpf-maps\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.809690 kubelet[2713]: I1106 00:17:16.809614 2713 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-lib-modules\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:16.809690 kubelet[2713]: I1106 00:17:16.809629 2713 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9b0aebf9-f411-4839-bf6e-13273982517b-xtables-lock\") on node \"localhost\" DevicePath \"\"" Nov 6 00:17:17.167135 kubelet[2713]: I1106 00:17:17.166966 2713 scope.go:117] "RemoveContainer" containerID="f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82" Nov 6 00:17:17.186682 containerd[1545]: time="2025-11-06T00:17:17.186617762Z" level=info msg="RemoveContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\"" Nov 6 00:17:17.212274 systemd[1]: Removed slice kubepods-besteffort-pod2633eddd_03ad_432e_944c_01548ec50159.slice - libcontainer container kubepods-besteffort-pod2633eddd_03ad_432e_944c_01548ec50159.slice. Nov 6 00:17:17.250925 systemd[1]: Removed slice kubepods-burstable-pod9b0aebf9_f411_4839_bf6e_13273982517b.slice - libcontainer container kubepods-burstable-pod9b0aebf9_f411_4839_bf6e_13273982517b.slice. Nov 6 00:17:17.251182 systemd[1]: kubepods-burstable-pod9b0aebf9_f411_4839_bf6e_13273982517b.slice: Consumed 8.450s CPU time, 125.3M memory peak, 220K read from disk, 16.6M written to disk. Nov 6 00:17:17.284120 containerd[1545]: time="2025-11-06T00:17:17.283602950Z" level=info msg="RemoveContainer for \"f85d631426df5dc98ad7c7168160bf522e4525609615547edf4faf563a3e8b82\" returns successfully" Nov 6 00:17:17.284970 kubelet[2713]: I1106 00:17:17.284789 2713 scope.go:117] "RemoveContainer" containerID="376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b" Nov 6 00:17:17.323461 containerd[1545]: time="2025-11-06T00:17:17.317907520Z" level=info msg="RemoveContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\"" Nov 6 00:17:17.350284 containerd[1545]: time="2025-11-06T00:17:17.350133519Z" level=info msg="RemoveContainer for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" returns successfully" Nov 6 00:17:17.352388 kubelet[2713]: I1106 00:17:17.352327 2713 scope.go:117] "RemoveContainer" containerID="0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc" Nov 6 00:17:17.360800 sshd[4462]: Connection closed by 10.0.0.1 port 54520 Nov 6 00:17:17.361677 sshd-session[4459]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:17.369515 containerd[1545]: time="2025-11-06T00:17:17.369452827Z" level=info msg="RemoveContainer for \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\"" Nov 6 00:17:17.394464 systemd[1]: sshd@30-10.0.0.38:22-10.0.0.1:54520.service: Deactivated successfully. Nov 6 00:17:17.402208 systemd[1]: session-31.scope: Deactivated successfully. Nov 6 00:17:17.402759 systemd[1]: session-31.scope: Consumed 1.076s CPU time, 26.2M memory peak. Nov 6 00:17:17.406414 containerd[1545]: time="2025-11-06T00:17:17.406315762Z" level=info msg="RemoveContainer for \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" returns successfully" Nov 6 00:17:17.406802 kubelet[2713]: I1106 00:17:17.406629 2713 scope.go:117] "RemoveContainer" containerID="7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b" Nov 6 00:17:17.408049 systemd-logind[1530]: Session 31 logged out. Waiting for processes to exit. Nov 6 00:17:17.411696 systemd-logind[1530]: Removed session 31. Nov 6 00:17:17.413305 systemd[1]: Started sshd@31-10.0.0.38:22-10.0.0.1:47794.service - OpenSSH per-connection server daemon (10.0.0.1:47794). Nov 6 00:17:17.415330 containerd[1545]: time="2025-11-06T00:17:17.415273477Z" level=info msg="RemoveContainer for \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\"" Nov 6 00:17:17.437169 containerd[1545]: time="2025-11-06T00:17:17.437102416Z" level=info msg="RemoveContainer for \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" returns successfully" Nov 6 00:17:17.438077 kubelet[2713]: I1106 00:17:17.438021 2713 scope.go:117] "RemoveContainer" containerID="fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa" Nov 6 00:17:17.445387 containerd[1545]: time="2025-11-06T00:17:17.445222331Z" level=info msg="RemoveContainer for \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\"" Nov 6 00:17:17.455067 containerd[1545]: time="2025-11-06T00:17:17.454972100Z" level=info msg="RemoveContainer for \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" returns successfully" Nov 6 00:17:17.455639 kubelet[2713]: I1106 00:17:17.455580 2713 scope.go:117] "RemoveContainer" containerID="c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54" Nov 6 00:17:17.460200 containerd[1545]: time="2025-11-06T00:17:17.460123285Z" level=info msg="RemoveContainer for \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\"" Nov 6 00:17:17.466296 containerd[1545]: time="2025-11-06T00:17:17.466176582Z" level=info msg="RemoveContainer for \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" returns successfully" Nov 6 00:17:17.466915 kubelet[2713]: I1106 00:17:17.466868 2713 scope.go:117] "RemoveContainer" containerID="376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b" Nov 6 00:17:17.467323 containerd[1545]: time="2025-11-06T00:17:17.467238585Z" level=error msg="ContainerStatus for \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\": not found" Nov 6 00:17:17.468890 kubelet[2713]: E1106 00:17:17.468743 2713 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\": not found" containerID="376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b" Nov 6 00:17:17.469001 kubelet[2713]: I1106 00:17:17.468891 2713 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b"} err="failed to get container status \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\": rpc error: code = NotFound desc = an error occurred when try to find container \"376d4df23265be412d2ded1ad6e8d4faa5a8b93cebd050468e107c3d76ff250b\": not found" Nov 6 00:17:17.469053 kubelet[2713]: I1106 00:17:17.469004 2713 scope.go:117] "RemoveContainer" containerID="0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc" Nov 6 00:17:17.469296 containerd[1545]: time="2025-11-06T00:17:17.469258233Z" level=error msg="ContainerStatus for \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\": not found" Nov 6 00:17:17.469722 kubelet[2713]: E1106 00:17:17.469695 2713 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\": not found" containerID="0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc" Nov 6 00:17:17.469887 kubelet[2713]: I1106 00:17:17.469825 2713 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc"} err="failed to get container status \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\": rpc error: code = NotFound desc = an error occurred when try to find container \"0daa47aa96dd5c460d38db6c70d7ba919fa3702da019fb1fbe651406ff4647bc\": not found" Nov 6 00:17:17.469887 kubelet[2713]: I1106 00:17:17.469854 2713 scope.go:117] "RemoveContainer" containerID="7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b" Nov 6 00:17:17.470131 containerd[1545]: time="2025-11-06T00:17:17.470085424Z" level=error msg="ContainerStatus for \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\": not found" Nov 6 00:17:17.470270 kubelet[2713]: E1106 00:17:17.470198 2713 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\": not found" containerID="7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b" Nov 6 00:17:17.470270 kubelet[2713]: I1106 00:17:17.470231 2713 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b"} err="failed to get container status \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\": rpc error: code = NotFound desc = an error occurred when try to find container \"7f2505f450dbc3dbc3dcd0b04de14e93cd0ce778f5661a65f5ef1f7efb7b4e9b\": not found" Nov 6 00:17:17.470270 kubelet[2713]: I1106 00:17:17.470251 2713 scope.go:117] "RemoveContainer" containerID="fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa" Nov 6 00:17:17.470534 containerd[1545]: time="2025-11-06T00:17:17.470498714Z" level=error msg="ContainerStatus for \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\": not found" Nov 6 00:17:17.470673 kubelet[2713]: E1106 00:17:17.470632 2713 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\": not found" containerID="fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa" Nov 6 00:17:17.470673 kubelet[2713]: I1106 00:17:17.470653 2713 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa"} err="failed to get container status \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\": rpc error: code = NotFound desc = an error occurred when try to find container \"fd8498bdcf36a4a1a332bdec04f3cd6abd3c93959e59cf2110efaf828acb8dfa\": not found" Nov 6 00:17:17.470673 kubelet[2713]: I1106 00:17:17.470669 2713 scope.go:117] "RemoveContainer" containerID="c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54" Nov 6 00:17:17.470909 containerd[1545]: time="2025-11-06T00:17:17.470868671Z" level=error msg="ContainerStatus for \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\": not found" Nov 6 00:17:17.471195 kubelet[2713]: E1106 00:17:17.471145 2713 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\": not found" containerID="c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54" Nov 6 00:17:17.471310 kubelet[2713]: I1106 00:17:17.471208 2713 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54"} err="failed to get container status \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\": rpc error: code = NotFound desc = an error occurred when try to find container \"c4385c1b9724ce006d0482dee85799cd9c40cc00a8de9793ca693a7dfc861f54\": not found" Nov 6 00:17:17.503114 sshd[4611]: Accepted publickey for core from 10.0.0.1 port 47794 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:17.504418 sshd-session[4611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:17.521929 systemd-logind[1530]: New session 32 of user core. Nov 6 00:17:17.537149 systemd[1]: Started session-32.scope - Session 32 of User core. Nov 6 00:17:18.155201 containerd[1545]: time="2025-11-06T00:17:18.154881476Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" id:\"e51dc040be06d795dab8b6df394cead33ee00af1a2c0c0253f4b2b599873ce11\" pid:2846 exit_status:137 exited_at:{seconds:1762388236 nanos:340816117}" Nov 6 00:17:18.529647 sshd[4614]: Connection closed by 10.0.0.1 port 47794 Nov 6 00:17:18.532062 sshd-session[4611]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:18.553652 systemd[1]: sshd@31-10.0.0.38:22-10.0.0.1:47794.service: Deactivated successfully. Nov 6 00:17:18.564803 systemd[1]: session-32.scope: Deactivated successfully. Nov 6 00:17:18.571559 systemd-logind[1530]: Session 32 logged out. Waiting for processes to exit. Nov 6 00:17:18.587124 systemd[1]: Started sshd@32-10.0.0.38:22-10.0.0.1:47800.service - OpenSSH per-connection server daemon (10.0.0.1:47800). Nov 6 00:17:18.594380 systemd-logind[1530]: Removed session 32. Nov 6 00:17:18.655153 kubelet[2713]: I1106 00:17:18.655057 2713 memory_manager.go:355] "RemoveStaleState removing state" podUID="9b0aebf9-f411-4839-bf6e-13273982517b" containerName="cilium-agent" Nov 6 00:17:18.655153 kubelet[2713]: I1106 00:17:18.655118 2713 memory_manager.go:355] "RemoveStaleState removing state" podUID="2633eddd-03ad-432e-944c-01548ec50159" containerName="cilium-operator" Nov 6 00:17:18.685981 systemd[1]: Created slice kubepods-burstable-pod548918fb_d6da_4927_ba36_11010c46e4f0.slice - libcontainer container kubepods-burstable-pod548918fb_d6da_4927_ba36_11010c46e4f0.slice. Nov 6 00:17:18.747446 sshd[4626]: Accepted publickey for core from 10.0.0.1 port 47800 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:18.756126 sshd-session[4626]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:18.759036 kubelet[2713]: I1106 00:17:18.758911 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-hostproc\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.759036 kubelet[2713]: I1106 00:17:18.758989 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-xtables-lock\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.759291 kubelet[2713]: I1106 00:17:18.759240 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-cilium-cgroup\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763469 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-host-proc-sys-kernel\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763598 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/548918fb-d6da-4927-ba36-11010c46e4f0-hubble-tls\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763630 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/548918fb-d6da-4927-ba36-11010c46e4f0-cilium-config-path\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763805 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-lib-modules\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763830 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-host-proc-sys-net\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764116 kubelet[2713]: I1106 00:17:18.763851 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-cni-path\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763884 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-cilium-run\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763903 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/548918fb-d6da-4927-ba36-11010c46e4f0-cilium-ipsec-secrets\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763926 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-bpf-maps\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763944 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/548918fb-d6da-4927-ba36-11010c46e4f0-etc-cni-netd\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763961 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb57d\" (UniqueName: \"kubernetes.io/projected/548918fb-d6da-4927-ba36-11010c46e4f0-kube-api-access-nb57d\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.764412 kubelet[2713]: I1106 00:17:18.763982 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/548918fb-d6da-4927-ba36-11010c46e4f0-clustermesh-secrets\") pod \"cilium-lm5sd\" (UID: \"548918fb-d6da-4927-ba36-11010c46e4f0\") " pod="kube-system/cilium-lm5sd" Nov 6 00:17:18.788590 systemd-logind[1530]: New session 33 of user core. Nov 6 00:17:18.804360 systemd[1]: Started session-33.scope - Session 33 of User core. Nov 6 00:17:18.894817 sshd[4629]: Connection closed by 10.0.0.1 port 47800 Nov 6 00:17:18.899617 sshd-session[4626]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:18.924991 systemd[1]: sshd@32-10.0.0.38:22-10.0.0.1:47800.service: Deactivated successfully. Nov 6 00:17:18.940095 systemd[1]: session-33.scope: Deactivated successfully. Nov 6 00:17:18.998899 kubelet[2713]: E1106 00:17:18.998812 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:19.001282 containerd[1545]: time="2025-11-06T00:17:19.001174398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lm5sd,Uid:548918fb-d6da-4927-ba36-11010c46e4f0,Namespace:kube-system,Attempt:0,}" Nov 6 00:17:19.032538 systemd-logind[1530]: Session 33 logged out. Waiting for processes to exit. Nov 6 00:17:19.059588 systemd[1]: Started sshd@33-10.0.0.38:22-10.0.0.1:47810.service - OpenSSH per-connection server daemon (10.0.0.1:47810). Nov 6 00:17:19.072492 systemd-logind[1530]: Removed session 33. Nov 6 00:17:19.168832 containerd[1545]: time="2025-11-06T00:17:19.168713846Z" level=info msg="connecting to shim 03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" namespace=k8s.io protocol=ttrpc version=3 Nov 6 00:17:19.227260 sshd[4641]: Accepted publickey for core from 10.0.0.1 port 47810 ssh2: RSA SHA256:PmSYF5WO1c+PbjRA1Pm6yQw5/JNmNUR55sY7don0Q4E Nov 6 00:17:19.240563 sshd-session[4641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 6 00:17:19.245795 kubelet[2713]: I1106 00:17:19.245725 2713 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2633eddd-03ad-432e-944c-01548ec50159" path="/var/lib/kubelet/pods/2633eddd-03ad-432e-944c-01548ec50159/volumes" Nov 6 00:17:19.246834 kubelet[2713]: I1106 00:17:19.246811 2713 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b0aebf9-f411-4839-bf6e-13273982517b" path="/var/lib/kubelet/pods/9b0aebf9-f411-4839-bf6e-13273982517b/volumes" Nov 6 00:17:19.274199 systemd[1]: Started cri-containerd-03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72.scope - libcontainer container 03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72. Nov 6 00:17:19.289475 systemd-logind[1530]: New session 34 of user core. Nov 6 00:17:19.292089 systemd[1]: Started session-34.scope - Session 34 of User core. Nov 6 00:17:19.463688 containerd[1545]: time="2025-11-06T00:17:19.460157377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-lm5sd,Uid:548918fb-d6da-4927-ba36-11010c46e4f0,Namespace:kube-system,Attempt:0,} returns sandbox id \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\"" Nov 6 00:17:19.463986 kubelet[2713]: E1106 00:17:19.463539 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:19.475564 containerd[1545]: time="2025-11-06T00:17:19.474911711Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Nov 6 00:17:19.531756 containerd[1545]: time="2025-11-06T00:17:19.531638834Z" level=info msg="Container cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:17:19.565132 containerd[1545]: time="2025-11-06T00:17:19.556933003Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\"" Nov 6 00:17:19.565132 containerd[1545]: time="2025-11-06T00:17:19.562124974Z" level=info msg="StartContainer for \"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\"" Nov 6 00:17:19.587870 containerd[1545]: time="2025-11-06T00:17:19.587784311Z" level=info msg="connecting to shim cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" protocol=ttrpc version=3 Nov 6 00:17:19.676488 systemd[1]: Started cri-containerd-cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb.scope - libcontainer container cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb. Nov 6 00:17:19.862276 systemd[1]: cri-containerd-cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb.scope: Deactivated successfully. Nov 6 00:17:19.870381 containerd[1545]: time="2025-11-06T00:17:19.870306740Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\" id:\"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\" pid:4710 exited_at:{seconds:1762388239 nanos:868691945}" Nov 6 00:17:19.873957 containerd[1545]: time="2025-11-06T00:17:19.872873519Z" level=info msg="received exit event container_id:\"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\" id:\"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\" pid:4710 exited_at:{seconds:1762388239 nanos:868691945}" Nov 6 00:17:19.913281 containerd[1545]: time="2025-11-06T00:17:19.913210526Z" level=info msg="StartContainer for \"cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb\" returns successfully" Nov 6 00:17:19.959862 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cc8f1346470bb2e35174d26c4b95c37e3230cd4c340430661bf8a11ad8aa99fb-rootfs.mount: Deactivated successfully. Nov 6 00:17:20.272391 kubelet[2713]: E1106 00:17:20.271838 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:20.281908 containerd[1545]: time="2025-11-06T00:17:20.281852596Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Nov 6 00:17:20.334952 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount74986279.mount: Deactivated successfully. Nov 6 00:17:20.367757 containerd[1545]: time="2025-11-06T00:17:20.367677440Z" level=info msg="Container db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:17:20.378031 kubelet[2713]: E1106 00:17:20.377753 2713 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Nov 6 00:17:20.435964 containerd[1545]: time="2025-11-06T00:17:20.435890453Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\"" Nov 6 00:17:20.443510 containerd[1545]: time="2025-11-06T00:17:20.439412876Z" level=info msg="StartContainer for \"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\"" Nov 6 00:17:20.443510 containerd[1545]: time="2025-11-06T00:17:20.442554299Z" level=info msg="connecting to shim db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" protocol=ttrpc version=3 Nov 6 00:17:20.515166 systemd[1]: Started cri-containerd-db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0.scope - libcontainer container db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0. Nov 6 00:17:20.699211 containerd[1545]: time="2025-11-06T00:17:20.698823876Z" level=info msg="StartContainer for \"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\" returns successfully" Nov 6 00:17:20.719028 systemd[1]: cri-containerd-db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0.scope: Deactivated successfully. Nov 6 00:17:20.726012 containerd[1545]: time="2025-11-06T00:17:20.725934697Z" level=info msg="received exit event container_id:\"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\" id:\"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\" pid:4758 exited_at:{seconds:1762388240 nanos:725551284}" Nov 6 00:17:20.726899 containerd[1545]: time="2025-11-06T00:17:20.726856244Z" level=info msg="TaskExit event in podsandbox handler container_id:\"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\" id:\"db8fa208b5f43e7d24d3e62e287a804508e3b0e81a1dadd8992542cdbd370cc0\" pid:4758 exited_at:{seconds:1762388240 nanos:725551284}" Nov 6 00:17:20.890946 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3733754124.mount: Deactivated successfully. Nov 6 00:17:21.296954 kubelet[2713]: E1106 00:17:21.296729 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:21.311166 containerd[1545]: time="2025-11-06T00:17:21.307457257Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Nov 6 00:17:21.370738 containerd[1545]: time="2025-11-06T00:17:21.370670626Z" level=info msg="Container 7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:17:21.394663 containerd[1545]: time="2025-11-06T00:17:21.394176776Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\"" Nov 6 00:17:21.400379 containerd[1545]: time="2025-11-06T00:17:21.396814149Z" level=info msg="StartContainer for \"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\"" Nov 6 00:17:21.411422 containerd[1545]: time="2025-11-06T00:17:21.409922484Z" level=info msg="connecting to shim 7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" protocol=ttrpc version=3 Nov 6 00:17:21.489483 systemd[1]: Started cri-containerd-7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d.scope - libcontainer container 7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d. Nov 6 00:17:21.653258 systemd[1]: cri-containerd-7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d.scope: Deactivated successfully. Nov 6 00:17:21.663857 containerd[1545]: time="2025-11-06T00:17:21.661705780Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\" id:\"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\" pid:4803 exited_at:{seconds:1762388241 nanos:661049693}" Nov 6 00:17:21.673808 containerd[1545]: time="2025-11-06T00:17:21.673515376Z" level=info msg="received exit event container_id:\"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\" id:\"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\" pid:4803 exited_at:{seconds:1762388241 nanos:661049693}" Nov 6 00:17:21.676744 containerd[1545]: time="2025-11-06T00:17:21.676613588Z" level=info msg="StartContainer for \"7b75cca6878beac02842e01ab21136003090adc1f63de4e0def50c5e8f19796d\" returns successfully" Nov 6 00:17:22.235728 kubelet[2713]: E1106 00:17:22.234978 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:22.321420 kubelet[2713]: E1106 00:17:22.320529 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:22.330979 containerd[1545]: time="2025-11-06T00:17:22.328126728Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Nov 6 00:17:22.398818 containerd[1545]: time="2025-11-06T00:17:22.398706416Z" level=info msg="Container 62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:17:22.504658 containerd[1545]: time="2025-11-06T00:17:22.504106634Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\"" Nov 6 00:17:22.510544 containerd[1545]: time="2025-11-06T00:17:22.510482484Z" level=info msg="StartContainer for \"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\"" Nov 6 00:17:22.513060 containerd[1545]: time="2025-11-06T00:17:22.512101697Z" level=info msg="connecting to shim 62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" protocol=ttrpc version=3 Nov 6 00:17:22.591251 systemd[1]: Started cri-containerd-62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562.scope - libcontainer container 62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562. Nov 6 00:17:22.706988 systemd[1]: cri-containerd-62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562.scope: Deactivated successfully. Nov 6 00:17:22.708063 containerd[1545]: time="2025-11-06T00:17:22.707990749Z" level=info msg="received exit event container_id:\"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\" id:\"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\" pid:4841 exited_at:{seconds:1762388242 nanos:706591561}" Nov 6 00:17:22.710421 containerd[1545]: time="2025-11-06T00:17:22.709895160Z" level=info msg="TaskExit event in podsandbox handler container_id:\"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\" id:\"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\" pid:4841 exited_at:{seconds:1762388242 nanos:706591561}" Nov 6 00:17:22.724713 containerd[1545]: time="2025-11-06T00:17:22.724069123Z" level=info msg="StartContainer for \"62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562\" returns successfully" Nov 6 00:17:22.760311 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-62fd64a755df02d0979fe385466f5f1cea3df7dd7b5189f49bcb863d1e79e562-rootfs.mount: Deactivated successfully. Nov 6 00:17:23.350593 kubelet[2713]: E1106 00:17:23.349804 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:23.372353 containerd[1545]: time="2025-11-06T00:17:23.366077597Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Nov 6 00:17:23.442178 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2809782428.mount: Deactivated successfully. Nov 6 00:17:23.456490 containerd[1545]: time="2025-11-06T00:17:23.456005432Z" level=info msg="Container 41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73: CDI devices from CRI Config.CDIDevices: []" Nov 6 00:17:23.616987 containerd[1545]: time="2025-11-06T00:17:23.611410419Z" level=info msg="CreateContainer within sandbox \"03ec07c02d08caaf89d02900ff6a76271786f707099719b1565560c5659aab72\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\"" Nov 6 00:17:23.616987 containerd[1545]: time="2025-11-06T00:17:23.612237297Z" level=info msg="StartContainer for \"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\"" Nov 6 00:17:23.616987 containerd[1545]: time="2025-11-06T00:17:23.616032171Z" level=info msg="connecting to shim 41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73" address="unix:///run/containerd/s/964f87ee2e69fc3fecabdf524988bc620667a05a04fabe54f5c18f4a55d41d1d" protocol=ttrpc version=3 Nov 6 00:17:23.716586 systemd[1]: Started cri-containerd-41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73.scope - libcontainer container 41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73. Nov 6 00:17:23.892810 containerd[1545]: time="2025-11-06T00:17:23.889807352Z" level=info msg="StartContainer for \"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" returns successfully" Nov 6 00:17:24.141581 containerd[1545]: time="2025-11-06T00:17:24.137001256Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"a77eaf2c128c76afe4e5921a7c6dca077b3f286142ea14eebdc201e7321a436f\" pid:4912 exited_at:{seconds:1762388244 nanos:135880954}" Nov 6 00:17:24.364663 kubelet[2713]: E1106 00:17:24.364594 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:24.412536 kubelet[2713]: I1106 00:17:24.412351 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-lm5sd" podStartSLOduration=6.412320727 podStartE2EDuration="6.412320727s" podCreationTimestamp="2025-11-06 00:17:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-06 00:17:24.412169291 +0000 UTC m=+139.277079456" watchObservedRunningTime="2025-11-06 00:17:24.412320727 +0000 UTC m=+139.277230872" Nov 6 00:17:25.139342 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni-avx)) Nov 6 00:17:25.373220 kubelet[2713]: E1106 00:17:25.373178 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:26.566585 containerd[1545]: time="2025-11-06T00:17:26.566274392Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"2defe74ba88f5e5d85b3ebe32bbcdd344d1c5908e2f20aabd37aa03c1ba5c7cb\" pid:5019 exit_status:1 exited_at:{seconds:1762388246 nanos:565711921}" Nov 6 00:17:29.010205 containerd[1545]: time="2025-11-06T00:17:29.007731568Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"1a6ce6dc866eac6a70048126e4086fcad16138dbaaa7e8842423f7ff3a527d63\" pid:5146 exit_status:1 exited_at:{seconds:1762388249 nanos:7210075}" Nov 6 00:17:31.536977 containerd[1545]: time="2025-11-06T00:17:31.535714961Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"84e1dc4fb133c6266c64fd909aa75000aea737e5b4d8605fd3ffb0d33245cbc7\" pid:5411 exit_status:1 exited_at:{seconds:1762388251 nanos:535336268}" Nov 6 00:17:32.051014 systemd-networkd[1470]: lxc_health: Link UP Nov 6 00:17:32.053282 systemd-networkd[1470]: lxc_health: Gained carrier Nov 6 00:17:32.239392 kubelet[2713]: E1106 00:17:32.238376 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:33.042881 kubelet[2713]: E1106 00:17:33.042333 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:33.446083 kubelet[2713]: E1106 00:17:33.444239 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:33.464310 systemd-networkd[1470]: lxc_health: Gained IPv6LL Nov 6 00:17:34.104974 containerd[1545]: time="2025-11-06T00:17:34.102815868Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"a3b4b7d0aae01850e428c54bd19f23a749ba8564262cac9e2c88c92fe0c85d79\" pid:5525 exited_at:{seconds:1762388254 nanos:102334129}" Nov 6 00:17:34.453185 kubelet[2713]: E1106 00:17:34.452473 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 6 00:17:36.408364 containerd[1545]: time="2025-11-06T00:17:36.407487926Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"ec09667827a8a0eb598c4f12019657b9f5a7b2de8675220379dcf1420d773769\" pid:5554 exited_at:{seconds:1762388256 nanos:407109863}" Nov 6 00:17:38.796969 containerd[1545]: time="2025-11-06T00:17:38.796883845Z" level=info msg="TaskExit event in podsandbox handler container_id:\"41650a43117fab425d27d90a47b460bfdcc68e9deaf540a6084931148de52b73\" id:\"0891c4ebe6049d949c2b5052337a4d6c20b41d0b650eadf00e2ed94246f5b906\" pid:5581 exited_at:{seconds:1762388258 nanos:796386849}" Nov 6 00:17:38.929973 sshd[4679]: Connection closed by 10.0.0.1 port 47810 Nov 6 00:17:38.933748 sshd-session[4641]: pam_unix(sshd:session): session closed for user core Nov 6 00:17:38.942970 systemd[1]: sshd@33-10.0.0.38:22-10.0.0.1:47810.service: Deactivated successfully. Nov 6 00:17:38.953659 systemd[1]: session-34.scope: Deactivated successfully. Nov 6 00:17:38.972246 systemd-logind[1530]: Session 34 logged out. Waiting for processes to exit. Nov 6 00:17:38.975042 systemd-logind[1530]: Removed session 34.