Jan 13 20:34:13.897636 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Jan 13 19:01:45 -00 2025 Jan 13 20:34:13.897673 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1175b5bd4028ce8485b23b7d346f787308cbfa43cca7b1fefd4254406dce7d07 Jan 13 20:34:13.897686 kernel: BIOS-provided physical RAM map: Jan 13 20:34:13.897692 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Jan 13 20:34:13.897698 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Jan 13 20:34:13.897705 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 13 20:34:13.897712 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Jan 13 20:34:13.897718 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Jan 13 20:34:13.897724 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jan 13 20:34:13.897733 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jan 13 20:34:13.897739 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 13 20:34:13.897745 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 13 20:34:13.897751 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 13 20:34:13.897758 kernel: NX (Execute Disable) protection: active Jan 13 20:34:13.897766 kernel: APIC: Static calls initialized Jan 13 20:34:13.897775 kernel: SMBIOS 2.8 present. Jan 13 20:34:13.897782 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Jan 13 20:34:13.897788 kernel: Hypervisor detected: KVM Jan 13 20:34:13.897795 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 13 20:34:13.897802 kernel: kvm-clock: using sched offset of 2726229164 cycles Jan 13 20:34:13.897809 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 13 20:34:13.897816 kernel: tsc: Detected 2794.748 MHz processor Jan 13 20:34:13.897823 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 13 20:34:13.897830 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 13 20:34:13.897837 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Jan 13 20:34:13.897846 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 13 20:34:13.897853 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 13 20:34:13.897860 kernel: Using GB pages for direct mapping Jan 13 20:34:13.897867 kernel: ACPI: Early table checksum verification disabled Jan 13 20:34:13.897873 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Jan 13 20:34:13.897880 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897887 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897894 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897903 kernel: ACPI: FACS 0x000000009CFE0000 000040 Jan 13 20:34:13.897910 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897917 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897924 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897930 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 13 20:34:13.897937 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db] Jan 13 20:34:13.897944 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7] Jan 13 20:34:13.897954 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Jan 13 20:34:13.897964 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b] Jan 13 20:34:13.897971 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3] Jan 13 20:34:13.897978 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df] Jan 13 20:34:13.897985 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407] Jan 13 20:34:13.897992 kernel: No NUMA configuration found Jan 13 20:34:13.897999 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Jan 13 20:34:13.898006 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Jan 13 20:34:13.898016 kernel: Zone ranges: Jan 13 20:34:13.898023 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 13 20:34:13.898030 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Jan 13 20:34:13.898037 kernel: Normal empty Jan 13 20:34:13.898044 kernel: Movable zone start for each node Jan 13 20:34:13.898051 kernel: Early memory node ranges Jan 13 20:34:13.898058 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 13 20:34:13.898065 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Jan 13 20:34:13.898073 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Jan 13 20:34:13.898082 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 13 20:34:13.898089 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 13 20:34:13.898096 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Jan 13 20:34:13.898104 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 13 20:34:13.898111 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 13 20:34:13.898118 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 13 20:34:13.898125 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 13 20:34:13.898132 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 13 20:34:13.898139 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 13 20:34:13.898148 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 13 20:34:13.898156 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 13 20:34:13.898163 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 13 20:34:13.898170 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 13 20:34:13.898177 kernel: TSC deadline timer available Jan 13 20:34:13.898184 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Jan 13 20:34:13.898191 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 13 20:34:13.898198 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 13 20:34:13.898205 kernel: kvm-guest: setup PV sched yield Jan 13 20:34:13.898212 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jan 13 20:34:13.898222 kernel: Booting paravirtualized kernel on KVM Jan 13 20:34:13.898229 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 13 20:34:13.898236 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 13 20:34:13.898244 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Jan 13 20:34:13.898251 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Jan 13 20:34:13.898258 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 13 20:34:13.898264 kernel: kvm-guest: PV spinlocks enabled Jan 13 20:34:13.898272 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 13 20:34:13.898280 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1175b5bd4028ce8485b23b7d346f787308cbfa43cca7b1fefd4254406dce7d07 Jan 13 20:34:13.898290 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 13 20:34:13.898297 kernel: random: crng init done Jan 13 20:34:13.898304 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 13 20:34:13.898311 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 13 20:34:13.898318 kernel: Fallback order for Node 0: 0 Jan 13 20:34:13.898326 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Jan 13 20:34:13.898333 kernel: Policy zone: DMA32 Jan 13 20:34:13.898340 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 13 20:34:13.898350 kernel: Memory: 2434592K/2571752K available (12288K kernel code, 2299K rwdata, 22736K rodata, 42976K init, 2216K bss, 136900K reserved, 0K cma-reserved) Jan 13 20:34:13.898365 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 13 20:34:13.898372 kernel: ftrace: allocating 37920 entries in 149 pages Jan 13 20:34:13.898379 kernel: ftrace: allocated 149 pages with 4 groups Jan 13 20:34:13.898387 kernel: Dynamic Preempt: voluntary Jan 13 20:34:13.898394 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 13 20:34:13.898402 kernel: rcu: RCU event tracing is enabled. Jan 13 20:34:13.898409 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 13 20:34:13.898416 kernel: Trampoline variant of Tasks RCU enabled. Jan 13 20:34:13.898426 kernel: Rude variant of Tasks RCU enabled. Jan 13 20:34:13.898433 kernel: Tracing variant of Tasks RCU enabled. Jan 13 20:34:13.898440 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 13 20:34:13.898447 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 13 20:34:13.898455 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 13 20:34:13.898462 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 13 20:34:13.898469 kernel: Console: colour VGA+ 80x25 Jan 13 20:34:13.898476 kernel: printk: console [ttyS0] enabled Jan 13 20:34:13.898483 kernel: ACPI: Core revision 20230628 Jan 13 20:34:13.898493 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 13 20:34:13.898500 kernel: APIC: Switch to symmetric I/O mode setup Jan 13 20:34:13.898507 kernel: x2apic enabled Jan 13 20:34:13.898514 kernel: APIC: Switched APIC routing to: physical x2apic Jan 13 20:34:13.898521 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 13 20:34:13.898529 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 13 20:34:13.898536 kernel: kvm-guest: setup PV IPIs Jan 13 20:34:13.898553 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 13 20:34:13.898561 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Jan 13 20:34:13.898568 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Jan 13 20:34:13.898576 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 13 20:34:13.898583 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 13 20:34:13.898593 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 13 20:34:13.898600 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 13 20:34:13.898608 kernel: Spectre V2 : Mitigation: Retpolines Jan 13 20:34:13.898615 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jan 13 20:34:13.898623 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Jan 13 20:34:13.898633 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Jan 13 20:34:13.898640 kernel: RETBleed: Mitigation: untrained return thunk Jan 13 20:34:13.898648 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 13 20:34:13.898667 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 13 20:34:13.898675 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 13 20:34:13.898683 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 13 20:34:13.898691 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 13 20:34:13.898699 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 13 20:34:13.898709 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 13 20:34:13.898716 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 13 20:34:13.898724 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 13 20:34:13.898731 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 13 20:34:13.898739 kernel: Freeing SMP alternatives memory: 32K Jan 13 20:34:13.898746 kernel: pid_max: default: 32768 minimum: 301 Jan 13 20:34:13.898754 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 13 20:34:13.898761 kernel: landlock: Up and running. Jan 13 20:34:13.898769 kernel: SELinux: Initializing. Jan 13 20:34:13.898778 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 13 20:34:13.898786 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 13 20:34:13.898794 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Jan 13 20:34:13.898801 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 13 20:34:13.898809 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 13 20:34:13.898817 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 13 20:34:13.898824 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jan 13 20:34:13.898832 kernel: ... version: 0 Jan 13 20:34:13.898842 kernel: ... bit width: 48 Jan 13 20:34:13.898849 kernel: ... generic registers: 6 Jan 13 20:34:13.898856 kernel: ... value mask: 0000ffffffffffff Jan 13 20:34:13.898864 kernel: ... max period: 00007fffffffffff Jan 13 20:34:13.898871 kernel: ... fixed-purpose events: 0 Jan 13 20:34:13.898879 kernel: ... event mask: 000000000000003f Jan 13 20:34:13.898886 kernel: signal: max sigframe size: 1776 Jan 13 20:34:13.898893 kernel: rcu: Hierarchical SRCU implementation. Jan 13 20:34:13.898901 kernel: rcu: Max phase no-delay instances is 400. Jan 13 20:34:13.898908 kernel: smp: Bringing up secondary CPUs ... Jan 13 20:34:13.899025 kernel: smpboot: x86: Booting SMP configuration: Jan 13 20:34:13.899032 kernel: .... node #0, CPUs: #1 #2 #3 Jan 13 20:34:13.899040 kernel: smp: Brought up 1 node, 4 CPUs Jan 13 20:34:13.899047 kernel: smpboot: Max logical packages: 1 Jan 13 20:34:13.899055 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Jan 13 20:34:13.899063 kernel: devtmpfs: initialized Jan 13 20:34:13.899070 kernel: x86/mm: Memory block size: 128MB Jan 13 20:34:13.899078 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 13 20:34:13.899085 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 13 20:34:13.899095 kernel: pinctrl core: initialized pinctrl subsystem Jan 13 20:34:13.899103 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 13 20:34:13.899110 kernel: audit: initializing netlink subsys (disabled) Jan 13 20:34:13.899118 kernel: audit: type=2000 audit(1736800453.402:1): state=initialized audit_enabled=0 res=1 Jan 13 20:34:13.899125 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 13 20:34:13.899133 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 13 20:34:13.899140 kernel: cpuidle: using governor menu Jan 13 20:34:13.899148 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 13 20:34:13.899155 kernel: dca service started, version 1.12.1 Jan 13 20:34:13.899165 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Jan 13 20:34:13.899172 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jan 13 20:34:13.899180 kernel: PCI: Using configuration type 1 for base access Jan 13 20:34:13.899188 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 13 20:34:13.899195 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 13 20:34:13.899203 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 13 20:34:13.899210 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 13 20:34:13.899218 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 13 20:34:13.899225 kernel: ACPI: Added _OSI(Module Device) Jan 13 20:34:13.899235 kernel: ACPI: Added _OSI(Processor Device) Jan 13 20:34:13.899242 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 13 20:34:13.899250 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 13 20:34:13.899257 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 13 20:34:13.899265 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 13 20:34:13.899272 kernel: ACPI: Interpreter enabled Jan 13 20:34:13.899279 kernel: ACPI: PM: (supports S0 S3 S5) Jan 13 20:34:13.899287 kernel: ACPI: Using IOAPIC for interrupt routing Jan 13 20:34:13.899295 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 13 20:34:13.899305 kernel: PCI: Using E820 reservations for host bridge windows Jan 13 20:34:13.899312 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 13 20:34:13.899320 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 13 20:34:13.899516 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 13 20:34:13.899645 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 13 20:34:13.899803 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 13 20:34:13.899815 kernel: PCI host bridge to bus 0000:00 Jan 13 20:34:13.899955 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 13 20:34:13.900069 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 13 20:34:13.900180 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 13 20:34:13.900304 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Jan 13 20:34:13.900433 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 13 20:34:13.900543 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Jan 13 20:34:13.900672 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 13 20:34:13.900868 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Jan 13 20:34:13.900998 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Jan 13 20:34:13.901119 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Jan 13 20:34:13.901239 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Jan 13 20:34:13.901365 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Jan 13 20:34:13.901487 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 13 20:34:13.901619 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Jan 13 20:34:13.901786 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Jan 13 20:34:13.901908 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Jan 13 20:34:13.902028 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Jan 13 20:34:13.902168 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Jan 13 20:34:13.902290 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Jan 13 20:34:13.902419 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Jan 13 20:34:13.902545 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Jan 13 20:34:13.902689 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Jan 13 20:34:13.902813 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Jan 13 20:34:13.902932 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Jan 13 20:34:13.903051 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Jan 13 20:34:13.903171 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Jan 13 20:34:13.903327 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Jan 13 20:34:13.903468 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 13 20:34:13.903596 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Jan 13 20:34:13.903735 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Jan 13 20:34:13.903861 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Jan 13 20:34:13.903990 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Jan 13 20:34:13.904109 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Jan 13 20:34:13.904119 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 13 20:34:13.904131 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 13 20:34:13.904139 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 13 20:34:13.904147 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 13 20:34:13.904154 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 13 20:34:13.904162 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 13 20:34:13.904169 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 13 20:34:13.904177 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 13 20:34:13.904184 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 13 20:34:13.904192 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 13 20:34:13.904203 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 13 20:34:13.904210 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 13 20:34:13.904235 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 13 20:34:13.904243 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 13 20:34:13.904250 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 13 20:34:13.904258 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 13 20:34:13.904265 kernel: iommu: Default domain type: Translated Jan 13 20:34:13.904273 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 13 20:34:13.904280 kernel: PCI: Using ACPI for IRQ routing Jan 13 20:34:13.904291 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 13 20:34:13.904298 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Jan 13 20:34:13.904305 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Jan 13 20:34:13.904439 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 13 20:34:13.904560 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 13 20:34:13.904744 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 13 20:34:13.904755 kernel: vgaarb: loaded Jan 13 20:34:13.904763 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 13 20:34:13.904775 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 13 20:34:13.904782 kernel: clocksource: Switched to clocksource kvm-clock Jan 13 20:34:13.904790 kernel: VFS: Disk quotas dquot_6.6.0 Jan 13 20:34:13.904798 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 13 20:34:13.904805 kernel: pnp: PnP ACPI init Jan 13 20:34:13.904940 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Jan 13 20:34:13.904951 kernel: pnp: PnP ACPI: found 6 devices Jan 13 20:34:13.904959 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 13 20:34:13.904970 kernel: NET: Registered PF_INET protocol family Jan 13 20:34:13.904977 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 13 20:34:13.904985 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 13 20:34:13.904993 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 13 20:34:13.905000 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 13 20:34:13.905008 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 13 20:34:13.905015 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 13 20:34:13.905023 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 13 20:34:13.905030 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 13 20:34:13.905041 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 13 20:34:13.905048 kernel: NET: Registered PF_XDP protocol family Jan 13 20:34:13.905185 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 13 20:34:13.905302 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 13 20:34:13.905422 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 13 20:34:13.905532 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Jan 13 20:34:13.905640 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jan 13 20:34:13.905764 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Jan 13 20:34:13.905780 kernel: PCI: CLS 0 bytes, default 64 Jan 13 20:34:13.905788 kernel: Initialise system trusted keyrings Jan 13 20:34:13.905795 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 13 20:34:13.905803 kernel: Key type asymmetric registered Jan 13 20:34:13.905810 kernel: Asymmetric key parser 'x509' registered Jan 13 20:34:13.905818 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 13 20:34:13.905826 kernel: io scheduler mq-deadline registered Jan 13 20:34:13.905833 kernel: io scheduler kyber registered Jan 13 20:34:13.905841 kernel: io scheduler bfq registered Jan 13 20:34:13.905851 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 13 20:34:13.905860 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 13 20:34:13.905867 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 13 20:34:13.905875 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 13 20:34:13.905882 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 13 20:34:13.905890 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 13 20:34:13.905898 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 13 20:34:13.905905 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 13 20:34:13.905913 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 13 20:34:13.906039 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 13 20:34:13.906154 kernel: rtc_cmos 00:04: registered as rtc0 Jan 13 20:34:13.906267 kernel: rtc_cmos 00:04: setting system clock to 2025-01-13T20:34:13 UTC (1736800453) Jan 13 20:34:13.906391 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jan 13 20:34:13.906401 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 13 20:34:13.906409 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 13 20:34:13.906417 kernel: NET: Registered PF_INET6 protocol family Jan 13 20:34:13.906424 kernel: Segment Routing with IPv6 Jan 13 20:34:13.906435 kernel: In-situ OAM (IOAM) with IPv6 Jan 13 20:34:13.906443 kernel: NET: Registered PF_PACKET protocol family Jan 13 20:34:13.906450 kernel: Key type dns_resolver registered Jan 13 20:34:13.906458 kernel: IPI shorthand broadcast: enabled Jan 13 20:34:13.906466 kernel: sched_clock: Marking stable (665004747, 109029471)->(791321530, -17287312) Jan 13 20:34:13.906473 kernel: registered taskstats version 1 Jan 13 20:34:13.906481 kernel: Loading compiled-in X.509 certificates Jan 13 20:34:13.906488 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: 98739e9049f62881f4df7ffd1e39335f7f55b344' Jan 13 20:34:13.906496 kernel: Key type .fscrypt registered Jan 13 20:34:13.906506 kernel: Key type fscrypt-provisioning registered Jan 13 20:34:13.906513 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 13 20:34:13.906521 kernel: ima: Allocated hash algorithm: sha1 Jan 13 20:34:13.906528 kernel: ima: No architecture policies found Jan 13 20:34:13.906536 kernel: clk: Disabling unused clocks Jan 13 20:34:13.906543 kernel: Freeing unused kernel image (initmem) memory: 42976K Jan 13 20:34:13.906551 kernel: Write protecting the kernel read-only data: 36864k Jan 13 20:34:13.906559 kernel: Freeing unused kernel image (rodata/data gap) memory: 1840K Jan 13 20:34:13.906567 kernel: Run /init as init process Jan 13 20:34:13.906577 kernel: with arguments: Jan 13 20:34:13.906585 kernel: /init Jan 13 20:34:13.906592 kernel: with environment: Jan 13 20:34:13.906599 kernel: HOME=/ Jan 13 20:34:13.906607 kernel: TERM=linux Jan 13 20:34:13.906614 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 13 20:34:13.906624 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 13 20:34:13.906633 systemd[1]: Detected virtualization kvm. Jan 13 20:34:13.906644 systemd[1]: Detected architecture x86-64. Jan 13 20:34:13.906703 systemd[1]: Running in initrd. Jan 13 20:34:13.906712 systemd[1]: No hostname configured, using default hostname. Jan 13 20:34:13.906720 systemd[1]: Hostname set to . Jan 13 20:34:13.906728 systemd[1]: Initializing machine ID from VM UUID. Jan 13 20:34:13.906736 systemd[1]: Queued start job for default target initrd.target. Jan 13 20:34:13.906745 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:34:13.906753 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:34:13.906765 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 13 20:34:13.906784 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 13 20:34:13.906795 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 13 20:34:13.906804 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 13 20:34:13.906814 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 13 20:34:13.906825 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 13 20:34:13.906833 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:34:13.906841 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:34:13.906849 systemd[1]: Reached target paths.target - Path Units. Jan 13 20:34:13.906858 systemd[1]: Reached target slices.target - Slice Units. Jan 13 20:34:13.906866 systemd[1]: Reached target swap.target - Swaps. Jan 13 20:34:13.906874 systemd[1]: Reached target timers.target - Timer Units. Jan 13 20:34:13.906882 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 13 20:34:13.906893 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 13 20:34:13.906901 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 13 20:34:13.906909 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 13 20:34:13.906918 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:34:13.906926 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 13 20:34:13.906934 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:34:13.906942 systemd[1]: Reached target sockets.target - Socket Units. Jan 13 20:34:13.906951 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 13 20:34:13.906959 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 13 20:34:13.906972 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 13 20:34:13.906980 systemd[1]: Starting systemd-fsck-usr.service... Jan 13 20:34:13.906989 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 13 20:34:13.906997 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 13 20:34:13.907005 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:34:13.907013 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 13 20:34:13.907022 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:34:13.907030 systemd[1]: Finished systemd-fsck-usr.service. Jan 13 20:34:13.907059 systemd-journald[194]: Collecting audit messages is disabled. Jan 13 20:34:13.907081 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 13 20:34:13.907092 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 13 20:34:13.907101 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 13 20:34:13.907109 systemd-journald[194]: Journal started Jan 13 20:34:13.907130 systemd-journald[194]: Runtime Journal (/run/log/journal/6ff90c8a49e14c22b1fa5321ffaba61a) is 6.0M, max 48.4M, 42.3M free. Jan 13 20:34:13.892718 systemd-modules-load[195]: Inserted module 'overlay' Jan 13 20:34:13.912682 systemd[1]: Started systemd-journald.service - Journal Service. Jan 13 20:34:13.919683 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 13 20:34:13.921322 systemd-modules-load[195]: Inserted module 'br_netfilter' Jan 13 20:34:13.946471 kernel: Bridge firewalling registered Jan 13 20:34:13.968208 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 13 20:34:13.969117 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:34:13.986846 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:34:13.990029 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:34:13.992692 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 13 20:34:13.994079 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:34:14.003433 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:34:14.009893 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 13 20:34:14.012104 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:34:14.014534 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:34:14.019214 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 13 20:34:14.024706 dracut-cmdline[226]: dracut-dracut-053 Jan 13 20:34:14.027883 dracut-cmdline[226]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=1175b5bd4028ce8485b23b7d346f787308cbfa43cca7b1fefd4254406dce7d07 Jan 13 20:34:14.062158 systemd-resolved[234]: Positive Trust Anchors: Jan 13 20:34:14.062173 systemd-resolved[234]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 13 20:34:14.062204 systemd-resolved[234]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 13 20:34:14.074020 systemd-resolved[234]: Defaulting to hostname 'linux'. Jan 13 20:34:14.076027 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 13 20:34:14.076402 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:34:14.118689 kernel: SCSI subsystem initialized Jan 13 20:34:14.128673 kernel: Loading iSCSI transport class v2.0-870. Jan 13 20:34:14.138676 kernel: iscsi: registered transport (tcp) Jan 13 20:34:14.160691 kernel: iscsi: registered transport (qla4xxx) Jan 13 20:34:14.160741 kernel: QLogic iSCSI HBA Driver Jan 13 20:34:14.213713 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 13 20:34:14.226810 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 13 20:34:14.254695 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 13 20:34:14.254735 kernel: device-mapper: uevent: version 1.0.3 Jan 13 20:34:14.256690 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 13 20:34:14.297684 kernel: raid6: avx2x4 gen() 30106 MB/s Jan 13 20:34:14.314696 kernel: raid6: avx2x2 gen() 31189 MB/s Jan 13 20:34:14.331948 kernel: raid6: avx2x1 gen() 25122 MB/s Jan 13 20:34:14.331973 kernel: raid6: using algorithm avx2x2 gen() 31189 MB/s Jan 13 20:34:14.349930 kernel: raid6: .... xor() 19767 MB/s, rmw enabled Jan 13 20:34:14.349965 kernel: raid6: using avx2x2 recovery algorithm Jan 13 20:34:14.370691 kernel: xor: automatically using best checksumming function avx Jan 13 20:34:14.526696 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 13 20:34:14.539062 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 13 20:34:14.550772 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:34:14.562708 systemd-udevd[414]: Using default interface naming scheme 'v255'. Jan 13 20:34:14.567277 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:34:14.582791 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 13 20:34:14.596931 dracut-pre-trigger[423]: rd.md=0: removing MD RAID activation Jan 13 20:34:14.626573 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 13 20:34:14.641842 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 13 20:34:14.714985 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:34:14.723896 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 13 20:34:14.739728 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 13 20:34:14.743148 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 13 20:34:14.746455 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:34:14.749429 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 13 20:34:14.759287 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 13 20:34:14.785839 kernel: cryptd: max_cpu_qlen set to 1000 Jan 13 20:34:14.785860 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Jan 13 20:34:14.786025 kernel: libata version 3.00 loaded. Jan 13 20:34:14.786037 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 13 20:34:14.786049 kernel: GPT:9289727 != 19775487 Jan 13 20:34:14.786062 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 13 20:34:14.786080 kernel: GPT:9289727 != 19775487 Jan 13 20:34:14.786093 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 13 20:34:14.786107 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 13 20:34:14.786121 kernel: ahci 0000:00:1f.2: version 3.0 Jan 13 20:34:14.822361 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 13 20:34:14.822381 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Jan 13 20:34:14.822535 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 13 20:34:14.822692 kernel: AVX2 version of gcm_enc/dec engaged. Jan 13 20:34:14.822704 kernel: AES CTR mode by8 optimization enabled Jan 13 20:34:14.822720 kernel: scsi host0: ahci Jan 13 20:34:14.822874 kernel: scsi host1: ahci Jan 13 20:34:14.823019 kernel: scsi host2: ahci Jan 13 20:34:14.823166 kernel: scsi host3: ahci Jan 13 20:34:14.823308 kernel: scsi host4: ahci Jan 13 20:34:14.823460 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (462) Jan 13 20:34:14.823476 kernel: scsi host5: ahci Jan 13 20:34:14.823639 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Jan 13 20:34:14.823650 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Jan 13 20:34:14.823673 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Jan 13 20:34:14.823684 kernel: BTRFS: device fsid 5e7921ba-229a-48a0-bc77-9b30aaa34aeb devid 1 transid 36 /dev/vda3 scanned by (udev-worker) (460) Jan 13 20:34:14.823694 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Jan 13 20:34:14.823705 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Jan 13 20:34:14.823715 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Jan 13 20:34:14.763255 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 13 20:34:14.777533 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 13 20:34:14.805252 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 13 20:34:14.805631 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:34:14.810539 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:34:14.812876 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 13 20:34:14.813072 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:34:14.817241 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:34:14.830921 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:34:14.837103 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 13 20:34:14.847202 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 13 20:34:14.884760 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:34:14.896375 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 13 20:34:14.901577 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jan 13 20:34:14.902832 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 13 20:34:14.923811 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 13 20:34:14.925865 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:34:14.936205 disk-uuid[557]: Primary Header is updated. Jan 13 20:34:14.936205 disk-uuid[557]: Secondary Entries is updated. Jan 13 20:34:14.936205 disk-uuid[557]: Secondary Header is updated. Jan 13 20:34:14.940686 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 13 20:34:14.945678 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 13 20:34:14.946913 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:34:15.127687 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 13 20:34:15.127762 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 13 20:34:15.136227 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 13 20:34:15.136267 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 13 20:34:15.136282 kernel: ata3.00: applying bridge limits Jan 13 20:34:15.136675 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 13 20:34:15.137683 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 13 20:34:15.138678 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 13 20:34:15.138693 kernel: ata3.00: configured for UDMA/100 Jan 13 20:34:15.139676 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 13 20:34:15.187701 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 13 20:34:15.201441 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 13 20:34:15.201461 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 13 20:34:15.946691 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 13 20:34:15.946756 disk-uuid[558]: The operation has completed successfully. Jan 13 20:34:15.981816 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 13 20:34:15.981931 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 13 20:34:15.997820 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 13 20:34:16.000848 sh[593]: Success Jan 13 20:34:16.012687 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Jan 13 20:34:16.049692 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 13 20:34:16.061191 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 13 20:34:16.064139 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 13 20:34:16.076247 kernel: BTRFS info (device dm-0): first mount of filesystem 5e7921ba-229a-48a0-bc77-9b30aaa34aeb Jan 13 20:34:16.076287 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 13 20:34:16.076310 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 13 20:34:16.077289 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 13 20:34:16.078690 kernel: BTRFS info (device dm-0): using free space tree Jan 13 20:34:16.083650 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 13 20:34:16.084791 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 13 20:34:16.096785 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 13 20:34:16.098467 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 13 20:34:16.106984 kernel: BTRFS info (device vda6): first mount of filesystem 1066b41d-395d-4ccb-b5ae-be36ea0fc11e Jan 13 20:34:16.107017 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 13 20:34:16.107028 kernel: BTRFS info (device vda6): using free space tree Jan 13 20:34:16.109696 kernel: BTRFS info (device vda6): auto enabling async discard Jan 13 20:34:16.119532 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 13 20:34:16.121583 kernel: BTRFS info (device vda6): last unmount of filesystem 1066b41d-395d-4ccb-b5ae-be36ea0fc11e Jan 13 20:34:16.131227 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 13 20:34:16.139830 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 13 20:34:16.239425 ignition[683]: Ignition 2.20.0 Jan 13 20:34:16.239441 ignition[683]: Stage: fetch-offline Jan 13 20:34:16.239478 ignition[683]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:16.239487 ignition[683]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:16.239591 ignition[683]: parsed url from cmdline: "" Jan 13 20:34:16.239595 ignition[683]: no config URL provided Jan 13 20:34:16.239600 ignition[683]: reading system config file "/usr/lib/ignition/user.ign" Jan 13 20:34:16.239612 ignition[683]: no config at "/usr/lib/ignition/user.ign" Jan 13 20:34:16.239642 ignition[683]: op(1): [started] loading QEMU firmware config module Jan 13 20:34:16.239648 ignition[683]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 13 20:34:16.247033 ignition[683]: op(1): [finished] loading QEMU firmware config module Jan 13 20:34:16.249551 ignition[683]: parsing config with SHA512: cb4ff5f15bdd1d31af0d66ab808d45b93dd7b918aca3581b933efa97f7256dfa80d3df72efec06177c529c552f1abd3bc81c93770aee21ad99ec4da9a5a8b271 Jan 13 20:34:16.253037 unknown[683]: fetched base config from "system" Jan 13 20:34:16.253051 unknown[683]: fetched user config from "qemu" Jan 13 20:34:16.253733 ignition[683]: fetch-offline: fetch-offline passed Jan 13 20:34:16.253851 ignition[683]: Ignition finished successfully Jan 13 20:34:16.256333 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 13 20:34:16.263459 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 13 20:34:16.275831 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 13 20:34:16.301054 systemd-networkd[782]: lo: Link UP Jan 13 20:34:16.301070 systemd-networkd[782]: lo: Gained carrier Jan 13 20:34:16.302829 systemd-networkd[782]: Enumeration completed Jan 13 20:34:16.302952 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 13 20:34:16.303252 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:34:16.303257 systemd-networkd[782]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 13 20:34:16.304647 systemd-networkd[782]: eth0: Link UP Jan 13 20:34:16.304665 systemd-networkd[782]: eth0: Gained carrier Jan 13 20:34:16.304674 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:34:16.305051 systemd[1]: Reached target network.target - Network. Jan 13 20:34:16.306621 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 13 20:34:16.320809 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 13 20:34:16.328703 systemd-networkd[782]: eth0: DHCPv4 address 10.0.0.53/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 13 20:34:16.339374 ignition[785]: Ignition 2.20.0 Jan 13 20:34:16.339387 ignition[785]: Stage: kargs Jan 13 20:34:16.339553 ignition[785]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:16.339565 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:16.340221 ignition[785]: kargs: kargs passed Jan 13 20:34:16.343523 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 13 20:34:16.340267 ignition[785]: Ignition finished successfully Jan 13 20:34:16.355958 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 13 20:34:16.371648 ignition[794]: Ignition 2.20.0 Jan 13 20:34:16.371673 ignition[794]: Stage: disks Jan 13 20:34:16.371846 ignition[794]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:16.371857 ignition[794]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:16.374822 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 13 20:34:16.372519 ignition[794]: disks: disks passed Jan 13 20:34:16.376592 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 13 20:34:16.372562 ignition[794]: Ignition finished successfully Jan 13 20:34:16.378712 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 13 20:34:16.380651 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 13 20:34:16.382741 systemd[1]: Reached target sysinit.target - System Initialization. Jan 13 20:34:16.384518 systemd[1]: Reached target basic.target - Basic System. Jan 13 20:34:16.395980 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 13 20:34:16.413504 systemd-fsck[804]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jan 13 20:34:16.420878 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 13 20:34:16.438889 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 13 20:34:16.532698 kernel: EXT4-fs (vda9): mounted filesystem 84bcd1b2-5573-4e91-8fd5-f97782397085 r/w with ordered data mode. Quota mode: none. Jan 13 20:34:16.533365 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 13 20:34:16.534438 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 13 20:34:16.546787 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 13 20:34:16.548860 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 13 20:34:16.550451 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 13 20:34:16.550503 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 13 20:34:16.558950 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (812) Jan 13 20:34:16.558973 kernel: BTRFS info (device vda6): first mount of filesystem 1066b41d-395d-4ccb-b5ae-be36ea0fc11e Jan 13 20:34:16.550531 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 13 20:34:16.565615 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 13 20:34:16.565640 kernel: BTRFS info (device vda6): using free space tree Jan 13 20:34:16.565652 kernel: BTRFS info (device vda6): auto enabling async discard Jan 13 20:34:16.558765 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 13 20:34:16.563789 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 13 20:34:16.566693 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 13 20:34:16.605922 initrd-setup-root[836]: cut: /sysroot/etc/passwd: No such file or directory Jan 13 20:34:16.610083 initrd-setup-root[843]: cut: /sysroot/etc/group: No such file or directory Jan 13 20:34:16.614970 initrd-setup-root[850]: cut: /sysroot/etc/shadow: No such file or directory Jan 13 20:34:16.618542 initrd-setup-root[857]: cut: /sysroot/etc/gshadow: No such file or directory Jan 13 20:34:16.708795 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 13 20:34:16.726833 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 13 20:34:16.728243 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 13 20:34:16.739692 kernel: BTRFS info (device vda6): last unmount of filesystem 1066b41d-395d-4ccb-b5ae-be36ea0fc11e Jan 13 20:34:16.758157 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 13 20:34:16.769285 ignition[927]: INFO : Ignition 2.20.0 Jan 13 20:34:16.769285 ignition[927]: INFO : Stage: mount Jan 13 20:34:16.771212 ignition[927]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:16.771212 ignition[927]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:16.771212 ignition[927]: INFO : mount: mount passed Jan 13 20:34:16.771212 ignition[927]: INFO : Ignition finished successfully Jan 13 20:34:16.775694 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 13 20:34:16.788828 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 13 20:34:17.075491 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 13 20:34:17.089837 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 13 20:34:17.095677 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (939) Jan 13 20:34:17.097682 kernel: BTRFS info (device vda6): first mount of filesystem 1066b41d-395d-4ccb-b5ae-be36ea0fc11e Jan 13 20:34:17.097696 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 13 20:34:17.097706 kernel: BTRFS info (device vda6): using free space tree Jan 13 20:34:17.100691 kernel: BTRFS info (device vda6): auto enabling async discard Jan 13 20:34:17.102116 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 13 20:34:17.121804 ignition[956]: INFO : Ignition 2.20.0 Jan 13 20:34:17.121804 ignition[956]: INFO : Stage: files Jan 13 20:34:17.123504 ignition[956]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:17.123504 ignition[956]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:17.126067 ignition[956]: DEBUG : files: compiled without relabeling support, skipping Jan 13 20:34:17.127453 ignition[956]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 13 20:34:17.127453 ignition[956]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 13 20:34:17.131261 ignition[956]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 13 20:34:17.132720 ignition[956]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 13 20:34:17.132720 ignition[956]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 13 20:34:17.132063 unknown[956]: wrote ssh authorized keys file for user: core Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jan 13 20:34:17.137365 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Jan 13 20:34:17.515836 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Jan 13 20:34:17.784821 systemd-networkd[782]: eth0: Gained IPv6LL Jan 13 20:34:17.945157 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jan 13 20:34:17.945157 ignition[956]: INFO : files: op(8): [started] processing unit "containerd.service" Jan 13 20:34:17.949067 ignition[956]: INFO : files: op(8): op(9): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jan 13 20:34:17.951948 ignition[956]: INFO : files: op(8): op(9): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jan 13 20:34:17.951948 ignition[956]: INFO : files: op(8): [finished] processing unit "containerd.service" Jan 13 20:34:17.951948 ignition[956]: INFO : files: op(a): [started] processing unit "coreos-metadata.service" Jan 13 20:34:17.957084 ignition[956]: INFO : files: op(a): op(b): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 13 20:34:17.959253 ignition[956]: INFO : files: op(a): op(b): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 13 20:34:17.959253 ignition[956]: INFO : files: op(a): [finished] processing unit "coreos-metadata.service" Jan 13 20:34:17.962732 ignition[956]: INFO : files: op(c): [started] setting preset to disabled for "coreos-metadata.service" Jan 13 20:34:17.985920 ignition[956]: INFO : files: op(c): op(d): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 13 20:34:17.991524 ignition[956]: INFO : files: op(c): op(d): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 13 20:34:17.993206 ignition[956]: INFO : files: op(c): [finished] setting preset to disabled for "coreos-metadata.service" Jan 13 20:34:17.993206 ignition[956]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 13 20:34:17.993206 ignition[956]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 13 20:34:17.993206 ignition[956]: INFO : files: files passed Jan 13 20:34:17.993206 ignition[956]: INFO : Ignition finished successfully Jan 13 20:34:18.002162 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 13 20:34:18.014783 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 13 20:34:18.015942 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 13 20:34:18.023183 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 13 20:34:18.047681 initrd-setup-root-after-ignition[985]: grep: /sysroot/oem/oem-release: No such file or directory Jan 13 20:34:18.047680 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 13 20:34:18.051300 initrd-setup-root-after-ignition[987]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:34:18.051300 initrd-setup-root-after-ignition[987]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:34:18.054358 initrd-setup-root-after-ignition[991]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:34:18.053497 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 13 20:34:18.057078 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 13 20:34:18.086804 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 13 20:34:18.111792 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 13 20:34:18.111944 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 13 20:34:18.112705 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 13 20:34:18.115591 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 13 20:34:18.117712 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 13 20:34:18.128347 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 13 20:34:18.147786 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 13 20:34:18.152826 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 13 20:34:18.164463 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:34:18.165013 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:34:18.167196 systemd[1]: Stopped target timers.target - Timer Units. Jan 13 20:34:18.167507 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 13 20:34:18.167640 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 13 20:34:18.172631 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 13 20:34:18.173193 systemd[1]: Stopped target basic.target - Basic System. Jan 13 20:34:18.173530 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 13 20:34:18.174033 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 13 20:34:18.179442 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 13 20:34:18.181644 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 13 20:34:18.184295 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 13 20:34:18.185061 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 13 20:34:18.188321 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 13 20:34:18.190049 systemd[1]: Stopped target swap.target - Swaps. Jan 13 20:34:18.191730 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 13 20:34:18.191865 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 13 20:34:18.193579 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:34:18.194091 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:34:18.197956 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 13 20:34:18.199811 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:34:18.200604 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 13 20:34:18.200788 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 13 20:34:18.206016 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 13 20:34:18.206189 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 13 20:34:18.206831 systemd[1]: Stopped target paths.target - Path Units. Jan 13 20:34:18.207191 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 13 20:34:18.211751 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:34:18.212304 systemd[1]: Stopped target slices.target - Slice Units. Jan 13 20:34:18.212619 systemd[1]: Stopped target sockets.target - Socket Units. Jan 13 20:34:18.216478 systemd[1]: iscsid.socket: Deactivated successfully. Jan 13 20:34:18.216581 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 13 20:34:18.218224 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 13 20:34:18.218337 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 13 20:34:18.220105 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 13 20:34:18.220233 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 13 20:34:18.221782 systemd[1]: ignition-files.service: Deactivated successfully. Jan 13 20:34:18.221900 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 13 20:34:18.232828 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 13 20:34:18.234851 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 13 20:34:18.235298 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 13 20:34:18.235432 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:34:18.237166 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 13 20:34:18.237303 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 13 20:34:18.244132 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 13 20:34:18.244249 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 13 20:34:18.248581 ignition[1011]: INFO : Ignition 2.20.0 Jan 13 20:34:18.248581 ignition[1011]: INFO : Stage: umount Jan 13 20:34:18.250154 ignition[1011]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:34:18.250154 ignition[1011]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 13 20:34:18.250154 ignition[1011]: INFO : umount: umount passed Jan 13 20:34:18.250154 ignition[1011]: INFO : Ignition finished successfully Jan 13 20:34:18.251452 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 13 20:34:18.251597 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 13 20:34:18.254623 systemd[1]: Stopped target network.target - Network. Jan 13 20:34:18.255376 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 13 20:34:18.255445 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 13 20:34:18.255905 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 13 20:34:18.255958 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 13 20:34:18.259012 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 13 20:34:18.259066 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 13 20:34:18.260564 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 13 20:34:18.260623 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 13 20:34:18.262525 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 13 20:34:18.264476 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 13 20:34:18.268693 systemd-networkd[782]: eth0: DHCPv6 lease lost Jan 13 20:34:18.271419 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 13 20:34:18.271567 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 13 20:34:18.272689 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 13 20:34:18.272821 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 13 20:34:18.276010 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 13 20:34:18.276066 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:34:18.280862 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 13 20:34:18.281331 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 13 20:34:18.281391 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 13 20:34:18.283207 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 13 20:34:18.283276 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:34:18.283524 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 13 20:34:18.283575 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 13 20:34:18.284030 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 13 20:34:18.284083 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:34:18.284462 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:34:18.301430 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 13 20:34:18.301579 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 13 20:34:18.307549 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 13 20:34:18.307744 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:34:18.310136 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 13 20:34:18.310194 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 13 20:34:18.312349 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 13 20:34:18.312399 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:34:18.314401 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 13 20:34:18.314459 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 13 20:34:18.316795 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 13 20:34:18.316851 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 13 20:34:18.318818 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 13 20:34:18.318874 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:34:18.327795 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 13 20:34:18.328981 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 13 20:34:18.329036 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:34:18.331591 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 13 20:34:18.331643 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 13 20:34:18.333997 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 13 20:34:18.334048 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:34:18.336815 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 13 20:34:18.336863 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:34:18.339154 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 13 20:34:18.339697 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 13 20:34:18.339801 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 13 20:34:18.544276 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 13 20:34:18.544408 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 13 20:34:18.546371 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 13 20:34:18.546893 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 13 20:34:18.546956 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 13 20:34:18.556804 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 13 20:34:18.565467 systemd[1]: Switching root. Jan 13 20:34:18.592773 systemd-journald[194]: Journal stopped Jan 13 20:34:19.667854 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Jan 13 20:34:19.667943 kernel: SELinux: policy capability network_peer_controls=1 Jan 13 20:34:19.667968 kernel: SELinux: policy capability open_perms=1 Jan 13 20:34:19.667990 kernel: SELinux: policy capability extended_socket_class=1 Jan 13 20:34:19.668005 kernel: SELinux: policy capability always_check_network=0 Jan 13 20:34:19.668020 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 13 20:34:19.668035 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 13 20:34:19.668050 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 13 20:34:19.668065 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 13 20:34:19.668081 kernel: audit: type=1403 audit(1736800458.976:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 13 20:34:19.668102 systemd[1]: Successfully loaded SELinux policy in 38.554ms. Jan 13 20:34:19.668126 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.164ms. Jan 13 20:34:19.668146 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 13 20:34:19.668162 systemd[1]: Detected virtualization kvm. Jan 13 20:34:19.668177 systemd[1]: Detected architecture x86-64. Jan 13 20:34:19.668195 systemd[1]: Detected first boot. Jan 13 20:34:19.668219 systemd[1]: Initializing machine ID from VM UUID. Jan 13 20:34:19.668240 zram_generator::config[1077]: No configuration found. Jan 13 20:34:19.668263 systemd[1]: Populated /etc with preset unit settings. Jan 13 20:34:19.668282 systemd[1]: Queued start job for default target multi-user.target. Jan 13 20:34:19.668297 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 13 20:34:19.668314 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 13 20:34:19.668331 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 13 20:34:19.668350 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 13 20:34:19.668366 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 13 20:34:19.668382 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 13 20:34:19.668397 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 13 20:34:19.668414 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 13 20:34:19.668435 systemd[1]: Created slice user.slice - User and Session Slice. Jan 13 20:34:19.668452 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:34:19.668468 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:34:19.668484 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 13 20:34:19.668502 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 13 20:34:19.668518 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 13 20:34:19.668533 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 13 20:34:19.668549 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 13 20:34:19.668564 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:34:19.668579 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 13 20:34:19.668594 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:34:19.668610 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 13 20:34:19.668628 systemd[1]: Reached target slices.target - Slice Units. Jan 13 20:34:19.668647 systemd[1]: Reached target swap.target - Swaps. Jan 13 20:34:19.668683 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 13 20:34:19.668702 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 13 20:34:19.668718 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 13 20:34:19.668735 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 13 20:34:19.668751 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:34:19.668767 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 13 20:34:19.668783 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:34:19.668799 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 13 20:34:19.668820 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 13 20:34:19.668836 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 13 20:34:19.668852 systemd[1]: Mounting media.mount - External Media Directory... Jan 13 20:34:19.668869 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:19.668886 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 13 20:34:19.668902 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 13 20:34:19.668919 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 13 20:34:19.668935 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 13 20:34:19.668954 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:34:19.668970 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 13 20:34:19.668987 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 13 20:34:19.669005 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:34:19.669021 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 13 20:34:19.669037 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:34:19.669053 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 13 20:34:19.669069 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:34:19.669090 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 13 20:34:19.669106 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Jan 13 20:34:19.669122 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Jan 13 20:34:19.669138 kernel: fuse: init (API version 7.39) Jan 13 20:34:19.669154 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 13 20:34:19.669169 kernel: loop: module loaded Jan 13 20:34:19.669184 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 13 20:34:19.669209 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 13 20:34:19.669254 systemd-journald[1155]: Collecting audit messages is disabled. Jan 13 20:34:19.669290 kernel: ACPI: bus type drm_connector registered Jan 13 20:34:19.669306 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 13 20:34:19.669322 systemd-journald[1155]: Journal started Jan 13 20:34:19.669349 systemd-journald[1155]: Runtime Journal (/run/log/journal/6ff90c8a49e14c22b1fa5321ffaba61a) is 6.0M, max 48.4M, 42.3M free. Jan 13 20:34:19.673919 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 13 20:34:19.680447 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:19.680490 systemd[1]: Started systemd-journald.service - Journal Service. Jan 13 20:34:19.683301 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 13 20:34:19.684542 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 13 20:34:19.685879 systemd[1]: Mounted media.mount - External Media Directory. Jan 13 20:34:19.687051 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 13 20:34:19.688296 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 13 20:34:19.689540 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 13 20:34:19.690926 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 13 20:34:19.692519 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:34:19.694123 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 13 20:34:19.694379 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 13 20:34:19.695946 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:34:19.696175 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:34:19.697648 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 13 20:34:19.697892 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 13 20:34:19.699695 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:34:19.699920 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:34:19.701602 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 13 20:34:19.701843 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 13 20:34:19.703487 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:34:19.703779 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:34:19.705431 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 13 20:34:19.707003 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 13 20:34:19.708812 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 13 20:34:19.722869 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 13 20:34:19.738736 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 13 20:34:19.741387 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 13 20:34:19.742941 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 13 20:34:19.745526 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 13 20:34:19.750286 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 13 20:34:19.751771 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 13 20:34:19.754027 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 13 20:34:19.755337 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 13 20:34:19.757906 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:34:19.768845 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 13 20:34:19.772242 systemd-journald[1155]: Time spent on flushing to /var/log/journal/6ff90c8a49e14c22b1fa5321ffaba61a is 18.927ms for 925 entries. Jan 13 20:34:19.772242 systemd-journald[1155]: System Journal (/var/log/journal/6ff90c8a49e14c22b1fa5321ffaba61a) is 8.0M, max 195.6M, 187.6M free. Jan 13 20:34:19.798123 systemd-journald[1155]: Received client request to flush runtime journal. Jan 13 20:34:19.773434 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:34:19.775012 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 13 20:34:19.776436 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 13 20:34:19.787123 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 13 20:34:19.791457 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:34:19.798261 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 13 20:34:19.807856 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 13 20:34:19.809799 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 13 20:34:19.817345 udevadm[1225]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 13 20:34:19.817417 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Jan 13 20:34:19.817435 systemd-tmpfiles[1212]: ACLs are not supported, ignoring. Jan 13 20:34:19.825464 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 13 20:34:19.839916 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 13 20:34:19.869082 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 13 20:34:19.877784 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 13 20:34:19.902447 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Jan 13 20:34:19.902477 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Jan 13 20:34:19.910162 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:34:20.297837 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 13 20:34:20.314941 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:34:20.340952 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Jan 13 20:34:20.356671 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:34:20.364839 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 13 20:34:20.383768 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 13 20:34:20.393630 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Jan 13 20:34:20.425690 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1252) Jan 13 20:34:20.454963 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 13 20:34:20.462853 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 13 20:34:20.469327 kernel: ACPI: button: Power Button [PWRF] Jan 13 20:34:20.473480 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 13 20:34:20.475134 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Jan 13 20:34:20.475344 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 13 20:34:20.490627 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 13 20:34:20.498708 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Jan 13 20:34:20.507796 kernel: mousedev: PS/2 mouse device common for all mice Jan 13 20:34:20.521927 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:34:20.540985 systemd-networkd[1243]: lo: Link UP Jan 13 20:34:20.540993 systemd-networkd[1243]: lo: Gained carrier Jan 13 20:34:20.543005 systemd-networkd[1243]: Enumeration completed Jan 13 20:34:20.543499 systemd-networkd[1243]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:34:20.543552 systemd-networkd[1243]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 13 20:34:20.544247 systemd-networkd[1243]: eth0: Link UP Jan 13 20:34:20.544311 systemd-networkd[1243]: eth0: Gained carrier Jan 13 20:34:20.544355 systemd-networkd[1243]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:34:20.546159 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 13 20:34:20.557830 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 13 20:34:20.561707 systemd-networkd[1243]: eth0: DHCPv4 address 10.0.0.53/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 13 20:34:20.608506 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:34:20.637833 kernel: kvm_amd: TSC scaling supported Jan 13 20:34:20.637889 kernel: kvm_amd: Nested Virtualization enabled Jan 13 20:34:20.637903 kernel: kvm_amd: Nested Paging enabled Jan 13 20:34:20.638799 kernel: kvm_amd: LBR virtualization supported Jan 13 20:34:20.638813 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 13 20:34:20.639792 kernel: kvm_amd: Virtual GIF supported Jan 13 20:34:20.658679 kernel: EDAC MC: Ver: 3.0.0 Jan 13 20:34:20.690156 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 13 20:34:20.702800 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 13 20:34:20.710424 lvm[1285]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 13 20:34:20.740220 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 13 20:34:20.742222 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:34:20.756826 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 13 20:34:20.762020 lvm[1288]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 13 20:34:20.795583 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 13 20:34:20.797788 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 13 20:34:20.799479 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 13 20:34:20.799515 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 13 20:34:20.800933 systemd[1]: Reached target machines.target - Containers. Jan 13 20:34:20.803936 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 13 20:34:20.831899 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 13 20:34:20.835073 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 13 20:34:20.836454 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:34:20.837749 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 13 20:34:20.841088 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 13 20:34:20.845959 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 13 20:34:20.848711 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 13 20:34:20.859081 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 13 20:34:20.864046 kernel: loop0: detected capacity change from 0 to 138184 Jan 13 20:34:20.874301 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 13 20:34:20.875338 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 13 20:34:20.886701 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 13 20:34:20.912675 kernel: loop1: detected capacity change from 0 to 140992 Jan 13 20:34:20.944692 kernel: loop2: detected capacity change from 0 to 211296 Jan 13 20:34:20.975682 kernel: loop3: detected capacity change from 0 to 138184 Jan 13 20:34:20.986673 kernel: loop4: detected capacity change from 0 to 140992 Jan 13 20:34:20.995957 kernel: loop5: detected capacity change from 0 to 211296 Jan 13 20:34:21.000846 (sd-merge)[1309]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Jan 13 20:34:21.001422 (sd-merge)[1309]: Merged extensions into '/usr'. Jan 13 20:34:21.005918 systemd[1]: Reloading requested from client PID 1296 ('systemd-sysext') (unit systemd-sysext.service)... Jan 13 20:34:21.005936 systemd[1]: Reloading... Jan 13 20:34:21.049814 zram_generator::config[1335]: No configuration found. Jan 13 20:34:21.087899 ldconfig[1292]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 13 20:34:21.190954 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:34:21.255396 systemd[1]: Reloading finished in 248 ms. Jan 13 20:34:21.275317 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 13 20:34:21.276934 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 13 20:34:21.296935 systemd[1]: Starting ensure-sysext.service... Jan 13 20:34:21.299684 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 13 20:34:21.303053 systemd[1]: Reloading requested from client PID 1381 ('systemctl') (unit ensure-sysext.service)... Jan 13 20:34:21.303067 systemd[1]: Reloading... Jan 13 20:34:21.327506 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 13 20:34:21.327991 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 13 20:34:21.330245 systemd-tmpfiles[1382]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 13 20:34:21.330675 systemd-tmpfiles[1382]: ACLs are not supported, ignoring. Jan 13 20:34:21.330777 systemd-tmpfiles[1382]: ACLs are not supported, ignoring. Jan 13 20:34:21.336635 systemd-tmpfiles[1382]: Detected autofs mount point /boot during canonicalization of boot. Jan 13 20:34:21.336774 systemd-tmpfiles[1382]: Skipping /boot Jan 13 20:34:21.350013 zram_generator::config[1413]: No configuration found. Jan 13 20:34:21.349535 systemd-tmpfiles[1382]: Detected autofs mount point /boot during canonicalization of boot. Jan 13 20:34:21.350247 systemd-tmpfiles[1382]: Skipping /boot Jan 13 20:34:21.472091 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:34:21.536866 systemd[1]: Reloading finished in 233 ms. Jan 13 20:34:21.553559 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:34:21.570876 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 13 20:34:21.573953 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 13 20:34:21.576874 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 13 20:34:21.580834 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 13 20:34:21.586773 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 13 20:34:21.596268 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:21.596670 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:34:21.600793 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:34:21.604945 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:34:21.609005 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:34:21.614206 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:34:21.614355 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:21.615410 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:34:21.616098 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:34:21.625194 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:34:21.625427 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:34:21.627800 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 13 20:34:21.630062 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:34:21.630352 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:34:21.638458 augenrules[1489]: No rules Jan 13 20:34:21.642237 systemd[1]: audit-rules.service: Deactivated successfully. Jan 13 20:34:21.642850 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 13 20:34:21.649312 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 13 20:34:21.653194 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:21.659875 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 13 20:34:21.661012 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:34:21.662587 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:34:21.666456 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 13 20:34:21.674006 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:34:21.691731 augenrules[1501]: /sbin/augenrules: No change Jan 13 20:34:21.689012 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:34:21.691488 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:34:21.696375 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 13 20:34:21.697596 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 13 20:34:21.699706 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 13 20:34:21.701868 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:34:21.702091 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:34:21.703467 systemd-resolved[1458]: Positive Trust Anchors: Jan 13 20:34:21.703761 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 13 20:34:21.703926 systemd-resolved[1458]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 13 20:34:21.703972 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 13 20:34:21.703973 systemd-resolved[1458]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 13 20:34:21.706094 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:34:21.706316 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:34:21.706396 augenrules[1528]: No rules Jan 13 20:34:21.708043 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:34:21.708368 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:34:21.709332 systemd-resolved[1458]: Defaulting to hostname 'linux'. Jan 13 20:34:21.710088 systemd[1]: audit-rules.service: Deactivated successfully. Jan 13 20:34:21.710465 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 13 20:34:21.712827 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 13 20:34:21.716625 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 13 20:34:21.718376 systemd[1]: Finished ensure-sysext.service. Jan 13 20:34:21.728781 systemd[1]: Reached target network.target - Network. Jan 13 20:34:21.729939 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:34:21.731399 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 13 20:34:21.731482 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 13 20:34:21.742936 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 13 20:34:21.744229 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 13 20:34:21.808610 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 13 20:34:21.810052 systemd[1]: Reached target sysinit.target - System Initialization. Jan 13 20:34:21.811313 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 13 20:34:22.380319 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 13 20:34:22.380332 systemd-resolved[1458]: Clock change detected. Flushing caches. Jan 13 20:34:22.380358 systemd-timesyncd[1545]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 13 20:34:22.381574 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 13 20:34:22.382172 systemd-timesyncd[1545]: Initial clock synchronization to Mon 2025-01-13 20:34:22.380262 UTC. Jan 13 20:34:22.382818 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 13 20:34:22.382847 systemd[1]: Reached target paths.target - Path Units. Jan 13 20:34:22.383741 systemd[1]: Reached target time-set.target - System Time Set. Jan 13 20:34:22.384981 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 13 20:34:22.386238 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 13 20:34:22.387664 systemd[1]: Reached target timers.target - Timer Units. Jan 13 20:34:22.389189 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 13 20:34:22.392616 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 13 20:34:22.395380 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 13 20:34:22.400419 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 13 20:34:22.401876 systemd[1]: Reached target sockets.target - Socket Units. Jan 13 20:34:22.403077 systemd[1]: Reached target basic.target - Basic System. Jan 13 20:34:22.404466 systemd[1]: System is tainted: cgroupsv1 Jan 13 20:34:22.404522 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 13 20:34:22.404551 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 13 20:34:22.406590 systemd[1]: Starting containerd.service - containerd container runtime... Jan 13 20:34:22.409456 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 13 20:34:22.412152 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 13 20:34:22.416534 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 13 20:34:22.417677 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 13 20:34:22.421463 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 13 20:34:22.422453 jq[1551]: false Jan 13 20:34:22.425447 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 13 20:34:22.432014 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 13 20:34:22.437665 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 13 20:34:22.439532 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 13 20:34:22.440048 dbus-daemon[1550]: [system] SELinux support is enabled Jan 13 20:34:22.443408 systemd[1]: Starting update-engine.service - Update Engine... Jan 13 20:34:22.448590 extend-filesystems[1553]: Found loop3 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found loop4 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found loop5 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found sr0 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda1 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda2 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda3 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found usr Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda4 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda6 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda7 Jan 13 20:34:22.448590 extend-filesystems[1553]: Found vda9 Jan 13 20:34:22.448590 extend-filesystems[1553]: Checking size of /dev/vda9 Jan 13 20:34:22.487782 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Jan 13 20:34:22.448157 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 13 20:34:22.487898 extend-filesystems[1553]: Resized partition /dev/vda9 Jan 13 20:34:22.450134 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 13 20:34:22.495300 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1254) Jan 13 20:34:22.495358 extend-filesystems[1577]: resize2fs 1.47.1 (20-May-2024) Jan 13 20:34:22.462762 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 13 20:34:22.496562 update_engine[1563]: I20250113 20:34:22.493624 1563 main.cc:92] Flatcar Update Engine starting Jan 13 20:34:22.496806 jq[1569]: true Jan 13 20:34:22.463103 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 13 20:34:22.464470 systemd[1]: motdgen.service: Deactivated successfully. Jan 13 20:34:22.464778 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 13 20:34:22.466913 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 13 20:34:22.467205 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 13 20:34:22.475195 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 13 20:34:22.475235 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 13 20:34:22.498493 update_engine[1563]: I20250113 20:34:22.498198 1563 update_check_scheduler.cc:74] Next update check in 11m43s Jan 13 20:34:22.475930 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 13 20:34:22.475947 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 13 20:34:22.498529 (ntainerd)[1584]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 13 20:34:22.500296 jq[1583]: true Jan 13 20:34:22.503881 systemd[1]: Started update-engine.service - Update Engine. Jan 13 20:34:22.509406 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 13 20:34:22.518303 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Jan 13 20:34:22.519245 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 13 20:34:22.545031 extend-filesystems[1577]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 13 20:34:22.545031 extend-filesystems[1577]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 13 20:34:22.545031 extend-filesystems[1577]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Jan 13 20:34:22.554000 extend-filesystems[1553]: Resized filesystem in /dev/vda9 Jan 13 20:34:22.546219 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 13 20:34:22.546596 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 13 20:34:22.560110 systemd-logind[1560]: Watching system buttons on /dev/input/event1 (Power Button) Jan 13 20:34:22.560137 systemd-logind[1560]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 13 20:34:22.561976 systemd-logind[1560]: New seat seat0. Jan 13 20:34:22.563771 systemd[1]: Started systemd-logind.service - User Login Management. Jan 13 20:34:22.570959 locksmithd[1592]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 13 20:34:22.572018 bash[1606]: Updated "/home/core/.ssh/authorized_keys" Jan 13 20:34:22.574093 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 13 20:34:22.578310 sshd_keygen[1573]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 13 20:34:22.577081 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 13 20:34:22.600903 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 13 20:34:22.610467 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 13 20:34:22.620328 systemd[1]: issuegen.service: Deactivated successfully. Jan 13 20:34:22.620680 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 13 20:34:22.630472 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 13 20:34:22.641608 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 13 20:34:22.649579 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 13 20:34:22.652094 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 13 20:34:22.653553 systemd[1]: Reached target getty.target - Login Prompts. Jan 13 20:34:22.710304 containerd[1584]: time="2025-01-13T20:34:22.710090560Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Jan 13 20:34:22.730966 containerd[1584]: time="2025-01-13T20:34:22.730923209Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.732680 containerd[1584]: time="2025-01-13T20:34:22.732634718Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:34:22.732680 containerd[1584]: time="2025-01-13T20:34:22.732661789Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 13 20:34:22.732742 containerd[1584]: time="2025-01-13T20:34:22.732680314Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 13 20:34:22.732923 containerd[1584]: time="2025-01-13T20:34:22.732887012Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 13 20:34:22.732923 containerd[1584]: time="2025-01-13T20:34:22.732910896Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733039 containerd[1584]: time="2025-01-13T20:34:22.732996908Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733039 containerd[1584]: time="2025-01-13T20:34:22.733018819Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733322 containerd[1584]: time="2025-01-13T20:34:22.733286100Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733322 containerd[1584]: time="2025-01-13T20:34:22.733305957Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733396 containerd[1584]: time="2025-01-13T20:34:22.733322208Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733396 containerd[1584]: time="2025-01-13T20:34:22.733335252Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733495 containerd[1584]: time="2025-01-13T20:34:22.733462571Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733757 containerd[1584]: time="2025-01-13T20:34:22.733713542Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733943 containerd[1584]: time="2025-01-13T20:34:22.733908287Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:34:22.733943 containerd[1584]: time="2025-01-13T20:34:22.733927513Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 13 20:34:22.734093 containerd[1584]: time="2025-01-13T20:34:22.734059751Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 13 20:34:22.734160 containerd[1584]: time="2025-01-13T20:34:22.734137897Z" level=info msg="metadata content store policy set" policy=shared Jan 13 20:34:22.740309 containerd[1584]: time="2025-01-13T20:34:22.740245471Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 13 20:34:22.740353 containerd[1584]: time="2025-01-13T20:34:22.740320011Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 13 20:34:22.740353 containerd[1584]: time="2025-01-13T20:34:22.740339337Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 13 20:34:22.740390 containerd[1584]: time="2025-01-13T20:34:22.740355428Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 13 20:34:22.740390 containerd[1584]: time="2025-01-13T20:34:22.740370346Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 13 20:34:22.740544 containerd[1584]: time="2025-01-13T20:34:22.740518533Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 13 20:34:22.740935 containerd[1584]: time="2025-01-13T20:34:22.740902634Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 13 20:34:22.741069 containerd[1584]: time="2025-01-13T20:34:22.741042155Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 13 20:34:22.741069 containerd[1584]: time="2025-01-13T20:34:22.741064277Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 13 20:34:22.741130 containerd[1584]: time="2025-01-13T20:34:22.741081609Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 13 20:34:22.741130 containerd[1584]: time="2025-01-13T20:34:22.741099984Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741130 containerd[1584]: time="2025-01-13T20:34:22.741116525Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741130882Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741147102Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741164555Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741179884Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741195012Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741216 containerd[1584]: time="2025-01-13T20:34:22.741208778Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741232542Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741249374Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741264502Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741297134Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741314646Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741331097Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741345444Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741389 containerd[1584]: time="2025-01-13T20:34:22.741390709Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741407931Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741427057Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741442055Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741462864Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741477462Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741494854Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741517757Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741533337Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741546571Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 13 20:34:22.741596 containerd[1584]: time="2025-01-13T20:34:22.741597587Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741617054Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741630369Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741645156Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741657399Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741671977Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741684650Z" level=info msg="NRI interface is disabled by configuration." Jan 13 20:34:22.741855 containerd[1584]: time="2025-01-13T20:34:22.741698045Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 13 20:34:22.742100 containerd[1584]: time="2025-01-13T20:34:22.742040147Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 13 20:34:22.742100 containerd[1584]: time="2025-01-13T20:34:22.742097004Z" level=info msg="Connect containerd service" Jan 13 20:34:22.742299 containerd[1584]: time="2025-01-13T20:34:22.742131639Z" level=info msg="using legacy CRI server" Jan 13 20:34:22.742299 containerd[1584]: time="2025-01-13T20:34:22.742140375Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 13 20:34:22.742299 containerd[1584]: time="2025-01-13T20:34:22.742240883Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 13 20:34:22.742847 containerd[1584]: time="2025-01-13T20:34:22.742813688Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 13 20:34:22.742999 containerd[1584]: time="2025-01-13T20:34:22.742953820Z" level=info msg="Start subscribing containerd event" Jan 13 20:34:22.743038 containerd[1584]: time="2025-01-13T20:34:22.743004596Z" level=info msg="Start recovering state" Jan 13 20:34:22.743178 containerd[1584]: time="2025-01-13T20:34:22.743151722Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 13 20:34:22.743214 containerd[1584]: time="2025-01-13T20:34:22.743151972Z" level=info msg="Start event monitor" Jan 13 20:34:22.743214 containerd[1584]: time="2025-01-13T20:34:22.743202036Z" level=info msg="Start snapshots syncer" Jan 13 20:34:22.743214 containerd[1584]: time="2025-01-13T20:34:22.743211774Z" level=info msg="Start cni network conf syncer for default" Jan 13 20:34:22.743307 containerd[1584]: time="2025-01-13T20:34:22.743218707Z" level=info msg="Start streaming server" Jan 13 20:34:22.743307 containerd[1584]: time="2025-01-13T20:34:22.743220330Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 13 20:34:22.743396 containerd[1584]: time="2025-01-13T20:34:22.743377435Z" level=info msg="containerd successfully booted in 0.034567s" Jan 13 20:34:22.743511 systemd[1]: Started containerd.service - containerd container runtime. Jan 13 20:34:23.024497 systemd-networkd[1243]: eth0: Gained IPv6LL Jan 13 20:34:23.028183 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 13 20:34:23.030338 systemd[1]: Reached target network-online.target - Network is Online. Jan 13 20:34:23.042490 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 13 20:34:23.045886 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:34:23.049188 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 13 20:34:23.074832 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 13 20:34:23.075311 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 13 20:34:23.077951 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 13 20:34:23.078593 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 13 20:34:23.678746 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:34:23.680378 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 13 20:34:23.681651 systemd[1]: Startup finished in 6.075s (kernel) + 4.173s (userspace) = 10.249s. Jan 13 20:34:23.702773 (kubelet)[1676]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 13 20:34:24.185116 kubelet[1676]: E0113 20:34:24.185029 1676 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 13 20:34:24.189738 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 13 20:34:24.190015 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 13 20:34:31.591977 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 13 20:34:31.603692 systemd[1]: Started sshd@0-10.0.0.53:22-10.0.0.1:46592.service - OpenSSH per-connection server daemon (10.0.0.1:46592). Jan 13 20:34:31.644356 sshd[1690]: Accepted publickey for core from 10.0.0.1 port 46592 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:31.646250 sshd-session[1690]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:31.655469 systemd-logind[1560]: New session 1 of user core. Jan 13 20:34:31.656749 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 13 20:34:31.665480 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 13 20:34:31.677335 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 13 20:34:31.687607 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 13 20:34:31.690740 (systemd)[1695]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 13 20:34:31.806306 systemd[1695]: Queued start job for default target default.target. Jan 13 20:34:31.806879 systemd[1695]: Created slice app.slice - User Application Slice. Jan 13 20:34:31.806903 systemd[1695]: Reached target paths.target - Paths. Jan 13 20:34:31.806920 systemd[1695]: Reached target timers.target - Timers. Jan 13 20:34:31.817453 systemd[1695]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 13 20:34:31.826568 systemd[1695]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 13 20:34:31.826657 systemd[1695]: Reached target sockets.target - Sockets. Jan 13 20:34:31.826675 systemd[1695]: Reached target basic.target - Basic System. Jan 13 20:34:31.826731 systemd[1695]: Reached target default.target - Main User Target. Jan 13 20:34:31.826771 systemd[1695]: Startup finished in 129ms. Jan 13 20:34:31.827333 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 13 20:34:31.830615 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 13 20:34:31.898771 systemd[1]: Started sshd@1-10.0.0.53:22-10.0.0.1:46600.service - OpenSSH per-connection server daemon (10.0.0.1:46600). Jan 13 20:34:31.952829 sshd[1708]: Accepted publickey for core from 10.0.0.1 port 46600 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:31.955200 sshd-session[1708]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:31.971292 systemd-logind[1560]: New session 2 of user core. Jan 13 20:34:31.984564 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 13 20:34:32.063581 sshd[1711]: Connection closed by 10.0.0.1 port 46600 Jan 13 20:34:32.062105 sshd-session[1708]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:32.078619 systemd[1]: Started sshd@2-10.0.0.53:22-10.0.0.1:46614.service - OpenSSH per-connection server daemon (10.0.0.1:46614). Jan 13 20:34:32.079287 systemd[1]: sshd@1-10.0.0.53:22-10.0.0.1:46600.service: Deactivated successfully. Jan 13 20:34:32.081411 systemd[1]: session-2.scope: Deactivated successfully. Jan 13 20:34:32.087027 systemd-logind[1560]: Session 2 logged out. Waiting for processes to exit. Jan 13 20:34:32.096391 systemd-logind[1560]: Removed session 2. Jan 13 20:34:32.157983 sshd[1714]: Accepted publickey for core from 10.0.0.1 port 46614 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:32.160523 sshd-session[1714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:32.172905 systemd-logind[1560]: New session 3 of user core. Jan 13 20:34:32.182769 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 13 20:34:32.256575 sshd[1719]: Connection closed by 10.0.0.1 port 46614 Jan 13 20:34:32.254250 sshd-session[1714]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:32.262859 systemd[1]: Started sshd@3-10.0.0.53:22-10.0.0.1:46622.service - OpenSSH per-connection server daemon (10.0.0.1:46622). Jan 13 20:34:32.263662 systemd[1]: sshd@2-10.0.0.53:22-10.0.0.1:46614.service: Deactivated successfully. Jan 13 20:34:32.268820 systemd-logind[1560]: Session 3 logged out. Waiting for processes to exit. Jan 13 20:34:32.269993 systemd[1]: session-3.scope: Deactivated successfully. Jan 13 20:34:32.271005 systemd-logind[1560]: Removed session 3. Jan 13 20:34:32.321488 sshd[1721]: Accepted publickey for core from 10.0.0.1 port 46622 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:32.323304 sshd-session[1721]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:32.339261 systemd-logind[1560]: New session 4 of user core. Jan 13 20:34:32.349855 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 13 20:34:32.409030 sshd[1727]: Connection closed by 10.0.0.1 port 46622 Jan 13 20:34:32.409352 sshd-session[1721]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:32.434641 systemd[1]: Started sshd@4-10.0.0.53:22-10.0.0.1:46626.service - OpenSSH per-connection server daemon (10.0.0.1:46626). Jan 13 20:34:32.435209 systemd[1]: sshd@3-10.0.0.53:22-10.0.0.1:46622.service: Deactivated successfully. Jan 13 20:34:32.437012 systemd[1]: session-4.scope: Deactivated successfully. Jan 13 20:34:32.440530 systemd-logind[1560]: Session 4 logged out. Waiting for processes to exit. Jan 13 20:34:32.441648 systemd-logind[1560]: Removed session 4. Jan 13 20:34:32.469589 sshd[1730]: Accepted publickey for core from 10.0.0.1 port 46626 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:32.471077 sshd-session[1730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:32.475059 systemd-logind[1560]: New session 5 of user core. Jan 13 20:34:32.485540 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 13 20:34:32.543141 sudo[1736]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 13 20:34:32.543478 sudo[1736]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 13 20:34:32.565553 sudo[1736]: pam_unix(sudo:session): session closed for user root Jan 13 20:34:32.567091 sshd[1735]: Connection closed by 10.0.0.1 port 46626 Jan 13 20:34:32.567570 sshd-session[1730]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:32.581523 systemd[1]: Started sshd@5-10.0.0.53:22-10.0.0.1:46640.service - OpenSSH per-connection server daemon (10.0.0.1:46640). Jan 13 20:34:32.582094 systemd[1]: sshd@4-10.0.0.53:22-10.0.0.1:46626.service: Deactivated successfully. Jan 13 20:34:32.584706 systemd-logind[1560]: Session 5 logged out. Waiting for processes to exit. Jan 13 20:34:32.586296 systemd[1]: session-5.scope: Deactivated successfully. Jan 13 20:34:32.587244 systemd-logind[1560]: Removed session 5. Jan 13 20:34:32.613802 sshd[1738]: Accepted publickey for core from 10.0.0.1 port 46640 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:32.615331 sshd-session[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:32.619516 systemd-logind[1560]: New session 6 of user core. Jan 13 20:34:32.629609 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 13 20:34:32.685105 sudo[1746]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 13 20:34:32.685537 sudo[1746]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 13 20:34:32.690035 sudo[1746]: pam_unix(sudo:session): session closed for user root Jan 13 20:34:32.697158 sudo[1745]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 13 20:34:32.697543 sudo[1745]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 13 20:34:32.716862 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 13 20:34:32.750012 augenrules[1768]: No rules Jan 13 20:34:32.751801 systemd[1]: audit-rules.service: Deactivated successfully. Jan 13 20:34:32.752177 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 13 20:34:32.753611 sudo[1745]: pam_unix(sudo:session): session closed for user root Jan 13 20:34:32.755436 sshd[1744]: Connection closed by 10.0.0.1 port 46640 Jan 13 20:34:32.756092 sshd-session[1738]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:32.770106 systemd[1]: Started sshd@6-10.0.0.53:22-10.0.0.1:46654.service - OpenSSH per-connection server daemon (10.0.0.1:46654). Jan 13 20:34:32.771335 systemd[1]: sshd@5-10.0.0.53:22-10.0.0.1:46640.service: Deactivated successfully. Jan 13 20:34:32.773964 systemd-logind[1560]: Session 6 logged out. Waiting for processes to exit. Jan 13 20:34:32.775774 systemd[1]: session-6.scope: Deactivated successfully. Jan 13 20:34:32.777113 systemd-logind[1560]: Removed session 6. Jan 13 20:34:32.808517 sshd[1774]: Accepted publickey for core from 10.0.0.1 port 46654 ssh2: RSA SHA256:uJ7Cm0ZiB1cKFsV9zv9H+G33T+grLCcYOUFbEs15LGg Jan 13 20:34:32.810171 sshd-session[1774]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:34:32.814754 systemd-logind[1560]: New session 7 of user core. Jan 13 20:34:32.828659 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 13 20:34:32.881837 sudo[1782]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 13 20:34:32.882237 sudo[1782]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 13 20:34:32.909558 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 13 20:34:32.931960 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 13 20:34:32.932361 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 13 20:34:33.398385 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:34:33.406445 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:34:33.422637 systemd[1]: Reloading requested from client PID 1834 ('systemctl') (unit session-7.scope)... Jan 13 20:34:33.422653 systemd[1]: Reloading... Jan 13 20:34:33.491378 zram_generator::config[1872]: No configuration found. Jan 13 20:34:33.788580 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:34:33.863091 systemd[1]: Reloading finished in 440 ms. Jan 13 20:34:33.915374 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 13 20:34:33.915490 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 13 20:34:33.915914 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:34:33.918968 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:34:34.058945 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:34:34.063763 (kubelet)[1933]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 13 20:34:34.108959 kubelet[1933]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:34:34.108959 kubelet[1933]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 13 20:34:34.108959 kubelet[1933]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:34:34.109472 kubelet[1933]: I0113 20:34:34.108997 1933 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 13 20:34:34.381737 kubelet[1933]: I0113 20:34:34.381623 1933 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Jan 13 20:34:34.381737 kubelet[1933]: I0113 20:34:34.381653 1933 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 13 20:34:34.383316 kubelet[1933]: I0113 20:34:34.382299 1933 server.go:919] "Client rotation is on, will bootstrap in background" Jan 13 20:34:34.398386 kubelet[1933]: I0113 20:34:34.398353 1933 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 13 20:34:34.410868 kubelet[1933]: I0113 20:34:34.410834 1933 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 13 20:34:34.411714 kubelet[1933]: I0113 20:34:34.411692 1933 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 13 20:34:34.411853 kubelet[1933]: I0113 20:34:34.411832 1933 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 13 20:34:34.412193 kubelet[1933]: I0113 20:34:34.412179 1933 topology_manager.go:138] "Creating topology manager with none policy" Jan 13 20:34:34.412193 kubelet[1933]: I0113 20:34:34.412192 1933 container_manager_linux.go:301] "Creating device plugin manager" Jan 13 20:34:34.412334 kubelet[1933]: I0113 20:34:34.412320 1933 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:34:34.412419 kubelet[1933]: I0113 20:34:34.412411 1933 kubelet.go:396] "Attempting to sync node with API server" Jan 13 20:34:34.412457 kubelet[1933]: I0113 20:34:34.412425 1933 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 13 20:34:34.412478 kubelet[1933]: I0113 20:34:34.412457 1933 kubelet.go:312] "Adding apiserver pod source" Jan 13 20:34:34.412496 kubelet[1933]: I0113 20:34:34.412478 1933 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 13 20:34:34.412602 kubelet[1933]: E0113 20:34:34.412585 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:34.412650 kubelet[1933]: E0113 20:34:34.412638 1933 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:34.413509 kubelet[1933]: I0113 20:34:34.413493 1933 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jan 13 20:34:34.415654 kubelet[1933]: I0113 20:34:34.415628 1933 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 13 20:34:34.415717 kubelet[1933]: W0113 20:34:34.415704 1933 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 13 20:34:34.416314 kubelet[1933]: W0113 20:34:34.416288 1933 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 13 20:34:34.416355 kubelet[1933]: E0113 20:34:34.416328 1933 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 13 20:34:34.416757 kubelet[1933]: W0113 20:34:34.416423 1933 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: nodes "10.0.0.53" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 13 20:34:34.416757 kubelet[1933]: I0113 20:34:34.416450 1933 server.go:1256] "Started kubelet" Jan 13 20:34:34.416757 kubelet[1933]: E0113 20:34:34.416454 1933 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: nodes "10.0.0.53" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 13 20:34:34.416757 kubelet[1933]: I0113 20:34:34.416492 1933 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Jan 13 20:34:34.416889 kubelet[1933]: I0113 20:34:34.416802 1933 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 13 20:34:34.417471 kubelet[1933]: I0113 20:34:34.417119 1933 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 13 20:34:34.417471 kubelet[1933]: I0113 20:34:34.417157 1933 server.go:461] "Adding debug handlers to kubelet server" Jan 13 20:34:34.420375 kubelet[1933]: I0113 20:34:34.420349 1933 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 13 20:34:34.421739 kubelet[1933]: I0113 20:34:34.421716 1933 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 13 20:34:34.423288 kubelet[1933]: I0113 20:34:34.422322 1933 factory.go:221] Registration of the systemd container factory successfully Jan 13 20:34:34.423288 kubelet[1933]: I0113 20:34:34.422399 1933 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 13 20:34:34.423573 kubelet[1933]: I0113 20:34:34.423549 1933 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Jan 13 20:34:34.424234 kubelet[1933]: I0113 20:34:34.424215 1933 reconciler_new.go:29] "Reconciler: start to sync state" Jan 13 20:34:34.425579 kubelet[1933]: E0113 20:34:34.425556 1933 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 13 20:34:34.426296 kubelet[1933]: I0113 20:34:34.426251 1933 factory.go:221] Registration of the containerd container factory successfully Jan 13 20:34:34.427648 kubelet[1933]: E0113 20:34:34.427621 1933 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"10.0.0.53\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Jan 13 20:34:34.431620 kubelet[1933]: W0113 20:34:34.431595 1933 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 13 20:34:34.431685 kubelet[1933]: E0113 20:34:34.431627 1933 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 13 20:34:34.435410 kubelet[1933]: E0113 20:34:34.435376 1933 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45b518de8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.416385512 +0000 UTC m=+0.348564354,LastTimestamp:2025-01-13 20:34:34.416385512 +0000 UTC m=+0.348564354,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.438569 kubelet[1933]: E0113 20:34:34.438550 1933 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45bdd52cc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.42554542 +0000 UTC m=+0.357724262,LastTimestamp:2025-01-13 20:34:34.42554542 +0000 UTC m=+0.357724262,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.443074 kubelet[1933]: I0113 20:34:34.443054 1933 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 13 20:34:34.443155 kubelet[1933]: I0113 20:34:34.443139 1933 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 13 20:34:34.443198 kubelet[1933]: I0113 20:34:34.443160 1933 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:34:34.446055 kubelet[1933]: E0113 20:34:34.446040 1933 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce2631f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.53 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442654495 +0000 UTC m=+0.374833337,LastTimestamp:2025-01-13 20:34:34.442654495 +0000 UTC m=+0.374833337,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.449240 kubelet[1933]: E0113 20:34:34.449219 1933 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce284a0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.53 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442663072 +0000 UTC m=+0.374841914,LastTimestamp:2025-01-13 20:34:34.442663072 +0000 UTC m=+0.374841914,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.452605 kubelet[1933]: E0113 20:34:34.452583 1933 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce29099 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.53 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442666137 +0000 UTC m=+0.374844979,LastTimestamp:2025-01-13 20:34:34.442666137 +0000 UTC m=+0.374844979,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.522557 kubelet[1933]: I0113 20:34:34.522524 1933 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.53" Jan 13 20:34:34.526114 kubelet[1933]: E0113 20:34:34.526081 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce2631f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce2631f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.53 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442654495 +0000 UTC m=+0.374833337,LastTimestamp:2025-01-13 20:34:34.522432928 +0000 UTC m=+0.454611771,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.526230 kubelet[1933]: E0113 20:34:34.526120 1933 kubelet_node_status.go:96] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="10.0.0.53" Jan 13 20:34:34.527880 kubelet[1933]: E0113 20:34:34.527864 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce284a0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce284a0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.53 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442663072 +0000 UTC m=+0.374841914,LastTimestamp:2025-01-13 20:34:34.522451603 +0000 UTC m=+0.454630446,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.529189 kubelet[1933]: E0113 20:34:34.529165 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce29099\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce29099 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.53 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442666137 +0000 UTC m=+0.374844979,LastTimestamp:2025-01-13 20:34:34.522454719 +0000 UTC m=+0.454633571,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.629467 kubelet[1933]: E0113 20:34:34.629402 1933 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"10.0.0.53\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="400ms" Jan 13 20:34:34.727479 kubelet[1933]: I0113 20:34:34.727343 1933 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.53" Jan 13 20:34:34.732780 kubelet[1933]: E0113 20:34:34.732742 1933 kubelet_node_status.go:96] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="10.0.0.53" Jan 13 20:34:34.732879 kubelet[1933]: E0113 20:34:34.732852 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce2631f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce2631f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.53 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442654495 +0000 UTC m=+0.374833337,LastTimestamp:2025-01-13 20:34:34.727303354 +0000 UTC m=+0.659482196,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.736410 kubelet[1933]: E0113 20:34:34.736374 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce284a0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce284a0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.53 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442663072 +0000 UTC m=+0.374841914,LastTimestamp:2025-01-13 20:34:34.727308022 +0000 UTC m=+0.659486864,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:34.739696 kubelet[1933]: E0113 20:34:34.739669 1933 event.go:346] "Server rejected event (will not retry!)" err="events \"10.0.0.53.181a5ad45ce29099\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.53.181a5ad45ce29099 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.53,UID:10.0.0.53,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.53 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.53,},FirstTimestamp:2025-01-13 20:34:34.442666137 +0000 UTC m=+0.374844979,LastTimestamp:2025-01-13 20:34:34.727311379 +0000 UTC m=+0.659490221,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.53,}" Jan 13 20:34:35.036948 kubelet[1933]: E0113 20:34:35.036812 1933 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"10.0.0.53\" not found" node="10.0.0.53" Jan 13 20:34:35.110046 kubelet[1933]: I0113 20:34:35.109973 1933 policy_none.go:49] "None policy: Start" Jan 13 20:34:35.111152 kubelet[1933]: I0113 20:34:35.111114 1933 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 13 20:34:35.111152 kubelet[1933]: I0113 20:34:35.111142 1933 state_mem.go:35] "Initializing new in-memory state store" Jan 13 20:34:35.120805 kubelet[1933]: I0113 20:34:35.119503 1933 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 13 20:34:35.120805 kubelet[1933]: I0113 20:34:35.119810 1933 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 13 20:34:35.121586 kubelet[1933]: E0113 20:34:35.121567 1933 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"10.0.0.53\" not found" Jan 13 20:34:35.133644 kubelet[1933]: I0113 20:34:35.133613 1933 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.53" Jan 13 20:34:35.136359 kubelet[1933]: I0113 20:34:35.136331 1933 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 13 20:34:35.137335 kubelet[1933]: I0113 20:34:35.137312 1933 kubelet_node_status.go:76] "Successfully registered node" node="10.0.0.53" Jan 13 20:34:35.137794 kubelet[1933]: I0113 20:34:35.137768 1933 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 13 20:34:35.137794 kubelet[1933]: I0113 20:34:35.137803 1933 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 13 20:34:35.137951 kubelet[1933]: I0113 20:34:35.137822 1933 kubelet.go:2329] "Starting kubelet main sync loop" Jan 13 20:34:35.138012 kubelet[1933]: E0113 20:34:35.137950 1933 kubelet.go:2353] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 13 20:34:35.149842 kubelet[1933]: E0113 20:34:35.149798 1933 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.53\" not found" Jan 13 20:34:35.250807 kubelet[1933]: E0113 20:34:35.250745 1933 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.53\" not found" Jan 13 20:34:35.351583 kubelet[1933]: E0113 20:34:35.351385 1933 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.53\" not found" Jan 13 20:34:35.385657 kubelet[1933]: I0113 20:34:35.385594 1933 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 13 20:34:35.385817 kubelet[1933]: W0113 20:34:35.385780 1933 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.RuntimeClass ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Jan 13 20:34:35.413035 kubelet[1933]: E0113 20:34:35.412983 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:35.451903 kubelet[1933]: E0113 20:34:35.451848 1933 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.53\" not found" Jan 13 20:34:35.552714 kubelet[1933]: E0113 20:34:35.552643 1933 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.53\" not found" Jan 13 20:34:35.576335 sudo[1782]: pam_unix(sudo:session): session closed for user root Jan 13 20:34:35.577768 sshd[1781]: Connection closed by 10.0.0.1 port 46654 Jan 13 20:34:35.578076 sshd-session[1774]: pam_unix(sshd:session): session closed for user core Jan 13 20:34:35.581696 systemd[1]: sshd@6-10.0.0.53:22-10.0.0.1:46654.service: Deactivated successfully. Jan 13 20:34:35.584096 systemd-logind[1560]: Session 7 logged out. Waiting for processes to exit. Jan 13 20:34:35.584135 systemd[1]: session-7.scope: Deactivated successfully. Jan 13 20:34:35.585203 systemd-logind[1560]: Removed session 7. Jan 13 20:34:35.653974 kubelet[1933]: I0113 20:34:35.653870 1933 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Jan 13 20:34:35.654223 containerd[1584]: time="2025-01-13T20:34:35.654178225Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 13 20:34:35.654686 kubelet[1933]: I0113 20:34:35.654394 1933 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Jan 13 20:34:36.413096 kubelet[1933]: I0113 20:34:36.413027 1933 apiserver.go:52] "Watching apiserver" Jan 13 20:34:36.413096 kubelet[1933]: E0113 20:34:36.413070 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:36.417138 kubelet[1933]: I0113 20:34:36.417105 1933 topology_manager.go:215] "Topology Admit Handler" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" podNamespace="kube-system" podName="cilium-jszvv" Jan 13 20:34:36.417227 kubelet[1933]: I0113 20:34:36.417219 1933 topology_manager.go:215] "Topology Admit Handler" podUID="28970548-af57-440e-bd56-609923b7b864" podNamespace="kube-system" podName="kube-proxy-ghqzl" Jan 13 20:34:36.424834 kubelet[1933]: I0113 20:34:36.424788 1933 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Jan 13 20:34:36.436564 kubelet[1933]: I0113 20:34:36.436522 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-run\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436564 kubelet[1933]: I0113 20:34:36.436563 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-etc-cni-netd\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436705 kubelet[1933]: I0113 20:34:36.436606 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-hubble-tls\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436705 kubelet[1933]: I0113 20:34:36.436646 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/28970548-af57-440e-bd56-609923b7b864-lib-modules\") pod \"kube-proxy-ghqzl\" (UID: \"28970548-af57-440e-bd56-609923b7b864\") " pod="kube-system/kube-proxy-ghqzl" Jan 13 20:34:36.436705 kubelet[1933]: I0113 20:34:36.436671 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aff58581-40f9-449b-ae11-6b18abd0053e-clustermesh-secrets\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436705 kubelet[1933]: I0113 20:34:36.436695 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-hostproc\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436830 kubelet[1933]: I0113 20:34:36.436719 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-lib-modules\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436830 kubelet[1933]: I0113 20:34:36.436743 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-xtables-lock\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436830 kubelet[1933]: I0113 20:34:36.436766 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/28970548-af57-440e-bd56-609923b7b864-kube-proxy\") pod \"kube-proxy-ghqzl\" (UID: \"28970548-af57-440e-bd56-609923b7b864\") " pod="kube-system/kube-proxy-ghqzl" Jan 13 20:34:36.436830 kubelet[1933]: I0113 20:34:36.436787 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-bpf-maps\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436830 kubelet[1933]: I0113 20:34:36.436822 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-cgroup\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436960 kubelet[1933]: I0113 20:34:36.436846 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cni-path\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436960 kubelet[1933]: I0113 20:34:36.436863 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-config-path\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436960 kubelet[1933]: I0113 20:34:36.436880 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-net\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.436960 kubelet[1933]: I0113 20:34:36.436904 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-kernel\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.437076 kubelet[1933]: I0113 20:34:36.436967 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqqmx\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-kube-api-access-vqqmx\") pod \"cilium-jszvv\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " pod="kube-system/cilium-jszvv" Jan 13 20:34:36.437076 kubelet[1933]: I0113 20:34:36.436997 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/28970548-af57-440e-bd56-609923b7b864-xtables-lock\") pod \"kube-proxy-ghqzl\" (UID: \"28970548-af57-440e-bd56-609923b7b864\") " pod="kube-system/kube-proxy-ghqzl" Jan 13 20:34:36.437076 kubelet[1933]: I0113 20:34:36.437025 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh8k5\" (UniqueName: \"kubernetes.io/projected/28970548-af57-440e-bd56-609923b7b864-kube-api-access-hh8k5\") pod \"kube-proxy-ghqzl\" (UID: \"28970548-af57-440e-bd56-609923b7b864\") " pod="kube-system/kube-proxy-ghqzl" Jan 13 20:34:36.721972 kubelet[1933]: E0113 20:34:36.721861 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:36.722550 containerd[1584]: time="2025-01-13T20:34:36.722502755Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ghqzl,Uid:28970548-af57-440e-bd56-609923b7b864,Namespace:kube-system,Attempt:0,}" Jan 13 20:34:36.723229 containerd[1584]: time="2025-01-13T20:34:36.723115444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jszvv,Uid:aff58581-40f9-449b-ae11-6b18abd0053e,Namespace:kube-system,Attempt:0,}" Jan 13 20:34:36.723255 kubelet[1933]: E0113 20:34:36.722877 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:37.317349 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3610492385.mount: Deactivated successfully. Jan 13 20:34:37.324527 containerd[1584]: time="2025-01-13T20:34:37.324455451Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:34:37.326254 containerd[1584]: time="2025-01-13T20:34:37.326136664Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jan 13 20:34:37.327618 containerd[1584]: time="2025-01-13T20:34:37.327572476Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:34:37.328901 containerd[1584]: time="2025-01-13T20:34:37.328848279Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:34:37.330121 containerd[1584]: time="2025-01-13T20:34:37.329776550Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 13 20:34:37.354631 containerd[1584]: time="2025-01-13T20:34:37.354560969Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:34:37.355323 containerd[1584]: time="2025-01-13T20:34:37.355258257Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 632.653972ms" Jan 13 20:34:37.356232 containerd[1584]: time="2025-01-13T20:34:37.356189273Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 633.001614ms" Jan 13 20:34:37.413837 kubelet[1933]: E0113 20:34:37.413776 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:37.470421 containerd[1584]: time="2025-01-13T20:34:37.470286296Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:34:37.470421 containerd[1584]: time="2025-01-13T20:34:37.470346158Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:34:37.470421 containerd[1584]: time="2025-01-13T20:34:37.470357159Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:37.470713 containerd[1584]: time="2025-01-13T20:34:37.470450634Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:37.471599 containerd[1584]: time="2025-01-13T20:34:37.469348046Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:34:37.471741 containerd[1584]: time="2025-01-13T20:34:37.471588788Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:34:37.471741 containerd[1584]: time="2025-01-13T20:34:37.471608475Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:37.471741 containerd[1584]: time="2025-01-13T20:34:37.471704165Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:37.568423 containerd[1584]: time="2025-01-13T20:34:37.568239313Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jszvv,Uid:aff58581-40f9-449b-ae11-6b18abd0053e,Namespace:kube-system,Attempt:0,} returns sandbox id \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\"" Jan 13 20:34:37.569453 kubelet[1933]: E0113 20:34:37.569413 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:37.572543 containerd[1584]: time="2025-01-13T20:34:37.572469486Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jan 13 20:34:37.573666 containerd[1584]: time="2025-01-13T20:34:37.573640983Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ghqzl,Uid:28970548-af57-440e-bd56-609923b7b864,Namespace:kube-system,Attempt:0,} returns sandbox id \"304fc832bcc7bd74825bcee37de518b29b11b01b93de087cad9ee10a5ee03336\"" Jan 13 20:34:37.574174 kubelet[1933]: E0113 20:34:37.574146 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:38.414291 kubelet[1933]: E0113 20:34:38.414233 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:39.414697 kubelet[1933]: E0113 20:34:39.414646 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:40.414980 kubelet[1933]: E0113 20:34:40.414932 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:40.998187 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1686143861.mount: Deactivated successfully. Jan 13 20:34:41.415727 kubelet[1933]: E0113 20:34:41.415588 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:42.415788 kubelet[1933]: E0113 20:34:42.415730 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:43.416761 kubelet[1933]: E0113 20:34:43.416696 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:44.240036 containerd[1584]: time="2025-01-13T20:34:44.239983695Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:34:44.242348 containerd[1584]: time="2025-01-13T20:34:44.242317441Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735307" Jan 13 20:34:44.244889 containerd[1584]: time="2025-01-13T20:34:44.244841595Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:34:44.246849 containerd[1584]: time="2025-01-13T20:34:44.246815998Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 6.674305294s" Jan 13 20:34:44.246849 containerd[1584]: time="2025-01-13T20:34:44.246844982Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Jan 13 20:34:44.247474 containerd[1584]: time="2025-01-13T20:34:44.247395224Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\"" Jan 13 20:34:44.248690 containerd[1584]: time="2025-01-13T20:34:44.248658633Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 13 20:34:44.319486 containerd[1584]: time="2025-01-13T20:34:44.319437479Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\"" Jan 13 20:34:44.319945 containerd[1584]: time="2025-01-13T20:34:44.319925124Z" level=info msg="StartContainer for \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\"" Jan 13 20:34:44.380983 containerd[1584]: time="2025-01-13T20:34:44.380897228Z" level=info msg="StartContainer for \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\" returns successfully" Jan 13 20:34:44.417098 kubelet[1933]: E0113 20:34:44.417035 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:45.129008 containerd[1584]: time="2025-01-13T20:34:45.128958700Z" level=error msg="collecting metrics for 9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee" error="cgroups: cgroup deleted: unknown" Jan 13 20:34:45.155103 kubelet[1933]: E0113 20:34:45.155066 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:45.210907 containerd[1584]: time="2025-01-13T20:34:45.210833294Z" level=info msg="shim disconnected" id=9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee namespace=k8s.io Jan 13 20:34:45.210907 containerd[1584]: time="2025-01-13T20:34:45.210900060Z" level=warning msg="cleaning up after shim disconnected" id=9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee namespace=k8s.io Jan 13 20:34:45.210907 containerd[1584]: time="2025-01-13T20:34:45.210912293Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:34:45.270174 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee-rootfs.mount: Deactivated successfully. Jan 13 20:34:45.418263 kubelet[1933]: E0113 20:34:45.417967 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:46.156912 kubelet[1933]: E0113 20:34:46.156882 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:46.158726 containerd[1584]: time="2025-01-13T20:34:46.158682099Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 13 20:34:46.418761 kubelet[1933]: E0113 20:34:46.418650 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:46.447446 containerd[1584]: time="2025-01-13T20:34:46.447403635Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\"" Jan 13 20:34:46.447786 containerd[1584]: time="2025-01-13T20:34:46.447756968Z" level=info msg="StartContainer for \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\"" Jan 13 20:34:46.512953 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 13 20:34:46.513284 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:34:46.513351 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:34:46.525751 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:34:46.539297 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:34:46.549909 containerd[1584]: time="2025-01-13T20:34:46.549862212Z" level=info msg="StartContainer for \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\" returns successfully" Jan 13 20:34:46.909387 containerd[1584]: time="2025-01-13T20:34:46.906232042Z" level=info msg="shim disconnected" id=de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f namespace=k8s.io Jan 13 20:34:46.909387 containerd[1584]: time="2025-01-13T20:34:46.906322362Z" level=warning msg="cleaning up after shim disconnected" id=de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f namespace=k8s.io Jan 13 20:34:46.909387 containerd[1584]: time="2025-01-13T20:34:46.906333162Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:34:47.159664 kubelet[1933]: E0113 20:34:47.159553 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:47.161426 containerd[1584]: time="2025-01-13T20:34:47.161369236Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 13 20:34:47.324801 containerd[1584]: time="2025-01-13T20:34:47.324740846Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\"" Jan 13 20:34:47.325535 containerd[1584]: time="2025-01-13T20:34:47.325503306Z" level=info msg="StartContainer for \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\"" Jan 13 20:34:47.352957 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f-rootfs.mount: Deactivated successfully. Jan 13 20:34:47.353146 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4184832364.mount: Deactivated successfully. Jan 13 20:34:47.419086 kubelet[1933]: E0113 20:34:47.418971 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:47.461370 containerd[1584]: time="2025-01-13T20:34:47.461335289Z" level=info msg="StartContainer for \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\" returns successfully" Jan 13 20:34:47.486442 containerd[1584]: time="2025-01-13T20:34:47.486379496Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:34:47.486424 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb-rootfs.mount: Deactivated successfully. Jan 13 20:34:47.965629 containerd[1584]: time="2025-01-13T20:34:47.964924068Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.12: active requests=0, bytes read=28619958" Jan 13 20:34:48.023536 containerd[1584]: time="2025-01-13T20:34:48.023441118Z" level=info msg="shim disconnected" id=5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb namespace=k8s.io Jan 13 20:34:48.023536 containerd[1584]: time="2025-01-13T20:34:48.023524595Z" level=warning msg="cleaning up after shim disconnected" id=5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb namespace=k8s.io Jan 13 20:34:48.023536 containerd[1584]: time="2025-01-13T20:34:48.023541506Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:34:48.030683 containerd[1584]: time="2025-01-13T20:34:48.030612657Z" level=info msg="ImageCreate event name:\"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:34:48.052390 containerd[1584]: time="2025-01-13T20:34:48.052336427Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:34:48.053024 containerd[1584]: time="2025-01-13T20:34:48.052991255Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.12\" with image id \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\", repo tag \"registry.k8s.io/kube-proxy:v1.29.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39\", size \"28618977\" in 3.805567618s" Jan 13 20:34:48.053024 containerd[1584]: time="2025-01-13T20:34:48.053019037Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\" returns image reference \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\"" Jan 13 20:34:48.054728 containerd[1584]: time="2025-01-13T20:34:48.054689760Z" level=info msg="CreateContainer within sandbox \"304fc832bcc7bd74825bcee37de518b29b11b01b93de087cad9ee10a5ee03336\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 13 20:34:48.163348 kubelet[1933]: E0113 20:34:48.163288 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:48.165560 containerd[1584]: time="2025-01-13T20:34:48.165502384Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 13 20:34:48.225924 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1084236522.mount: Deactivated successfully. Jan 13 20:34:48.391392 containerd[1584]: time="2025-01-13T20:34:48.391318890Z" level=info msg="CreateContainer within sandbox \"304fc832bcc7bd74825bcee37de518b29b11b01b93de087cad9ee10a5ee03336\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"55dccd8d6ae7b1a5c5be2c189b2cd5bedc0e7ab64a7792a9de83d4e6514effe7\"" Jan 13 20:34:48.391900 containerd[1584]: time="2025-01-13T20:34:48.391861197Z" level=info msg="StartContainer for \"55dccd8d6ae7b1a5c5be2c189b2cd5bedc0e7ab64a7792a9de83d4e6514effe7\"" Jan 13 20:34:48.435544 kubelet[1933]: E0113 20:34:48.435468 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:48.571153 containerd[1584]: time="2025-01-13T20:34:48.571025354Z" level=info msg="StartContainer for \"55dccd8d6ae7b1a5c5be2c189b2cd5bedc0e7ab64a7792a9de83d4e6514effe7\" returns successfully" Jan 13 20:34:48.571153 containerd[1584]: time="2025-01-13T20:34:48.571090917Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\"" Jan 13 20:34:48.571728 containerd[1584]: time="2025-01-13T20:34:48.571704447Z" level=info msg="StartContainer for \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\"" Jan 13 20:34:48.677557 containerd[1584]: time="2025-01-13T20:34:48.677504911Z" level=info msg="StartContainer for \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\" returns successfully" Jan 13 20:34:48.802496 containerd[1584]: time="2025-01-13T20:34:48.802419981Z" level=info msg="shim disconnected" id=1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3 namespace=k8s.io Jan 13 20:34:48.802496 containerd[1584]: time="2025-01-13T20:34:48.802494571Z" level=warning msg="cleaning up after shim disconnected" id=1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3 namespace=k8s.io Jan 13 20:34:48.802710 containerd[1584]: time="2025-01-13T20:34:48.802506544Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:34:49.165954 kubelet[1933]: E0113 20:34:49.165927 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:49.167863 kubelet[1933]: E0113 20:34:49.167840 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:49.169704 containerd[1584]: time="2025-01-13T20:34:49.169671289Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 13 20:34:49.219455 kubelet[1933]: I0113 20:34:49.219415 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-ghqzl" podStartSLOduration=3.741054199 podStartE2EDuration="14.219366315s" podCreationTimestamp="2025-01-13 20:34:35 +0000 UTC" firstStartedPulling="2025-01-13 20:34:37.57492478 +0000 UTC m=+3.507103622" lastFinishedPulling="2025-01-13 20:34:48.053236896 +0000 UTC m=+13.985415738" observedRunningTime="2025-01-13 20:34:49.219320258 +0000 UTC m=+15.151499100" watchObservedRunningTime="2025-01-13 20:34:49.219366315 +0000 UTC m=+15.151545157" Jan 13 20:34:49.436387 kubelet[1933]: E0113 20:34:49.436230 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:49.534070 containerd[1584]: time="2025-01-13T20:34:49.534006758Z" level=info msg="CreateContainer within sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\"" Jan 13 20:34:49.534574 containerd[1584]: time="2025-01-13T20:34:49.534515171Z" level=info msg="StartContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\"" Jan 13 20:34:49.662517 containerd[1584]: time="2025-01-13T20:34:49.662434586Z" level=info msg="StartContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" returns successfully" Jan 13 20:34:49.718089 systemd[1]: run-containerd-runc-k8s.io-0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1-runc.WLYJqB.mount: Deactivated successfully. Jan 13 20:34:49.809511 kubelet[1933]: I0113 20:34:49.809486 1933 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Jan 13 20:34:50.176048 kubelet[1933]: E0113 20:34:50.175721 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:50.180332 kubelet[1933]: E0113 20:34:50.180067 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:50.296038 kernel: Initializing XFRM netlink socket Jan 13 20:34:50.436857 kubelet[1933]: E0113 20:34:50.436704 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:51.177530 kubelet[1933]: E0113 20:34:51.177495 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:51.437826 kubelet[1933]: E0113 20:34:51.437689 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:51.830437 kubelet[1933]: I0113 20:34:51.830238 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-jszvv" podStartSLOduration=10.153034071 podStartE2EDuration="16.830181195s" podCreationTimestamp="2025-01-13 20:34:35 +0000 UTC" firstStartedPulling="2025-01-13 20:34:37.570095764 +0000 UTC m=+3.502274606" lastFinishedPulling="2025-01-13 20:34:44.247242888 +0000 UTC m=+10.179421730" observedRunningTime="2025-01-13 20:34:50.226787138 +0000 UTC m=+16.158965980" watchObservedRunningTime="2025-01-13 20:34:51.830181195 +0000 UTC m=+17.762360047" Jan 13 20:34:51.830637 kubelet[1933]: I0113 20:34:51.830583 1933 topology_manager.go:215] "Topology Admit Handler" podUID="7e592dc2-7025-4d14-93fd-0eed6e44066d" podNamespace="default" podName="nginx-deployment-6d5f899847-sgz2x" Jan 13 20:34:51.958421 kubelet[1933]: I0113 20:34:51.958215 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkhg\" (UniqueName: \"kubernetes.io/projected/7e592dc2-7025-4d14-93fd-0eed6e44066d-kube-api-access-6mkhg\") pod \"nginx-deployment-6d5f899847-sgz2x\" (UID: \"7e592dc2-7025-4d14-93fd-0eed6e44066d\") " pod="default/nginx-deployment-6d5f899847-sgz2x" Jan 13 20:34:52.178614 containerd[1584]: time="2025-01-13T20:34:52.177473902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-sgz2x,Uid:7e592dc2-7025-4d14-93fd-0eed6e44066d,Namespace:default,Attempt:0,}" Jan 13 20:34:52.180689 kubelet[1933]: E0113 20:34:52.180065 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:52.262586 systemd-networkd[1243]: cilium_host: Link UP Jan 13 20:34:52.262796 systemd-networkd[1243]: cilium_net: Link UP Jan 13 20:34:52.262801 systemd-networkd[1243]: cilium_net: Gained carrier Jan 13 20:34:52.263042 systemd-networkd[1243]: cilium_host: Gained carrier Jan 13 20:34:52.270745 systemd-networkd[1243]: cilium_host: Gained IPv6LL Jan 13 20:34:52.439586 kubelet[1933]: E0113 20:34:52.439403 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:52.476431 systemd-networkd[1243]: cilium_vxlan: Link UP Jan 13 20:34:52.476445 systemd-networkd[1243]: cilium_vxlan: Gained carrier Jan 13 20:34:52.536460 systemd-networkd[1243]: cilium_net: Gained IPv6LL Jan 13 20:34:52.703313 kernel: NET: Registered PF_ALG protocol family Jan 13 20:34:53.359080 systemd-networkd[1243]: lxc_health: Link UP Jan 13 20:34:53.376555 systemd-networkd[1243]: lxc_health: Gained carrier Jan 13 20:34:53.440013 kubelet[1933]: E0113 20:34:53.439962 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:53.872453 systemd-networkd[1243]: cilium_vxlan: Gained IPv6LL Jan 13 20:34:53.939678 systemd-networkd[1243]: lxcdd17473e9be7: Link UP Jan 13 20:34:53.951317 kernel: eth0: renamed from tmp610db Jan 13 20:34:53.956726 systemd-networkd[1243]: lxcdd17473e9be7: Gained carrier Jan 13 20:34:54.413590 kubelet[1933]: E0113 20:34:54.413541 1933 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:54.440806 kubelet[1933]: E0113 20:34:54.440725 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:54.725573 kubelet[1933]: E0113 20:34:54.725447 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:34:55.024545 systemd-networkd[1243]: lxc_health: Gained IPv6LL Jan 13 20:34:55.455313 kubelet[1933]: E0113 20:34:55.446100 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:55.670407 systemd-networkd[1243]: lxcdd17473e9be7: Gained IPv6LL Jan 13 20:34:56.452253 kubelet[1933]: E0113 20:34:56.452183 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:57.453685 kubelet[1933]: E0113 20:34:57.453330 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:58.235602 containerd[1584]: time="2025-01-13T20:34:58.235311355Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:34:58.235602 containerd[1584]: time="2025-01-13T20:34:58.235385767Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:34:58.235602 containerd[1584]: time="2025-01-13T20:34:58.235405585Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:58.236248 containerd[1584]: time="2025-01-13T20:34:58.235661909Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:34:58.299151 systemd-resolved[1458]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 13 20:34:58.342633 containerd[1584]: time="2025-01-13T20:34:58.342493406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-sgz2x,Uid:7e592dc2-7025-4d14-93fd-0eed6e44066d,Namespace:default,Attempt:0,} returns sandbox id \"610db1586525753fae6045d1a22de255f569f97a052c69c392fd0b17bac22114\"" Jan 13 20:34:58.348428 containerd[1584]: time="2025-01-13T20:34:58.344325540Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Jan 13 20:34:58.454211 kubelet[1933]: E0113 20:34:58.454133 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:34:59.455156 kubelet[1933]: E0113 20:34:59.455106 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:00.455740 kubelet[1933]: E0113 20:35:00.455696 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:01.195236 kubelet[1933]: I0113 20:35:01.194797 1933 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 13 20:35:01.196020 kubelet[1933]: E0113 20:35:01.195836 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:01.204288 kubelet[1933]: E0113 20:35:01.204201 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:01.456441 kubelet[1933]: E0113 20:35:01.456246 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:02.458296 kubelet[1933]: E0113 20:35:02.457304 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:03.025886 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount554248789.mount: Deactivated successfully. Jan 13 20:35:03.457779 kubelet[1933]: E0113 20:35:03.457734 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:04.458738 kubelet[1933]: E0113 20:35:04.458691 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:05.459527 kubelet[1933]: E0113 20:35:05.459305 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:05.485995 containerd[1584]: time="2025-01-13T20:35:05.485833703Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:05.488209 containerd[1584]: time="2025-01-13T20:35:05.488139176Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=71036018" Jan 13 20:35:05.490982 containerd[1584]: time="2025-01-13T20:35:05.490664680Z" level=info msg="ImageCreate event name:\"sha256:29ef6eaebfc53650f3a4609edbf9d35e866f56b2c5e01d32d93439031b300f0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:05.496759 containerd[1584]: time="2025-01-13T20:35:05.495964110Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:eca1d1ff18c7af45f86b7e0b572090f563a676ddca3da2ecff678390366335ad\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:05.497398 containerd[1584]: time="2025-01-13T20:35:05.497040081Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:29ef6eaebfc53650f3a4609edbf9d35e866f56b2c5e01d32d93439031b300f0b\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:eca1d1ff18c7af45f86b7e0b572090f563a676ddca3da2ecff678390366335ad\", size \"71035896\" in 7.152674394s" Jan 13 20:35:05.497398 containerd[1584]: time="2025-01-13T20:35:05.497109765Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:29ef6eaebfc53650f3a4609edbf9d35e866f56b2c5e01d32d93439031b300f0b\"" Jan 13 20:35:05.501776 containerd[1584]: time="2025-01-13T20:35:05.501600934Z" level=info msg="CreateContainer within sandbox \"610db1586525753fae6045d1a22de255f569f97a052c69c392fd0b17bac22114\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Jan 13 20:35:05.537614 containerd[1584]: time="2025-01-13T20:35:05.537488284Z" level=info msg="CreateContainer within sandbox \"610db1586525753fae6045d1a22de255f569f97a052c69c392fd0b17bac22114\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"9bdadb26dc80008937e28fef3a6ae1d511216d696522aa656b4eb1db89b9a826\"" Jan 13 20:35:05.538693 containerd[1584]: time="2025-01-13T20:35:05.538595805Z" level=info msg="StartContainer for \"9bdadb26dc80008937e28fef3a6ae1d511216d696522aa656b4eb1db89b9a826\"" Jan 13 20:35:05.578533 systemd[1]: run-containerd-runc-k8s.io-9bdadb26dc80008937e28fef3a6ae1d511216d696522aa656b4eb1db89b9a826-runc.bSAMyi.mount: Deactivated successfully. Jan 13 20:35:05.663359 containerd[1584]: time="2025-01-13T20:35:05.662717139Z" level=info msg="StartContainer for \"9bdadb26dc80008937e28fef3a6ae1d511216d696522aa656b4eb1db89b9a826\" returns successfully" Jan 13 20:35:06.235554 kubelet[1933]: I0113 20:35:06.235490 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-sgz2x" podStartSLOduration=8.081571245 podStartE2EDuration="15.23545828s" podCreationTimestamp="2025-01-13 20:34:51 +0000 UTC" firstStartedPulling="2025-01-13 20:34:58.343838683 +0000 UTC m=+24.276017525" lastFinishedPulling="2025-01-13 20:35:05.497725718 +0000 UTC m=+31.429904560" observedRunningTime="2025-01-13 20:35:06.235439825 +0000 UTC m=+32.167618667" watchObservedRunningTime="2025-01-13 20:35:06.23545828 +0000 UTC m=+32.167637122" Jan 13 20:35:06.459683 kubelet[1933]: E0113 20:35:06.459607 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:07.459856 kubelet[1933]: E0113 20:35:07.459767 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:07.464421 update_engine[1563]: I20250113 20:35:07.464342 1563 update_attempter.cc:509] Updating boot flags... Jan 13 20:35:07.495306 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (3136) Jan 13 20:35:07.532306 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (3137) Jan 13 20:35:07.564415 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (3137) Jan 13 20:35:08.460916 kubelet[1933]: E0113 20:35:08.460853 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:09.461174 kubelet[1933]: E0113 20:35:09.461009 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:10.086781 kubelet[1933]: I0113 20:35:10.086732 1933 topology_manager.go:215] "Topology Admit Handler" podUID="160acbc1-1acb-43e5-be48-db9ab1abfac7" podNamespace="default" podName="nfs-server-provisioner-0" Jan 13 20:35:10.204235 kubelet[1933]: I0113 20:35:10.204170 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/160acbc1-1acb-43e5-be48-db9ab1abfac7-data\") pod \"nfs-server-provisioner-0\" (UID: \"160acbc1-1acb-43e5-be48-db9ab1abfac7\") " pod="default/nfs-server-provisioner-0" Jan 13 20:35:10.204235 kubelet[1933]: I0113 20:35:10.204220 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz9lt\" (UniqueName: \"kubernetes.io/projected/160acbc1-1acb-43e5-be48-db9ab1abfac7-kube-api-access-rz9lt\") pod \"nfs-server-provisioner-0\" (UID: \"160acbc1-1acb-43e5-be48-db9ab1abfac7\") " pod="default/nfs-server-provisioner-0" Jan 13 20:35:10.391811 containerd[1584]: time="2025-01-13T20:35:10.391753701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:160acbc1-1acb-43e5-be48-db9ab1abfac7,Namespace:default,Attempt:0,}" Jan 13 20:35:10.461731 kubelet[1933]: E0113 20:35:10.461647 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:10.869395 systemd-networkd[1243]: lxce357c00c31d7: Link UP Jan 13 20:35:10.882099 kernel: eth0: renamed from tmp23eac Jan 13 20:35:10.894555 systemd-networkd[1243]: lxce357c00c31d7: Gained carrier Jan 13 20:35:11.227048 containerd[1584]: time="2025-01-13T20:35:11.226566665Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:35:11.227048 containerd[1584]: time="2025-01-13T20:35:11.226713163Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:35:11.227048 containerd[1584]: time="2025-01-13T20:35:11.226738582Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:11.229432 containerd[1584]: time="2025-01-13T20:35:11.229284148Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:11.274988 systemd-resolved[1458]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 13 20:35:11.313209 containerd[1584]: time="2025-01-13T20:35:11.313067414Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:160acbc1-1acb-43e5-be48-db9ab1abfac7,Namespace:default,Attempt:0,} returns sandbox id \"23eac1f02b7ea7bd0159d40d672a09ff7f4b2af614b9ecf49b49b33fbbc12322\"" Jan 13 20:35:11.315864 containerd[1584]: time="2025-01-13T20:35:11.315820063Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Jan 13 20:35:11.462316 kubelet[1933]: E0113 20:35:11.462198 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:12.049585 systemd-networkd[1243]: lxce357c00c31d7: Gained IPv6LL Jan 13 20:35:12.462422 kubelet[1933]: E0113 20:35:12.462368 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:13.463490 kubelet[1933]: E0113 20:35:13.463445 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:13.686070 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3032726493.mount: Deactivated successfully. Jan 13 20:35:14.413200 kubelet[1933]: E0113 20:35:14.413156 1933 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:14.464621 kubelet[1933]: E0113 20:35:14.464577 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:15.465462 kubelet[1933]: E0113 20:35:15.465387 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:16.465695 kubelet[1933]: E0113 20:35:16.465625 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:17.466387 kubelet[1933]: E0113 20:35:17.466335 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:17.835788 containerd[1584]: time="2025-01-13T20:35:17.833982604Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:17.835788 containerd[1584]: time="2025-01-13T20:35:17.835214211Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=91039406" Jan 13 20:35:17.840817 containerd[1584]: time="2025-01-13T20:35:17.840738079Z" level=info msg="ImageCreate event name:\"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:17.856325 containerd[1584]: time="2025-01-13T20:35:17.854707222Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"91036984\" in 6.538843767s" Jan 13 20:35:17.856325 containerd[1584]: time="2025-01-13T20:35:17.854762767Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Jan 13 20:35:17.856325 containerd[1584]: time="2025-01-13T20:35:17.854960541Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:17.865545 containerd[1584]: time="2025-01-13T20:35:17.865370020Z" level=info msg="CreateContainer within sandbox \"23eac1f02b7ea7bd0159d40d672a09ff7f4b2af614b9ecf49b49b33fbbc12322\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Jan 13 20:35:17.916134 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1954715519.mount: Deactivated successfully. Jan 13 20:35:17.936369 containerd[1584]: time="2025-01-13T20:35:17.936246950Z" level=info msg="CreateContainer within sandbox \"23eac1f02b7ea7bd0159d40d672a09ff7f4b2af614b9ecf49b49b33fbbc12322\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"77bffff649f23c40e41384e32fade213970b84f175bc3e02f1775bb254c66f37\"" Jan 13 20:35:17.937933 containerd[1584]: time="2025-01-13T20:35:17.937862061Z" level=info msg="StartContainer for \"77bffff649f23c40e41384e32fade213970b84f175bc3e02f1775bb254c66f37\"" Jan 13 20:35:18.096105 containerd[1584]: time="2025-01-13T20:35:18.095962545Z" level=info msg="StartContainer for \"77bffff649f23c40e41384e32fade213970b84f175bc3e02f1775bb254c66f37\" returns successfully" Jan 13 20:35:18.273660 kubelet[1933]: I0113 20:35:18.273618 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=1.7319289759999998 podStartE2EDuration="8.273576175s" podCreationTimestamp="2025-01-13 20:35:10 +0000 UTC" firstStartedPulling="2025-01-13 20:35:11.315150784 +0000 UTC m=+37.247329636" lastFinishedPulling="2025-01-13 20:35:17.856797993 +0000 UTC m=+43.788976835" observedRunningTime="2025-01-13 20:35:18.272838482 +0000 UTC m=+44.205017324" watchObservedRunningTime="2025-01-13 20:35:18.273576175 +0000 UTC m=+44.205755017" Jan 13 20:35:18.467549 kubelet[1933]: E0113 20:35:18.467489 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:19.467831 kubelet[1933]: E0113 20:35:19.467723 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:20.468203 kubelet[1933]: E0113 20:35:20.468131 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:21.469489 kubelet[1933]: E0113 20:35:21.469343 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:22.470090 kubelet[1933]: E0113 20:35:22.470032 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:23.471522 kubelet[1933]: E0113 20:35:23.471091 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:24.471529 kubelet[1933]: E0113 20:35:24.471467 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:25.471920 kubelet[1933]: E0113 20:35:25.471832 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:26.472862 kubelet[1933]: E0113 20:35:26.472799 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:27.473428 kubelet[1933]: E0113 20:35:27.473359 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:28.304845 kubelet[1933]: I0113 20:35:28.304802 1933 topology_manager.go:215] "Topology Admit Handler" podUID="cfd60918-b153-4e08-8e5f-5df165dbbce5" podNamespace="default" podName="test-pod-1" Jan 13 20:35:28.443702 kubelet[1933]: I0113 20:35:28.443643 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0091771d-c10a-4968-8875-1f0bcdf5c853\" (UniqueName: \"kubernetes.io/nfs/cfd60918-b153-4e08-8e5f-5df165dbbce5-pvc-0091771d-c10a-4968-8875-1f0bcdf5c853\") pod \"test-pod-1\" (UID: \"cfd60918-b153-4e08-8e5f-5df165dbbce5\") " pod="default/test-pod-1" Jan 13 20:35:28.443702 kubelet[1933]: I0113 20:35:28.443717 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtrm\" (UniqueName: \"kubernetes.io/projected/cfd60918-b153-4e08-8e5f-5df165dbbce5-kube-api-access-qxtrm\") pod \"test-pod-1\" (UID: \"cfd60918-b153-4e08-8e5f-5df165dbbce5\") " pod="default/test-pod-1" Jan 13 20:35:28.473701 kubelet[1933]: E0113 20:35:28.473517 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:28.588311 kernel: FS-Cache: Loaded Jan 13 20:35:28.675334 kernel: RPC: Registered named UNIX socket transport module. Jan 13 20:35:28.675482 kernel: RPC: Registered udp transport module. Jan 13 20:35:28.675519 kernel: RPC: Registered tcp transport module. Jan 13 20:35:28.676981 kernel: RPC: Registered tcp-with-tls transport module. Jan 13 20:35:28.677032 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Jan 13 20:35:28.960596 kernel: NFS: Registering the id_resolver key type Jan 13 20:35:28.960766 kernel: Key type id_resolver registered Jan 13 20:35:28.960794 kernel: Key type id_legacy registered Jan 13 20:35:28.990062 nfsidmap[3332]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Jan 13 20:35:28.995655 nfsidmap[3335]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Jan 13 20:35:29.208882 containerd[1584]: time="2025-01-13T20:35:29.208803603Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:cfd60918-b153-4e08-8e5f-5df165dbbce5,Namespace:default,Attempt:0,}" Jan 13 20:35:29.239873 systemd-networkd[1243]: lxc4d9e5500785d: Link UP Jan 13 20:35:29.251995 kernel: eth0: renamed from tmpca333 Jan 13 20:35:29.257010 systemd-networkd[1243]: lxc4d9e5500785d: Gained carrier Jan 13 20:35:29.463450 containerd[1584]: time="2025-01-13T20:35:29.463228326Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:35:29.463450 containerd[1584]: time="2025-01-13T20:35:29.463303497Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:35:29.463450 containerd[1584]: time="2025-01-13T20:35:29.463314718Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:29.463450 containerd[1584]: time="2025-01-13T20:35:29.463410318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:29.474144 kubelet[1933]: E0113 20:35:29.474098 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:29.497091 systemd-resolved[1458]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 13 20:35:29.525563 containerd[1584]: time="2025-01-13T20:35:29.525530891Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:cfd60918-b153-4e08-8e5f-5df165dbbce5,Namespace:default,Attempt:0,} returns sandbox id \"ca333607049bd78000429c2919faf45e035059ebf6b1f7ef5b39b7e8310c0df0\"" Jan 13 20:35:29.527482 containerd[1584]: time="2025-01-13T20:35:29.527438851Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Jan 13 20:35:29.995804 containerd[1584]: time="2025-01-13T20:35:29.995722918Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:29.996623 containerd[1584]: time="2025-01-13T20:35:29.996567298Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Jan 13 20:35:29.999083 containerd[1584]: time="2025-01-13T20:35:29.999051382Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:29ef6eaebfc53650f3a4609edbf9d35e866f56b2c5e01d32d93439031b300f0b\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:eca1d1ff18c7af45f86b7e0b572090f563a676ddca3da2ecff678390366335ad\", size \"71035896\" in 471.571583ms" Jan 13 20:35:29.999083 containerd[1584]: time="2025-01-13T20:35:29.999080938Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:29ef6eaebfc53650f3a4609edbf9d35e866f56b2c5e01d32d93439031b300f0b\"" Jan 13 20:35:30.000617 containerd[1584]: time="2025-01-13T20:35:30.000570812Z" level=info msg="CreateContainer within sandbox \"ca333607049bd78000429c2919faf45e035059ebf6b1f7ef5b39b7e8310c0df0\" for container &ContainerMetadata{Name:test,Attempt:0,}" Jan 13 20:35:30.013790 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1721603909.mount: Deactivated successfully. Jan 13 20:35:30.017996 containerd[1584]: time="2025-01-13T20:35:30.017961722Z" level=info msg="CreateContainer within sandbox \"ca333607049bd78000429c2919faf45e035059ebf6b1f7ef5b39b7e8310c0df0\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"200d3736bda7c1497e89b71f576aa59d78c08c1e8e86baa597fab3c85904b3af\"" Jan 13 20:35:30.018406 containerd[1584]: time="2025-01-13T20:35:30.018381180Z" level=info msg="StartContainer for \"200d3736bda7c1497e89b71f576aa59d78c08c1e8e86baa597fab3c85904b3af\"" Jan 13 20:35:30.069743 containerd[1584]: time="2025-01-13T20:35:30.069575851Z" level=info msg="StartContainer for \"200d3736bda7c1497e89b71f576aa59d78c08c1e8e86baa597fab3c85904b3af\" returns successfully" Jan 13 20:35:30.308151 kubelet[1933]: I0113 20:35:30.307976 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=19.835583552 podStartE2EDuration="20.307921348s" podCreationTimestamp="2025-01-13 20:35:10 +0000 UTC" firstStartedPulling="2025-01-13 20:35:29.526984296 +0000 UTC m=+55.459163138" lastFinishedPulling="2025-01-13 20:35:29.999322092 +0000 UTC m=+55.931500934" observedRunningTime="2025-01-13 20:35:30.307239205 +0000 UTC m=+56.239418057" watchObservedRunningTime="2025-01-13 20:35:30.307921348 +0000 UTC m=+56.240100190" Jan 13 20:35:30.475161 kubelet[1933]: E0113 20:35:30.475124 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:30.544721 systemd-networkd[1243]: lxc4d9e5500785d: Gained IPv6LL Jan 13 20:35:31.475659 kubelet[1933]: E0113 20:35:31.475556 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:32.476211 kubelet[1933]: E0113 20:35:32.476054 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:33.462187 containerd[1584]: time="2025-01-13T20:35:33.462112675Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 13 20:35:33.473076 containerd[1584]: time="2025-01-13T20:35:33.473029522Z" level=info msg="StopContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" with timeout 2 (s)" Jan 13 20:35:33.473393 containerd[1584]: time="2025-01-13T20:35:33.473359703Z" level=info msg="Stop container \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" with signal terminated" Jan 13 20:35:33.476385 kubelet[1933]: E0113 20:35:33.476297 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:33.482576 systemd-networkd[1243]: lxc_health: Link DOWN Jan 13 20:35:33.482584 systemd-networkd[1243]: lxc_health: Lost carrier Jan 13 20:35:33.547386 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1-rootfs.mount: Deactivated successfully. Jan 13 20:35:33.575525 containerd[1584]: time="2025-01-13T20:35:33.575438079Z" level=info msg="shim disconnected" id=0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1 namespace=k8s.io Jan 13 20:35:33.575525 containerd[1584]: time="2025-01-13T20:35:33.575508572Z" level=warning msg="cleaning up after shim disconnected" id=0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1 namespace=k8s.io Jan 13 20:35:33.575525 containerd[1584]: time="2025-01-13T20:35:33.575523590Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:33.598305 containerd[1584]: time="2025-01-13T20:35:33.598191645Z" level=info msg="StopContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" returns successfully" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.598971812Z" level=info msg="StopPodSandbox for \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\"" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.599014662Z" level=info msg="Container to stop \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.599057282Z" level=info msg="Container to stop \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.599068984Z" level=info msg="Container to stop \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.599080235Z" level=info msg="Container to stop \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 13 20:35:33.599545 containerd[1584]: time="2025-01-13T20:35:33.599092518Z" level=info msg="Container to stop \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 13 20:35:33.601715 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa-shm.mount: Deactivated successfully. Jan 13 20:35:33.625892 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa-rootfs.mount: Deactivated successfully. Jan 13 20:35:33.629120 containerd[1584]: time="2025-01-13T20:35:33.629052976Z" level=info msg="shim disconnected" id=c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa namespace=k8s.io Jan 13 20:35:33.629120 containerd[1584]: time="2025-01-13T20:35:33.629116736Z" level=warning msg="cleaning up after shim disconnected" id=c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa namespace=k8s.io Jan 13 20:35:33.629359 containerd[1584]: time="2025-01-13T20:35:33.629126154Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:33.644214 containerd[1584]: time="2025-01-13T20:35:33.644155835Z" level=info msg="TearDown network for sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" successfully" Jan 13 20:35:33.644214 containerd[1584]: time="2025-01-13T20:35:33.644194368Z" level=info msg="StopPodSandbox for \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" returns successfully" Jan 13 20:35:33.794458 kubelet[1933]: I0113 20:35:33.794334 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aff58581-40f9-449b-ae11-6b18abd0053e-clustermesh-secrets\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794458 kubelet[1933]: I0113 20:35:33.794425 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-config-path\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794458 kubelet[1933]: I0113 20:35:33.794454 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqqmx\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-kube-api-access-vqqmx\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794481 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-run\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794507 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-lib-modules\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794531 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-bpf-maps\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794552 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cni-path\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794574 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-kernel\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794657 kubelet[1933]: I0113 20:35:33.794597 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-hostproc\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794817 kubelet[1933]: I0113 20:35:33.794621 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-xtables-lock\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794817 kubelet[1933]: I0113 20:35:33.794644 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-cgroup\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794817 kubelet[1933]: I0113 20:35:33.794666 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-net\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794817 kubelet[1933]: I0113 20:35:33.794688 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-etc-cni-netd\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.794817 kubelet[1933]: I0113 20:35:33.794713 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-hubble-tls\") pod \"aff58581-40f9-449b-ae11-6b18abd0053e\" (UID: \"aff58581-40f9-449b-ae11-6b18abd0053e\") " Jan 13 20:35:33.798523 kubelet[1933]: I0113 20:35:33.795584 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798523 kubelet[1933]: I0113 20:35:33.795651 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798523 kubelet[1933]: I0113 20:35:33.795673 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798523 kubelet[1933]: I0113 20:35:33.795696 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798523 kubelet[1933]: I0113 20:35:33.795716 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cni-path" (OuterVolumeSpecName: "cni-path") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798832 kubelet[1933]: I0113 20:35:33.795747 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798832 kubelet[1933]: I0113 20:35:33.795770 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-hostproc" (OuterVolumeSpecName: "hostproc") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798832 kubelet[1933]: I0113 20:35:33.795791 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.798685 systemd[1]: var-lib-kubelet-pods-aff58581\x2d40f9\x2d449b\x2dae11\x2d6b18abd0053e-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jan 13 20:35:33.799015 kubelet[1933]: I0113 20:35:33.798894 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aff58581-40f9-449b-ae11-6b18abd0053e-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 13 20:35:33.798889 systemd[1]: var-lib-kubelet-pods-aff58581\x2d40f9\x2d449b\x2dae11\x2d6b18abd0053e-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jan 13 20:35:33.799143 kubelet[1933]: I0113 20:35:33.799029 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 13 20:35:33.799143 kubelet[1933]: I0113 20:35:33.799071 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.799143 kubelet[1933]: I0113 20:35:33.799094 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 13 20:35:33.800894 kubelet[1933]: I0113 20:35:33.800852 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 13 20:35:33.801600 kubelet[1933]: I0113 20:35:33.801571 1933 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-kube-api-access-vqqmx" (OuterVolumeSpecName: "kube-api-access-vqqmx") pod "aff58581-40f9-449b-ae11-6b18abd0053e" (UID: "aff58581-40f9-449b-ae11-6b18abd0053e"). InnerVolumeSpecName "kube-api-access-vqqmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 13 20:35:33.895030 kubelet[1933]: I0113 20:35:33.894965 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-run\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895030 kubelet[1933]: I0113 20:35:33.895017 1933 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-lib-modules\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895030 kubelet[1933]: I0113 20:35:33.895030 1933 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-bpf-maps\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895030 kubelet[1933]: I0113 20:35:33.895045 1933 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cni-path\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895060 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-kernel\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895074 1933 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-hostproc\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895087 1933 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-xtables-lock\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895100 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-cgroup\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895114 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-host-proc-sys-net\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895126 1933 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aff58581-40f9-449b-ae11-6b18abd0053e-etc-cni-netd\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895138 1933 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-hubble-tls\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895261 kubelet[1933]: I0113 20:35:33.895151 1933 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aff58581-40f9-449b-ae11-6b18abd0053e-clustermesh-secrets\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895455 kubelet[1933]: I0113 20:35:33.895164 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aff58581-40f9-449b-ae11-6b18abd0053e-cilium-config-path\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:33.895455 kubelet[1933]: I0113 20:35:33.895178 1933 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-vqqmx\" (UniqueName: \"kubernetes.io/projected/aff58581-40f9-449b-ae11-6b18abd0053e-kube-api-access-vqqmx\") on node \"10.0.0.53\" DevicePath \"\"" Jan 13 20:35:34.310594 kubelet[1933]: I0113 20:35:34.310563 1933 scope.go:117] "RemoveContainer" containerID="0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1" Jan 13 20:35:34.311878 containerd[1584]: time="2025-01-13T20:35:34.311605219Z" level=info msg="RemoveContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\"" Jan 13 20:35:34.315107 containerd[1584]: time="2025-01-13T20:35:34.315085332Z" level=info msg="RemoveContainer for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" returns successfully" Jan 13 20:35:34.315353 kubelet[1933]: I0113 20:35:34.315329 1933 scope.go:117] "RemoveContainer" containerID="1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3" Jan 13 20:35:34.316069 containerd[1584]: time="2025-01-13T20:35:34.316044264Z" level=info msg="RemoveContainer for \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\"" Jan 13 20:35:34.319177 containerd[1584]: time="2025-01-13T20:35:34.319143511Z" level=info msg="RemoveContainer for \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\" returns successfully" Jan 13 20:35:34.319353 kubelet[1933]: I0113 20:35:34.319329 1933 scope.go:117] "RemoveContainer" containerID="5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb" Jan 13 20:35:34.320081 containerd[1584]: time="2025-01-13T20:35:34.320057900Z" level=info msg="RemoveContainer for \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\"" Jan 13 20:35:34.323418 containerd[1584]: time="2025-01-13T20:35:34.323388652Z" level=info msg="RemoveContainer for \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\" returns successfully" Jan 13 20:35:34.323546 kubelet[1933]: I0113 20:35:34.323521 1933 scope.go:117] "RemoveContainer" containerID="de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f" Jan 13 20:35:34.324396 containerd[1584]: time="2025-01-13T20:35:34.324373044Z" level=info msg="RemoveContainer for \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\"" Jan 13 20:35:34.327175 containerd[1584]: time="2025-01-13T20:35:34.327156877Z" level=info msg="RemoveContainer for \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\" returns successfully" Jan 13 20:35:34.327318 kubelet[1933]: I0113 20:35:34.327301 1933 scope.go:117] "RemoveContainer" containerID="9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee" Jan 13 20:35:34.327989 containerd[1584]: time="2025-01-13T20:35:34.327965888Z" level=info msg="RemoveContainer for \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\"" Jan 13 20:35:34.330645 containerd[1584]: time="2025-01-13T20:35:34.330614758Z" level=info msg="RemoveContainer for \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\" returns successfully" Jan 13 20:35:34.330756 kubelet[1933]: I0113 20:35:34.330739 1933 scope.go:117] "RemoveContainer" containerID="0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1" Jan 13 20:35:34.330899 containerd[1584]: time="2025-01-13T20:35:34.330866251Z" level=error msg="ContainerStatus for \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\": not found" Jan 13 20:35:34.331012 kubelet[1933]: E0113 20:35:34.330996 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\": not found" containerID="0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1" Jan 13 20:35:34.331095 kubelet[1933]: I0113 20:35:34.331081 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1"} err="failed to get container status \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\": rpc error: code = NotFound desc = an error occurred when try to find container \"0bf77e6860c4a952417a2330015819112152cf719b65fc6cfa0f52f554ff7bc1\": not found" Jan 13 20:35:34.331133 kubelet[1933]: I0113 20:35:34.331098 1933 scope.go:117] "RemoveContainer" containerID="1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3" Jan 13 20:35:34.331245 containerd[1584]: time="2025-01-13T20:35:34.331220848Z" level=error msg="ContainerStatus for \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\": not found" Jan 13 20:35:34.331356 kubelet[1933]: E0113 20:35:34.331340 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\": not found" containerID="1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3" Jan 13 20:35:34.331404 kubelet[1933]: I0113 20:35:34.331367 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3"} err="failed to get container status \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\": rpc error: code = NotFound desc = an error occurred when try to find container \"1c02fc23fd67e807aaec40d052ab671f8cfff1be74b1d7100fbd878f605ad4f3\": not found" Jan 13 20:35:34.331404 kubelet[1933]: I0113 20:35:34.331378 1933 scope.go:117] "RemoveContainer" containerID="5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb" Jan 13 20:35:34.331530 containerd[1584]: time="2025-01-13T20:35:34.331503929Z" level=error msg="ContainerStatus for \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\": not found" Jan 13 20:35:34.331628 kubelet[1933]: E0113 20:35:34.331616 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\": not found" containerID="5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb" Jan 13 20:35:34.331667 kubelet[1933]: I0113 20:35:34.331636 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb"} err="failed to get container status \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\": rpc error: code = NotFound desc = an error occurred when try to find container \"5f5f53541ca0c172580ce14e3d65ade42e0545cb85af3f3aaba5a10156b3eedb\": not found" Jan 13 20:35:34.331667 kubelet[1933]: I0113 20:35:34.331643 1933 scope.go:117] "RemoveContainer" containerID="de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f" Jan 13 20:35:34.331780 containerd[1584]: time="2025-01-13T20:35:34.331757427Z" level=error msg="ContainerStatus for \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\": not found" Jan 13 20:35:34.331869 kubelet[1933]: E0113 20:35:34.331855 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\": not found" containerID="de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f" Jan 13 20:35:34.331912 kubelet[1933]: I0113 20:35:34.331879 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f"} err="failed to get container status \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\": rpc error: code = NotFound desc = an error occurred when try to find container \"de4a81cae830261e7d55f7f8272f05d15fe6f00ceb90ba33982167c826453b5f\": not found" Jan 13 20:35:34.331912 kubelet[1933]: I0113 20:35:34.331888 1933 scope.go:117] "RemoveContainer" containerID="9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee" Jan 13 20:35:34.332015 containerd[1584]: time="2025-01-13T20:35:34.331993050Z" level=error msg="ContainerStatus for \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\": not found" Jan 13 20:35:34.332112 kubelet[1933]: E0113 20:35:34.332097 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\": not found" containerID="9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee" Jan 13 20:35:34.332152 kubelet[1933]: I0113 20:35:34.332123 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee"} err="failed to get container status \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\": rpc error: code = NotFound desc = an error occurred when try to find container \"9f288b2621207927a083cb1cf53aae3eaa1e4e0b7622f4b05b4b4866cfddc1ee\": not found" Jan 13 20:35:34.412949 kubelet[1933]: E0113 20:35:34.412912 1933 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:34.428674 containerd[1584]: time="2025-01-13T20:35:34.428642998Z" level=info msg="StopPodSandbox for \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\"" Jan 13 20:35:34.428760 containerd[1584]: time="2025-01-13T20:35:34.428740922Z" level=info msg="TearDown network for sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" successfully" Jan 13 20:35:34.428784 containerd[1584]: time="2025-01-13T20:35:34.428757934Z" level=info msg="StopPodSandbox for \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" returns successfully" Jan 13 20:35:34.429061 containerd[1584]: time="2025-01-13T20:35:34.429038001Z" level=info msg="RemovePodSandbox for \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\"" Jan 13 20:35:34.429061 containerd[1584]: time="2025-01-13T20:35:34.429059180Z" level=info msg="Forcibly stopping sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\"" Jan 13 20:35:34.429146 containerd[1584]: time="2025-01-13T20:35:34.429102301Z" level=info msg="TearDown network for sandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" successfully" Jan 13 20:35:34.431974 containerd[1584]: time="2025-01-13T20:35:34.431952781Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 13 20:35:34.432016 containerd[1584]: time="2025-01-13T20:35:34.431981916Z" level=info msg="RemovePodSandbox \"c6b883c9489afc247c5464c11c55663bee3dc83ac670901879e82a6ced2302aa\" returns successfully" Jan 13 20:35:34.440542 systemd[1]: var-lib-kubelet-pods-aff58581\x2d40f9\x2d449b\x2dae11\x2d6b18abd0053e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dvqqmx.mount: Deactivated successfully. Jan 13 20:35:34.476629 kubelet[1933]: E0113 20:35:34.476584 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:35.136199 kubelet[1933]: E0113 20:35:35.136156 1933 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jan 13 20:35:35.140379 kubelet[1933]: I0113 20:35:35.140340 1933 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" path="/var/lib/kubelet/pods/aff58581-40f9-449b-ae11-6b18abd0053e/volumes" Jan 13 20:35:35.477299 kubelet[1933]: E0113 20:35:35.477165 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:35.728970 kubelet[1933]: I0113 20:35:35.728910 1933 topology_manager.go:215] "Topology Admit Handler" podUID="c67645e4-57cc-4a79-8e0d-47c63a814454" podNamespace="kube-system" podName="cilium-bmnpd" Jan 13 20:35:35.728970 kubelet[1933]: E0113 20:35:35.728951 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="mount-bpf-fs" Jan 13 20:35:35.728970 kubelet[1933]: E0113 20:35:35.728963 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="cilium-agent" Jan 13 20:35:35.728970 kubelet[1933]: E0113 20:35:35.728970 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="mount-cgroup" Jan 13 20:35:35.729129 kubelet[1933]: E0113 20:35:35.728978 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="apply-sysctl-overwrites" Jan 13 20:35:35.729129 kubelet[1933]: E0113 20:35:35.728985 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="clean-cilium-state" Jan 13 20:35:35.729129 kubelet[1933]: I0113 20:35:35.729004 1933 memory_manager.go:354] "RemoveStaleState removing state" podUID="aff58581-40f9-449b-ae11-6b18abd0053e" containerName="cilium-agent" Jan 13 20:35:35.729220 kubelet[1933]: I0113 20:35:35.729137 1933 topology_manager.go:215] "Topology Admit Handler" podUID="dc04ff52-4d0d-4c8b-8b52-545bdf31a642" podNamespace="kube-system" podName="cilium-operator-5cc964979-5gnj6" Jan 13 20:35:35.907224 kubelet[1933]: I0113 20:35:35.907146 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c67645e4-57cc-4a79-8e0d-47c63a814454-hubble-tls\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907224 kubelet[1933]: I0113 20:35:35.907208 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-cilium-run\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907224 kubelet[1933]: I0113 20:35:35.907236 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-hostproc\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907468 kubelet[1933]: I0113 20:35:35.907261 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-cilium-cgroup\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907468 kubelet[1933]: I0113 20:35:35.907335 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pmrg\" (UniqueName: \"kubernetes.io/projected/dc04ff52-4d0d-4c8b-8b52-545bdf31a642-kube-api-access-8pmrg\") pod \"cilium-operator-5cc964979-5gnj6\" (UID: \"dc04ff52-4d0d-4c8b-8b52-545bdf31a642\") " pod="kube-system/cilium-operator-5cc964979-5gnj6" Jan 13 20:35:35.907468 kubelet[1933]: I0113 20:35:35.907384 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-host-proc-sys-net\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907468 kubelet[1933]: I0113 20:35:35.907432 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-host-proc-sys-kernel\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907468 kubelet[1933]: I0113 20:35:35.907463 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/dc04ff52-4d0d-4c8b-8b52-545bdf31a642-cilium-config-path\") pod \"cilium-operator-5cc964979-5gnj6\" (UID: \"dc04ff52-4d0d-4c8b-8b52-545bdf31a642\") " pod="kube-system/cilium-operator-5cc964979-5gnj6" Jan 13 20:35:35.907585 kubelet[1933]: I0113 20:35:35.907518 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-cni-path\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907680 kubelet[1933]: I0113 20:35:35.907623 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-etc-cni-netd\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907730 kubelet[1933]: I0113 20:35:35.907713 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-bpf-maps\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907763 kubelet[1933]: I0113 20:35:35.907747 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-xtables-lock\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907799 kubelet[1933]: I0113 20:35:35.907780 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c67645e4-57cc-4a79-8e0d-47c63a814454-cilium-config-path\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907830 kubelet[1933]: I0113 20:35:35.907817 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qrmd\" (UniqueName: \"kubernetes.io/projected/c67645e4-57cc-4a79-8e0d-47c63a814454-kube-api-access-9qrmd\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907869 kubelet[1933]: I0113 20:35:35.907851 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c67645e4-57cc-4a79-8e0d-47c63a814454-clustermesh-secrets\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907905 kubelet[1933]: I0113 20:35:35.907898 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/c67645e4-57cc-4a79-8e0d-47c63a814454-cilium-ipsec-secrets\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:35.907992 kubelet[1933]: I0113 20:35:35.907962 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c67645e4-57cc-4a79-8e0d-47c63a814454-lib-modules\") pod \"cilium-bmnpd\" (UID: \"c67645e4-57cc-4a79-8e0d-47c63a814454\") " pod="kube-system/cilium-bmnpd" Jan 13 20:35:36.032825 kubelet[1933]: E0113 20:35:36.032725 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:36.033310 containerd[1584]: time="2025-01-13T20:35:36.033216276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-5gnj6,Uid:dc04ff52-4d0d-4c8b-8b52-545bdf31a642,Namespace:kube-system,Attempt:0,}" Jan 13 20:35:36.034422 kubelet[1933]: E0113 20:35:36.034393 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:36.034675 containerd[1584]: time="2025-01-13T20:35:36.034644140Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bmnpd,Uid:c67645e4-57cc-4a79-8e0d-47c63a814454,Namespace:kube-system,Attempt:0,}" Jan 13 20:35:36.056978 containerd[1584]: time="2025-01-13T20:35:36.056752822Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:35:36.056978 containerd[1584]: time="2025-01-13T20:35:36.056854674Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:35:36.056978 containerd[1584]: time="2025-01-13T20:35:36.056869272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:36.057158 containerd[1584]: time="2025-01-13T20:35:36.057030084Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:36.058258 containerd[1584]: time="2025-01-13T20:35:36.057972595Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:35:36.058258 containerd[1584]: time="2025-01-13T20:35:36.058030424Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:35:36.058258 containerd[1584]: time="2025-01-13T20:35:36.058048688Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:36.058258 containerd[1584]: time="2025-01-13T20:35:36.058155318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:35:36.102874 containerd[1584]: time="2025-01-13T20:35:36.102766966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bmnpd,Uid:c67645e4-57cc-4a79-8e0d-47c63a814454,Namespace:kube-system,Attempt:0,} returns sandbox id \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\"" Jan 13 20:35:36.103624 kubelet[1933]: E0113 20:35:36.103594 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:36.105763 containerd[1584]: time="2025-01-13T20:35:36.105727912Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 13 20:35:36.113996 kubelet[1933]: I0113 20:35:36.113956 1933 setters.go:568] "Node became not ready" node="10.0.0.53" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-01-13T20:35:36Z","lastTransitionTime":"2025-01-13T20:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jan 13 20:35:36.116132 containerd[1584]: time="2025-01-13T20:35:36.116094946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-5gnj6,Uid:dc04ff52-4d0d-4c8b-8b52-545bdf31a642,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e7c2639c66f8d99d50f6cfae893f814dc324e43fcfffcba00c71c6eab08c15e\"" Jan 13 20:35:36.116767 kubelet[1933]: E0113 20:35:36.116738 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:36.117460 containerd[1584]: time="2025-01-13T20:35:36.117425296Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jan 13 20:35:36.123234 containerd[1584]: time="2025-01-13T20:35:36.123179572Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3940d9bbeb30e21dfb27bba39a0c677e6b1b05663aa3fe96a96255db3afc93ef\"" Jan 13 20:35:36.123909 containerd[1584]: time="2025-01-13T20:35:36.123874748Z" level=info msg="StartContainer for \"3940d9bbeb30e21dfb27bba39a0c677e6b1b05663aa3fe96a96255db3afc93ef\"" Jan 13 20:35:36.177976 containerd[1584]: time="2025-01-13T20:35:36.177916149Z" level=info msg="StartContainer for \"3940d9bbeb30e21dfb27bba39a0c677e6b1b05663aa3fe96a96255db3afc93ef\" returns successfully" Jan 13 20:35:36.222441 containerd[1584]: time="2025-01-13T20:35:36.222362877Z" level=info msg="shim disconnected" id=3940d9bbeb30e21dfb27bba39a0c677e6b1b05663aa3fe96a96255db3afc93ef namespace=k8s.io Jan 13 20:35:36.222441 containerd[1584]: time="2025-01-13T20:35:36.222438168Z" level=warning msg="cleaning up after shim disconnected" id=3940d9bbeb30e21dfb27bba39a0c677e6b1b05663aa3fe96a96255db3afc93ef namespace=k8s.io Jan 13 20:35:36.222441 containerd[1584]: time="2025-01-13T20:35:36.222453206Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:36.316855 kubelet[1933]: E0113 20:35:36.316466 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:36.318595 containerd[1584]: time="2025-01-13T20:35:36.318552966Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 13 20:35:36.336005 containerd[1584]: time="2025-01-13T20:35:36.335936548Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"15753a94b78196fd6828c5f6e94a44e21de378a0405142ae19746a183c3db79b\"" Jan 13 20:35:36.336556 containerd[1584]: time="2025-01-13T20:35:36.336508212Z" level=info msg="StartContainer for \"15753a94b78196fd6828c5f6e94a44e21de378a0405142ae19746a183c3db79b\"" Jan 13 20:35:36.404881 containerd[1584]: time="2025-01-13T20:35:36.404833360Z" level=info msg="StartContainer for \"15753a94b78196fd6828c5f6e94a44e21de378a0405142ae19746a183c3db79b\" returns successfully" Jan 13 20:35:36.439027 containerd[1584]: time="2025-01-13T20:35:36.438950272Z" level=info msg="shim disconnected" id=15753a94b78196fd6828c5f6e94a44e21de378a0405142ae19746a183c3db79b namespace=k8s.io Jan 13 20:35:36.439027 containerd[1584]: time="2025-01-13T20:35:36.439016467Z" level=warning msg="cleaning up after shim disconnected" id=15753a94b78196fd6828c5f6e94a44e21de378a0405142ae19746a183c3db79b namespace=k8s.io Jan 13 20:35:36.439027 containerd[1584]: time="2025-01-13T20:35:36.439028229Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:36.477640 kubelet[1933]: E0113 20:35:36.477571 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:37.321558 kubelet[1933]: E0113 20:35:37.321522 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:37.323951 containerd[1584]: time="2025-01-13T20:35:37.323907916Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 13 20:35:37.341123 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount631918753.mount: Deactivated successfully. Jan 13 20:35:37.344619 containerd[1584]: time="2025-01-13T20:35:37.344568481Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5\"" Jan 13 20:35:37.345169 containerd[1584]: time="2025-01-13T20:35:37.345126530Z" level=info msg="StartContainer for \"dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5\"" Jan 13 20:35:37.419925 containerd[1584]: time="2025-01-13T20:35:37.419842530Z" level=info msg="StartContainer for \"dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5\" returns successfully" Jan 13 20:35:37.463330 containerd[1584]: time="2025-01-13T20:35:37.462975758Z" level=info msg="shim disconnected" id=dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5 namespace=k8s.io Jan 13 20:35:37.463330 containerd[1584]: time="2025-01-13T20:35:37.463060597Z" level=warning msg="cleaning up after shim disconnected" id=dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5 namespace=k8s.io Jan 13 20:35:37.463330 containerd[1584]: time="2025-01-13T20:35:37.463074903Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:37.477954 kubelet[1933]: E0113 20:35:37.477853 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:38.015004 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dedd69552ec834d5abd8d6fb2db90da8531ad4c6bdcc8506172ae17befae9ad5-rootfs.mount: Deactivated successfully. Jan 13 20:35:38.340154 kubelet[1933]: E0113 20:35:38.339880 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:38.342982 containerd[1584]: time="2025-01-13T20:35:38.342907276Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 13 20:35:38.369437 containerd[1584]: time="2025-01-13T20:35:38.369388806Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2\"" Jan 13 20:35:38.369932 containerd[1584]: time="2025-01-13T20:35:38.369903523Z" level=info msg="StartContainer for \"0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2\"" Jan 13 20:35:38.450909 containerd[1584]: time="2025-01-13T20:35:38.450841568Z" level=info msg="StartContainer for \"0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2\" returns successfully" Jan 13 20:35:38.478135 kubelet[1933]: E0113 20:35:38.478061 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:38.489133 containerd[1584]: time="2025-01-13T20:35:38.488995789Z" level=info msg="shim disconnected" id=0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2 namespace=k8s.io Jan 13 20:35:38.489133 containerd[1584]: time="2025-01-13T20:35:38.489059208Z" level=warning msg="cleaning up after shim disconnected" id=0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2 namespace=k8s.io Jan 13 20:35:38.489133 containerd[1584]: time="2025-01-13T20:35:38.489068105Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:35:39.014394 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0cee77a430b218ca160aca11335db8c880abccc5f0e309c4b467a3d2e18951a2-rootfs.mount: Deactivated successfully. Jan 13 20:35:39.374412 kubelet[1933]: E0113 20:35:39.374262 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:39.380060 containerd[1584]: time="2025-01-13T20:35:39.379873729Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 13 20:35:39.478447 kubelet[1933]: E0113 20:35:39.478258 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:39.712298 containerd[1584]: time="2025-01-13T20:35:39.712081070Z" level=info msg="CreateContainer within sandbox \"72f13393c5b169c6bac3eef820b31cf240a0c10f7c2216550b50d0eb40669926\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"85badbb796160319e87879c7958110be6e5b8e7bb314ee6f09c536631959f8ec\"" Jan 13 20:35:39.714019 containerd[1584]: time="2025-01-13T20:35:39.713791514Z" level=info msg="StartContainer for \"85badbb796160319e87879c7958110be6e5b8e7bb314ee6f09c536631959f8ec\"" Jan 13 20:35:39.830167 containerd[1584]: time="2025-01-13T20:35:39.830104710Z" level=info msg="StartContainer for \"85badbb796160319e87879c7958110be6e5b8e7bb314ee6f09c536631959f8ec\" returns successfully" Jan 13 20:35:40.283300 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Jan 13 20:35:40.380536 kubelet[1933]: E0113 20:35:40.380485 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:40.479437 kubelet[1933]: E0113 20:35:40.479386 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:41.480221 kubelet[1933]: E0113 20:35:41.480153 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:42.035793 kubelet[1933]: E0113 20:35:42.035742 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:42.481416 kubelet[1933]: E0113 20:35:42.481346 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:43.481665 kubelet[1933]: E0113 20:35:43.481620 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:43.715508 systemd-networkd[1243]: lxc_health: Link UP Jan 13 20:35:43.728676 systemd-networkd[1243]: lxc_health: Gained carrier Jan 13 20:35:44.036245 kubelet[1933]: E0113 20:35:44.036207 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:44.052513 kubelet[1933]: I0113 20:35:44.052443 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-bmnpd" podStartSLOduration=9.052393985 podStartE2EDuration="9.052393985s" podCreationTimestamp="2025-01-13 20:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:35:40.498712834 +0000 UTC m=+66.430891686" watchObservedRunningTime="2025-01-13 20:35:44.052393985 +0000 UTC m=+69.984572828" Jan 13 20:35:44.393535 kubelet[1933]: E0113 20:35:44.393453 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:44.482353 kubelet[1933]: E0113 20:35:44.482308 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:45.391130 kubelet[1933]: E0113 20:35:45.391096 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:45.482761 kubelet[1933]: E0113 20:35:45.482728 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:45.584813 systemd-networkd[1243]: lxc_health: Gained IPv6LL Jan 13 20:35:46.483159 kubelet[1933]: E0113 20:35:46.483126 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:47.484038 kubelet[1933]: E0113 20:35:47.483998 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:47.945474 containerd[1584]: time="2025-01-13T20:35:47.945403334Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:47.953975 containerd[1584]: time="2025-01-13T20:35:47.953940551Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18906589" Jan 13 20:35:47.972504 containerd[1584]: time="2025-01-13T20:35:47.972469478Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:35:47.973912 containerd[1584]: time="2025-01-13T20:35:47.973872131Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 11.856405458s" Jan 13 20:35:47.973954 containerd[1584]: time="2025-01-13T20:35:47.973917997Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Jan 13 20:35:47.975725 containerd[1584]: time="2025-01-13T20:35:47.975698520Z" level=info msg="CreateContainer within sandbox \"5e7c2639c66f8d99d50f6cfae893f814dc324e43fcfffcba00c71c6eab08c15e\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jan 13 20:35:48.166901 containerd[1584]: time="2025-01-13T20:35:48.166844815Z" level=info msg="CreateContainer within sandbox \"5e7c2639c66f8d99d50f6cfae893f814dc324e43fcfffcba00c71c6eab08c15e\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"ce5556427a5141d2378a2dd429c1676e83b4a3a4edd8970128a2dce3d7bded83\"" Jan 13 20:35:48.167460 containerd[1584]: time="2025-01-13T20:35:48.167397093Z" level=info msg="StartContainer for \"ce5556427a5141d2378a2dd429c1676e83b4a3a4edd8970128a2dce3d7bded83\"" Jan 13 20:35:48.270891 containerd[1584]: time="2025-01-13T20:35:48.270765071Z" level=info msg="StartContainer for \"ce5556427a5141d2378a2dd429c1676e83b4a3a4edd8970128a2dce3d7bded83\" returns successfully" Jan 13 20:35:48.396522 kubelet[1933]: E0113 20:35:48.396487 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:48.451034 kubelet[1933]: I0113 20:35:48.450997 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-5cc964979-5gnj6" podStartSLOduration=1.59386285 podStartE2EDuration="13.45095666s" podCreationTimestamp="2025-01-13 20:35:35 +0000 UTC" firstStartedPulling="2025-01-13 20:35:36.117118921 +0000 UTC m=+62.049297763" lastFinishedPulling="2025-01-13 20:35:47.974212731 +0000 UTC m=+73.906391573" observedRunningTime="2025-01-13 20:35:48.450877642 +0000 UTC m=+74.383056494" watchObservedRunningTime="2025-01-13 20:35:48.45095666 +0000 UTC m=+74.383135502" Jan 13 20:35:48.485035 kubelet[1933]: E0113 20:35:48.484979 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:49.397857 kubelet[1933]: E0113 20:35:49.397813 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 13 20:35:49.486052 kubelet[1933]: E0113 20:35:49.485989 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:50.487226 kubelet[1933]: E0113 20:35:50.486994 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 13 20:35:51.487771 kubelet[1933]: E0113 20:35:51.487702 1933 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"