Jan 30 13:41:54.898490 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Wed Jan 29 10:09:32 -00 2025 Jan 30 13:41:54.898516 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=befc9792b021bef43c896e00e1d5172b6224dbafc9b6c92b267e5e544378e681 Jan 30 13:41:54.898531 kernel: BIOS-provided physical RAM map: Jan 30 13:41:54.898540 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Jan 30 13:41:54.898548 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Jan 30 13:41:54.898556 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 30 13:41:54.898566 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Jan 30 13:41:54.898574 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Jan 30 13:41:54.898602 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jan 30 13:41:54.898612 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jan 30 13:41:54.898619 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 30 13:41:54.898625 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 30 13:41:54.898631 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 30 13:41:54.898637 kernel: NX (Execute Disable) protection: active Jan 30 13:41:54.898645 kernel: APIC: Static calls initialized Jan 30 13:41:54.898654 kernel: SMBIOS 2.8 present. Jan 30 13:41:54.898661 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Jan 30 13:41:54.898667 kernel: Hypervisor detected: KVM Jan 30 13:41:54.898674 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 30 13:41:54.898681 kernel: kvm-clock: using sched offset of 2232501968 cycles Jan 30 13:41:54.898688 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 30 13:41:54.898695 kernel: tsc: Detected 2794.748 MHz processor Jan 30 13:41:54.898702 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 30 13:41:54.898709 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 30 13:41:54.898716 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Jan 30 13:41:54.898725 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 30 13:41:54.898732 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 30 13:41:54.898739 kernel: Using GB pages for direct mapping Jan 30 13:41:54.898746 kernel: ACPI: Early table checksum verification disabled Jan 30 13:41:54.898753 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Jan 30 13:41:54.898760 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898767 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898773 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898791 kernel: ACPI: FACS 0x000000009CFE0000 000040 Jan 30 13:41:54.898799 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898805 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898812 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898819 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 13:41:54.898826 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db] Jan 30 13:41:54.898833 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7] Jan 30 13:41:54.898843 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Jan 30 13:41:54.898853 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b] Jan 30 13:41:54.898860 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3] Jan 30 13:41:54.898867 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df] Jan 30 13:41:54.898874 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407] Jan 30 13:41:54.898881 kernel: No NUMA configuration found Jan 30 13:41:54.898889 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Jan 30 13:41:54.898896 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Jan 30 13:41:54.898905 kernel: Zone ranges: Jan 30 13:41:54.898913 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 30 13:41:54.898920 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Jan 30 13:41:54.898927 kernel: Normal empty Jan 30 13:41:54.898934 kernel: Movable zone start for each node Jan 30 13:41:54.898942 kernel: Early memory node ranges Jan 30 13:41:54.898949 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 30 13:41:54.898956 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Jan 30 13:41:54.898963 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Jan 30 13:41:54.898975 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 30 13:41:54.898985 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 30 13:41:54.898994 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Jan 30 13:41:54.899003 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 30 13:41:54.899012 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 30 13:41:54.899019 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 30 13:41:54.899026 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 30 13:41:54.899035 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 30 13:41:54.899045 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 30 13:41:54.899059 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 30 13:41:54.899069 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 30 13:41:54.899079 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 30 13:41:54.899089 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 30 13:41:54.899099 kernel: TSC deadline timer available Jan 30 13:41:54.899109 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Jan 30 13:41:54.899119 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 30 13:41:54.899130 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 30 13:41:54.899139 kernel: kvm-guest: setup PV sched yield Jan 30 13:41:54.899149 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jan 30 13:41:54.899163 kernel: Booting paravirtualized kernel on KVM Jan 30 13:41:54.899173 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 30 13:41:54.899183 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 30 13:41:54.899193 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Jan 30 13:41:54.899202 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Jan 30 13:41:54.899212 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 30 13:41:54.899221 kernel: kvm-guest: PV spinlocks enabled Jan 30 13:41:54.899231 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 30 13:41:54.899242 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=befc9792b021bef43c896e00e1d5172b6224dbafc9b6c92b267e5e544378e681 Jan 30 13:41:54.899256 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 30 13:41:54.899266 kernel: random: crng init done Jan 30 13:41:54.899276 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 30 13:41:54.899286 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 30 13:41:54.899296 kernel: Fallback order for Node 0: 0 Jan 30 13:41:54.899306 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Jan 30 13:41:54.899315 kernel: Policy zone: DMA32 Jan 30 13:41:54.899323 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 30 13:41:54.899333 kernel: Memory: 2434588K/2571752K available (12288K kernel code, 2301K rwdata, 22728K rodata, 42844K init, 2348K bss, 136904K reserved, 0K cma-reserved) Jan 30 13:41:54.899341 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 30 13:41:54.899348 kernel: ftrace: allocating 37921 entries in 149 pages Jan 30 13:41:54.899355 kernel: ftrace: allocated 149 pages with 4 groups Jan 30 13:41:54.899362 kernel: Dynamic Preempt: voluntary Jan 30 13:41:54.899369 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 30 13:41:54.899377 kernel: rcu: RCU event tracing is enabled. Jan 30 13:41:54.899384 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 30 13:41:54.899392 kernel: Trampoline variant of Tasks RCU enabled. Jan 30 13:41:54.899401 kernel: Rude variant of Tasks RCU enabled. Jan 30 13:41:54.899408 kernel: Tracing variant of Tasks RCU enabled. Jan 30 13:41:54.899416 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 30 13:41:54.899423 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 30 13:41:54.899430 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 30 13:41:54.899437 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 30 13:41:54.899444 kernel: Console: colour VGA+ 80x25 Jan 30 13:41:54.899451 kernel: printk: console [ttyS0] enabled Jan 30 13:41:54.899458 kernel: ACPI: Core revision 20230628 Jan 30 13:41:54.899468 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 30 13:41:54.899475 kernel: APIC: Switch to symmetric I/O mode setup Jan 30 13:41:54.899483 kernel: x2apic enabled Jan 30 13:41:54.899492 kernel: APIC: Switched APIC routing to: physical x2apic Jan 30 13:41:54.899499 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 30 13:41:54.899508 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 30 13:41:54.899516 kernel: kvm-guest: setup PV IPIs Jan 30 13:41:54.899533 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 30 13:41:54.899540 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Jan 30 13:41:54.899548 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Jan 30 13:41:54.899555 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 30 13:41:54.899562 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 30 13:41:54.899572 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 30 13:41:54.899607 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 30 13:41:54.899616 kernel: Spectre V2 : Mitigation: Retpolines Jan 30 13:41:54.899624 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jan 30 13:41:54.899631 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Jan 30 13:41:54.899642 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Jan 30 13:41:54.899649 kernel: RETBleed: Mitigation: untrained return thunk Jan 30 13:41:54.899657 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 30 13:41:54.899665 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 30 13:41:54.899672 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 30 13:41:54.899682 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 30 13:41:54.899692 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 30 13:41:54.899702 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 30 13:41:54.899715 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 30 13:41:54.899725 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 30 13:41:54.899735 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 30 13:41:54.899746 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 30 13:41:54.899756 kernel: Freeing SMP alternatives memory: 32K Jan 30 13:41:54.899767 kernel: pid_max: default: 32768 minimum: 301 Jan 30 13:41:54.899777 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 30 13:41:54.899799 kernel: landlock: Up and running. Jan 30 13:41:54.899810 kernel: SELinux: Initializing. Jan 30 13:41:54.899824 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:41:54.899834 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:41:54.899845 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Jan 30 13:41:54.899855 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 30 13:41:54.899864 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 30 13:41:54.899874 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 30 13:41:54.899884 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jan 30 13:41:54.899894 kernel: ... version: 0 Jan 30 13:41:54.899903 kernel: ... bit width: 48 Jan 30 13:41:54.899917 kernel: ... generic registers: 6 Jan 30 13:41:54.899927 kernel: ... value mask: 0000ffffffffffff Jan 30 13:41:54.899937 kernel: ... max period: 00007fffffffffff Jan 30 13:41:54.899947 kernel: ... fixed-purpose events: 0 Jan 30 13:41:54.899957 kernel: ... event mask: 000000000000003f Jan 30 13:41:54.899967 kernel: signal: max sigframe size: 1776 Jan 30 13:41:54.899978 kernel: rcu: Hierarchical SRCU implementation. Jan 30 13:41:54.899988 kernel: rcu: Max phase no-delay instances is 400. Jan 30 13:41:54.899999 kernel: smp: Bringing up secondary CPUs ... Jan 30 13:41:54.900013 kernel: smpboot: x86: Booting SMP configuration: Jan 30 13:41:54.900024 kernel: .... node #0, CPUs: #1 #2 #3 Jan 30 13:41:54.900032 kernel: smp: Brought up 1 node, 4 CPUs Jan 30 13:41:54.900040 kernel: smpboot: Max logical packages: 1 Jan 30 13:41:54.900047 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Jan 30 13:41:54.900055 kernel: devtmpfs: initialized Jan 30 13:41:54.900062 kernel: x86/mm: Memory block size: 128MB Jan 30 13:41:54.900070 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 30 13:41:54.900077 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 30 13:41:54.900087 kernel: pinctrl core: initialized pinctrl subsystem Jan 30 13:41:54.900095 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 30 13:41:54.900102 kernel: audit: initializing netlink subsys (disabled) Jan 30 13:41:54.900110 kernel: audit: type=2000 audit(1738244514.594:1): state=initialized audit_enabled=0 res=1 Jan 30 13:41:54.900117 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 30 13:41:54.900124 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 30 13:41:54.900132 kernel: cpuidle: using governor menu Jan 30 13:41:54.900139 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 30 13:41:54.900147 kernel: dca service started, version 1.12.1 Jan 30 13:41:54.900156 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Jan 30 13:41:54.900164 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jan 30 13:41:54.900171 kernel: PCI: Using configuration type 1 for base access Jan 30 13:41:54.900179 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 30 13:41:54.900186 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 30 13:41:54.900193 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 30 13:41:54.900201 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 30 13:41:54.900208 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 30 13:41:54.900215 kernel: ACPI: Added _OSI(Module Device) Jan 30 13:41:54.900225 kernel: ACPI: Added _OSI(Processor Device) Jan 30 13:41:54.900232 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 30 13:41:54.900239 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 30 13:41:54.900247 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 30 13:41:54.900254 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 30 13:41:54.900262 kernel: ACPI: Interpreter enabled Jan 30 13:41:54.900269 kernel: ACPI: PM: (supports S0 S3 S5) Jan 30 13:41:54.900276 kernel: ACPI: Using IOAPIC for interrupt routing Jan 30 13:41:54.900284 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 30 13:41:54.900293 kernel: PCI: Using E820 reservations for host bridge windows Jan 30 13:41:54.900301 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 30 13:41:54.900308 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 30 13:41:54.900489 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 30 13:41:54.900675 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 30 13:41:54.900818 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 30 13:41:54.900828 kernel: PCI host bridge to bus 0000:00 Jan 30 13:41:54.900956 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 30 13:41:54.901077 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 30 13:41:54.901212 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 30 13:41:54.901354 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Jan 30 13:41:54.901490 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 30 13:41:54.901633 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Jan 30 13:41:54.901807 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 30 13:41:54.902013 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Jan 30 13:41:54.902180 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Jan 30 13:41:54.902306 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Jan 30 13:41:54.902427 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Jan 30 13:41:54.902546 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Jan 30 13:41:54.902697 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 30 13:41:54.902876 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Jan 30 13:41:54.903088 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Jan 30 13:41:54.903263 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Jan 30 13:41:54.903431 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Jan 30 13:41:54.903652 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Jan 30 13:41:54.903809 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Jan 30 13:41:54.903933 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Jan 30 13:41:54.904068 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Jan 30 13:41:54.904220 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Jan 30 13:41:54.904377 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Jan 30 13:41:54.904523 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Jan 30 13:41:54.904678 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Jan 30 13:41:54.904850 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Jan 30 13:41:54.905047 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Jan 30 13:41:54.905207 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 30 13:41:54.905359 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Jan 30 13:41:54.905481 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Jan 30 13:41:54.905655 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Jan 30 13:41:54.905836 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Jan 30 13:41:54.906017 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Jan 30 13:41:54.906032 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 30 13:41:54.906044 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 30 13:41:54.906057 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 30 13:41:54.906065 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 30 13:41:54.906072 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 30 13:41:54.906080 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 30 13:41:54.906088 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 30 13:41:54.906096 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 30 13:41:54.906103 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 30 13:41:54.906111 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 30 13:41:54.906121 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 30 13:41:54.906129 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 30 13:41:54.906137 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 30 13:41:54.906145 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 30 13:41:54.906152 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 30 13:41:54.906160 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 30 13:41:54.906168 kernel: iommu: Default domain type: Translated Jan 30 13:41:54.906176 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 30 13:41:54.906184 kernel: PCI: Using ACPI for IRQ routing Jan 30 13:41:54.906194 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 30 13:41:54.906202 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Jan 30 13:41:54.906210 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Jan 30 13:41:54.906335 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 30 13:41:54.906492 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 30 13:41:54.906683 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 30 13:41:54.906696 kernel: vgaarb: loaded Jan 30 13:41:54.906703 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 30 13:41:54.906715 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 30 13:41:54.906723 kernel: clocksource: Switched to clocksource kvm-clock Jan 30 13:41:54.906731 kernel: VFS: Disk quotas dquot_6.6.0 Jan 30 13:41:54.906739 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 30 13:41:54.906746 kernel: pnp: PnP ACPI init Jan 30 13:41:54.906886 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Jan 30 13:41:54.906898 kernel: pnp: PnP ACPI: found 6 devices Jan 30 13:41:54.906906 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 30 13:41:54.906917 kernel: NET: Registered PF_INET protocol family Jan 30 13:41:54.906925 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 30 13:41:54.906933 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 30 13:41:54.906940 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 30 13:41:54.906948 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 30 13:41:54.906956 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 30 13:41:54.906963 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 30 13:41:54.906973 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:41:54.906984 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:41:54.906998 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 30 13:41:54.907006 kernel: NET: Registered PF_XDP protocol family Jan 30 13:41:54.907120 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 30 13:41:54.907241 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 30 13:41:54.907381 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 30 13:41:54.907523 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Jan 30 13:41:54.907716 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jan 30 13:41:54.907838 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Jan 30 13:41:54.907853 kernel: PCI: CLS 0 bytes, default 64 Jan 30 13:41:54.907861 kernel: Initialise system trusted keyrings Jan 30 13:41:54.907869 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 30 13:41:54.907876 kernel: Key type asymmetric registered Jan 30 13:41:54.907884 kernel: Asymmetric key parser 'x509' registered Jan 30 13:41:54.907892 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 30 13:41:54.907899 kernel: io scheduler mq-deadline registered Jan 30 13:41:54.907907 kernel: io scheduler kyber registered Jan 30 13:41:54.907914 kernel: io scheduler bfq registered Jan 30 13:41:54.907924 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 30 13:41:54.907934 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 30 13:41:54.907944 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 30 13:41:54.907955 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 30 13:41:54.907966 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 30 13:41:54.907977 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 30 13:41:54.907987 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 30 13:41:54.907997 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 30 13:41:54.908007 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 30 13:41:54.908170 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 30 13:41:54.908316 kernel: rtc_cmos 00:04: registered as rtc0 Jan 30 13:41:54.908328 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 30 13:41:54.908439 kernel: rtc_cmos 00:04: setting system clock to 2025-01-30T13:41:54 UTC (1738244514) Jan 30 13:41:54.908552 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jan 30 13:41:54.908561 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 30 13:41:54.908569 kernel: NET: Registered PF_INET6 protocol family Jan 30 13:41:54.908591 kernel: Segment Routing with IPv6 Jan 30 13:41:54.908603 kernel: In-situ OAM (IOAM) with IPv6 Jan 30 13:41:54.908611 kernel: NET: Registered PF_PACKET protocol family Jan 30 13:41:54.908618 kernel: Key type dns_resolver registered Jan 30 13:41:54.908626 kernel: IPI shorthand broadcast: enabled Jan 30 13:41:54.908634 kernel: sched_clock: Marking stable (565002509, 111799942)->(727404634, -50602183) Jan 30 13:41:54.908642 kernel: registered taskstats version 1 Jan 30 13:41:54.908650 kernel: Loading compiled-in X.509 certificates Jan 30 13:41:54.908658 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 1efdcbe72fc44d29e4e6411cf9a3e64046be4375' Jan 30 13:41:54.908667 kernel: Key type .fscrypt registered Jan 30 13:41:54.908680 kernel: Key type fscrypt-provisioning registered Jan 30 13:41:54.908691 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 30 13:41:54.908701 kernel: ima: Allocated hash algorithm: sha1 Jan 30 13:41:54.908711 kernel: ima: No architecture policies found Jan 30 13:41:54.908721 kernel: clk: Disabling unused clocks Jan 30 13:41:54.908732 kernel: Freeing unused kernel image (initmem) memory: 42844K Jan 30 13:41:54.908743 kernel: Write protecting the kernel read-only data: 36864k Jan 30 13:41:54.908753 kernel: Freeing unused kernel image (rodata/data gap) memory: 1848K Jan 30 13:41:54.908764 kernel: Run /init as init process Jan 30 13:41:54.908779 kernel: with arguments: Jan 30 13:41:54.908800 kernel: /init Jan 30 13:41:54.908810 kernel: with environment: Jan 30 13:41:54.908820 kernel: HOME=/ Jan 30 13:41:54.908830 kernel: TERM=linux Jan 30 13:41:54.908840 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 30 13:41:54.908852 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 13:41:54.908864 systemd[1]: Detected virtualization kvm. Jan 30 13:41:54.908879 systemd[1]: Detected architecture x86-64. Jan 30 13:41:54.908890 systemd[1]: Running in initrd. Jan 30 13:41:54.908900 systemd[1]: No hostname configured, using default hostname. Jan 30 13:41:54.908911 systemd[1]: Hostname set to . Jan 30 13:41:54.908922 systemd[1]: Initializing machine ID from VM UUID. Jan 30 13:41:54.908932 systemd[1]: Queued start job for default target initrd.target. Jan 30 13:41:54.908943 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:41:54.908955 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:41:54.908971 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 30 13:41:54.908997 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 13:41:54.909011 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 30 13:41:54.909023 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 30 13:41:54.909033 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 30 13:41:54.909044 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 30 13:41:54.909053 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:41:54.909061 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:41:54.909069 systemd[1]: Reached target paths.target - Path Units. Jan 30 13:41:54.909078 systemd[1]: Reached target slices.target - Slice Units. Jan 30 13:41:54.909086 systemd[1]: Reached target swap.target - Swaps. Jan 30 13:41:54.909095 systemd[1]: Reached target timers.target - Timer Units. Jan 30 13:41:54.909103 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:41:54.909114 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:41:54.909122 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 30 13:41:54.909131 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 30 13:41:54.909139 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:41:54.909147 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 13:41:54.909156 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:41:54.909164 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 13:41:54.909173 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 30 13:41:54.909181 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 13:41:54.909192 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 30 13:41:54.909200 systemd[1]: Starting systemd-fsck-usr.service... Jan 30 13:41:54.909209 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 13:41:54.909217 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 13:41:54.909226 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:41:54.909234 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 30 13:41:54.909244 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:41:54.909253 systemd[1]: Finished systemd-fsck-usr.service. Jan 30 13:41:54.909284 systemd-journald[192]: Collecting audit messages is disabled. Jan 30 13:41:54.909306 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 13:41:54.909317 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:41:54.909326 systemd-journald[192]: Journal started Jan 30 13:41:54.909346 systemd-journald[192]: Runtime Journal (/run/log/journal/b292f1083fb9452b857790fff709c5b7) is 6.0M, max 48.4M, 42.3M free. Jan 30 13:41:54.903605 systemd-modules-load[193]: Inserted module 'overlay' Jan 30 13:41:54.940300 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 30 13:41:54.940317 kernel: Bridge firewalling registered Jan 30 13:41:54.932209 systemd-modules-load[193]: Inserted module 'br_netfilter' Jan 30 13:41:54.943005 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 13:41:54.943394 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 13:41:54.945725 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:41:54.968767 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:41:54.972062 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:41:54.974728 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 13:41:54.977760 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 13:41:54.986344 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:41:54.987838 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:41:54.992511 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 30 13:41:54.994713 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:41:54.997048 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:41:55.001758 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 13:41:55.009239 dracut-cmdline[226]: dracut-dracut-053 Jan 30 13:41:55.012797 dracut-cmdline[226]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=befc9792b021bef43c896e00e1d5172b6224dbafc9b6c92b267e5e544378e681 Jan 30 13:41:55.035801 systemd-resolved[232]: Positive Trust Anchors: Jan 30 13:41:55.035816 systemd-resolved[232]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 13:41:55.035847 systemd-resolved[232]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 13:41:55.038278 systemd-resolved[232]: Defaulting to hostname 'linux'. Jan 30 13:41:55.039316 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 13:41:55.044638 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:41:55.109612 kernel: SCSI subsystem initialized Jan 30 13:41:55.119604 kernel: Loading iSCSI transport class v2.0-870. Jan 30 13:41:55.129608 kernel: iscsi: registered transport (tcp) Jan 30 13:41:55.151600 kernel: iscsi: registered transport (qla4xxx) Jan 30 13:41:55.151629 kernel: QLogic iSCSI HBA Driver Jan 30 13:41:55.200421 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 30 13:41:55.211705 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 30 13:41:55.236077 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 30 13:41:55.236123 kernel: device-mapper: uevent: version 1.0.3 Jan 30 13:41:55.237118 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 30 13:41:55.278626 kernel: raid6: avx2x4 gen() 29821 MB/s Jan 30 13:41:55.295605 kernel: raid6: avx2x2 gen() 22479 MB/s Jan 30 13:41:55.312707 kernel: raid6: avx2x1 gen() 23110 MB/s Jan 30 13:41:55.312750 kernel: raid6: using algorithm avx2x4 gen() 29821 MB/s Jan 30 13:41:55.330694 kernel: raid6: .... xor() 8083 MB/s, rmw enabled Jan 30 13:41:55.330747 kernel: raid6: using avx2x2 recovery algorithm Jan 30 13:41:55.355607 kernel: xor: automatically using best checksumming function avx Jan 30 13:41:55.552616 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 30 13:41:55.564717 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:41:55.585809 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:41:55.598436 systemd-udevd[413]: Using default interface naming scheme 'v255'. Jan 30 13:41:55.603961 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:41:55.624814 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 30 13:41:55.639597 dracut-pre-trigger[422]: rd.md=0: removing MD RAID activation Jan 30 13:41:55.675420 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:41:55.687750 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 13:41:55.754110 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:41:55.762459 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 30 13:41:55.774900 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 30 13:41:55.777970 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:41:55.780772 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:41:55.783184 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 13:41:55.787644 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 30 13:41:55.813470 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Jan 30 13:41:55.813663 kernel: cryptd: max_cpu_qlen set to 1000 Jan 30 13:41:55.813675 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 30 13:41:55.813686 kernel: GPT:9289727 != 19775487 Jan 30 13:41:55.813697 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 30 13:41:55.813707 kernel: GPT:9289727 != 19775487 Jan 30 13:41:55.813725 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 30 13:41:55.813735 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 30 13:41:55.813746 kernel: AVX2 version of gcm_enc/dec engaged. Jan 30 13:41:55.813756 kernel: AES CTR mode by8 optimization enabled Jan 30 13:41:55.791743 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 30 13:41:55.809888 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:41:55.829595 kernel: libata version 3.00 loaded. Jan 30 13:41:55.832513 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:41:55.835659 kernel: ahci 0000:00:1f.2: version 3.0 Jan 30 13:41:55.865505 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 30 13:41:55.865523 kernel: BTRFS: device fsid 64bb5b5a-85cc-41cc-a02b-2cfaa3e93b0a devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (462) Jan 30 13:41:55.865540 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Jan 30 13:41:55.865922 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 30 13:41:55.866062 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (476) Jan 30 13:41:55.866074 kernel: scsi host0: ahci Jan 30 13:41:55.866250 kernel: scsi host1: ahci Jan 30 13:41:55.866397 kernel: scsi host2: ahci Jan 30 13:41:55.866573 kernel: scsi host3: ahci Jan 30 13:41:55.866754 kernel: scsi host4: ahci Jan 30 13:41:55.866948 kernel: scsi host5: ahci Jan 30 13:41:55.867108 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Jan 30 13:41:55.867121 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Jan 30 13:41:55.867131 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Jan 30 13:41:55.867142 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Jan 30 13:41:55.867152 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Jan 30 13:41:55.867167 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Jan 30 13:41:55.833451 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:41:55.840310 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:41:55.844063 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:41:55.844217 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:41:55.847177 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:41:55.862324 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:41:55.874606 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 30 13:41:55.883025 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 30 13:41:55.890781 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 30 13:41:55.918314 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jan 30 13:41:55.923369 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 30 13:41:55.939775 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 30 13:41:55.941279 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:41:55.944953 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:41:55.963018 disk-uuid[556]: Primary Header is updated. Jan 30 13:41:55.963018 disk-uuid[556]: Secondary Entries is updated. Jan 30 13:41:55.963018 disk-uuid[556]: Secondary Header is updated. Jan 30 13:41:55.966607 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 30 13:41:55.966804 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:41:55.970605 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 30 13:41:55.974618 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 30 13:41:55.976599 kernel: block device autoloading is deprecated and will be removed. Jan 30 13:41:56.174467 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 30 13:41:56.174536 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 30 13:41:56.174547 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 30 13:41:56.174557 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 30 13:41:56.175605 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 30 13:41:56.176605 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 30 13:41:56.177821 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 30 13:41:56.177849 kernel: ata3.00: applying bridge limits Jan 30 13:41:56.178611 kernel: ata3.00: configured for UDMA/100 Jan 30 13:41:56.179604 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 30 13:41:56.223130 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 30 13:41:56.235231 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 30 13:41:56.235245 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 30 13:41:56.976571 disk-uuid[567]: The operation has completed successfully. Jan 30 13:41:56.977878 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 30 13:41:57.009712 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 30 13:41:57.009877 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 30 13:41:57.044870 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 30 13:41:57.048300 sh[595]: Success Jan 30 13:41:57.061644 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Jan 30 13:41:57.097966 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 30 13:41:57.111998 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 30 13:41:57.117402 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 30 13:41:57.126898 kernel: BTRFS info (device dm-0): first mount of filesystem 64bb5b5a-85cc-41cc-a02b-2cfaa3e93b0a Jan 30 13:41:57.126924 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 30 13:41:57.126935 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 30 13:41:57.128766 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 30 13:41:57.128784 kernel: BTRFS info (device dm-0): using free space tree Jan 30 13:41:57.134532 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 30 13:41:57.135129 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 30 13:41:57.144865 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 30 13:41:57.148115 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 30 13:41:57.156679 kernel: BTRFS info (device vda6): first mount of filesystem aa75aabd-8755-4402-b4b6-23093345fe03 Jan 30 13:41:57.156740 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 30 13:41:57.156757 kernel: BTRFS info (device vda6): using free space tree Jan 30 13:41:57.160806 kernel: BTRFS info (device vda6): auto enabling async discard Jan 30 13:41:57.170421 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 30 13:41:57.172481 kernel: BTRFS info (device vda6): last unmount of filesystem aa75aabd-8755-4402-b4b6-23093345fe03 Jan 30 13:41:57.181955 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 30 13:41:57.189818 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 30 13:41:57.244225 ignition[689]: Ignition 2.19.0 Jan 30 13:41:57.244240 ignition[689]: Stage: fetch-offline Jan 30 13:41:57.244285 ignition[689]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:57.244298 ignition[689]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:57.244404 ignition[689]: parsed url from cmdline: "" Jan 30 13:41:57.244412 ignition[689]: no config URL provided Jan 30 13:41:57.244418 ignition[689]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 13:41:57.244431 ignition[689]: no config at "/usr/lib/ignition/user.ign" Jan 30 13:41:57.244463 ignition[689]: op(1): [started] loading QEMU firmware config module Jan 30 13:41:57.244470 ignition[689]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 30 13:41:57.256922 ignition[689]: op(1): [finished] loading QEMU firmware config module Jan 30 13:41:57.258548 ignition[689]: parsing config with SHA512: cb22d1e3aba392b141685b1e4bd730cd49ebdc2ba100c8993ceb65eb93f994c340faef6857d0ff0d37fa99d3149db07661bb79846d61cad0adf223758c3b3f16 Jan 30 13:41:57.261486 unknown[689]: fetched base config from "system" Jan 30 13:41:57.261503 unknown[689]: fetched user config from "qemu" Jan 30 13:41:57.262442 ignition[689]: fetch-offline: fetch-offline passed Jan 30 13:41:57.262505 ignition[689]: Ignition finished successfully Jan 30 13:41:57.265043 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:41:57.268570 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:41:57.280790 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 13:41:57.302704 systemd-networkd[785]: lo: Link UP Jan 30 13:41:57.302715 systemd-networkd[785]: lo: Gained carrier Jan 30 13:41:57.304258 systemd-networkd[785]: Enumeration completed Jan 30 13:41:57.304428 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 13:41:57.304647 systemd-networkd[785]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:41:57.304651 systemd-networkd[785]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:41:57.305964 systemd-networkd[785]: eth0: Link UP Jan 30 13:41:57.305967 systemd-networkd[785]: eth0: Gained carrier Jan 30 13:41:57.305974 systemd-networkd[785]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:41:57.306227 systemd[1]: Reached target network.target - Network. Jan 30 13:41:57.307895 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 30 13:41:57.314709 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 30 13:41:57.319633 systemd-networkd[785]: eth0: DHCPv4 address 10.0.0.28/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 30 13:41:57.328545 ignition[787]: Ignition 2.19.0 Jan 30 13:41:57.328556 ignition[787]: Stage: kargs Jan 30 13:41:57.328773 ignition[787]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:57.328786 ignition[787]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:57.329560 ignition[787]: kargs: kargs passed Jan 30 13:41:57.329632 ignition[787]: Ignition finished successfully Jan 30 13:41:57.333353 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 30 13:41:57.349792 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 30 13:41:57.361560 ignition[796]: Ignition 2.19.0 Jan 30 13:41:57.361570 ignition[796]: Stage: disks Jan 30 13:41:57.361779 ignition[796]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:57.361793 ignition[796]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:57.364932 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 30 13:41:57.362419 ignition[796]: disks: disks passed Jan 30 13:41:57.367052 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 30 13:41:57.362461 ignition[796]: Ignition finished successfully Jan 30 13:41:57.369158 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 30 13:41:57.371461 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 13:41:57.371893 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 13:41:57.372232 systemd[1]: Reached target basic.target - Basic System. Jan 30 13:41:57.381792 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 30 13:41:57.393533 systemd-resolved[232]: Detected conflict on linux IN A 10.0.0.28 Jan 30 13:41:57.393550 systemd-resolved[232]: Hostname conflict, changing published hostname from 'linux' to 'linux6'. Jan 30 13:41:57.395315 systemd-fsck[807]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jan 30 13:41:57.402788 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 30 13:41:57.414735 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 30 13:41:57.520592 kernel: EXT4-fs (vda9): mounted filesystem 9f41abed-fd12-4e57-bcd4-5c0ef7f8a1bf r/w with ordered data mode. Quota mode: none. Jan 30 13:41:57.521019 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 30 13:41:57.522570 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 30 13:41:57.538704 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:41:57.540556 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 30 13:41:57.545649 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (815) Jan 30 13:41:57.542255 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 30 13:41:57.551182 kernel: BTRFS info (device vda6): first mount of filesystem aa75aabd-8755-4402-b4b6-23093345fe03 Jan 30 13:41:57.551212 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 30 13:41:57.551226 kernel: BTRFS info (device vda6): using free space tree Jan 30 13:41:57.551239 kernel: BTRFS info (device vda6): auto enabling async discard Jan 30 13:41:57.542308 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 30 13:41:57.542337 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:41:57.549546 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 30 13:41:57.552082 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 30 13:41:57.557861 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:41:57.612611 initrd-setup-root[839]: cut: /sysroot/etc/passwd: No such file or directory Jan 30 13:41:57.617824 initrd-setup-root[846]: cut: /sysroot/etc/group: No such file or directory Jan 30 13:41:57.622799 initrd-setup-root[853]: cut: /sysroot/etc/shadow: No such file or directory Jan 30 13:41:57.627896 initrd-setup-root[860]: cut: /sysroot/etc/gshadow: No such file or directory Jan 30 13:41:57.709405 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 30 13:41:57.717692 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 30 13:41:57.718587 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 30 13:41:57.729635 kernel: BTRFS info (device vda6): last unmount of filesystem aa75aabd-8755-4402-b4b6-23093345fe03 Jan 30 13:41:57.741895 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 30 13:41:57.818104 ignition[932]: INFO : Ignition 2.19.0 Jan 30 13:41:57.818104 ignition[932]: INFO : Stage: mount Jan 30 13:41:57.819999 ignition[932]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:57.819999 ignition[932]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:57.819999 ignition[932]: INFO : mount: mount passed Jan 30 13:41:57.819999 ignition[932]: INFO : Ignition finished successfully Jan 30 13:41:57.825496 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 30 13:41:57.836748 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 30 13:41:58.126279 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 30 13:41:58.141850 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:41:58.150600 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (941) Jan 30 13:41:58.150624 kernel: BTRFS info (device vda6): first mount of filesystem aa75aabd-8755-4402-b4b6-23093345fe03 Jan 30 13:41:58.150635 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 30 13:41:58.152069 kernel: BTRFS info (device vda6): using free space tree Jan 30 13:41:58.154606 kernel: BTRFS info (device vda6): auto enabling async discard Jan 30 13:41:58.156040 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:41:58.177107 ignition[958]: INFO : Ignition 2.19.0 Jan 30 13:41:58.177107 ignition[958]: INFO : Stage: files Jan 30 13:41:58.186330 ignition[958]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:58.186330 ignition[958]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:58.186330 ignition[958]: DEBUG : files: compiled without relabeling support, skipping Jan 30 13:41:58.186330 ignition[958]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 30 13:41:58.186330 ignition[958]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 30 13:41:58.193351 ignition[958]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 30 13:41:58.193351 ignition[958]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 30 13:41:58.193351 ignition[958]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 30 13:41:58.193351 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Jan 30 13:41:58.187508 unknown[958]: wrote ssh authorized keys file for user: core Jan 30 13:41:58.566412 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Jan 30 13:41:58.882286 ignition[958]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 30 13:41:58.882286 ignition[958]: INFO : files: op(8): [started] processing unit "containerd.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(8): op(9): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(8): op(9): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(8): [finished] processing unit "containerd.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(a): [started] processing unit "coreos-metadata.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(a): op(b): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(a): op(b): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(a): [finished] processing unit "coreos-metadata.service" Jan 30 13:41:58.886601 ignition[958]: INFO : files: op(c): [started] setting preset to disabled for "coreos-metadata.service" Jan 30 13:41:58.905419 ignition[958]: INFO : files: op(c): op(d): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 30 13:41:58.910938 ignition[958]: INFO : files: op(c): op(d): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 30 13:41:58.912537 ignition[958]: INFO : files: op(c): [finished] setting preset to disabled for "coreos-metadata.service" Jan 30 13:41:58.912537 ignition[958]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:41:58.912537 ignition[958]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:41:58.912537 ignition[958]: INFO : files: files passed Jan 30 13:41:58.912537 ignition[958]: INFO : Ignition finished successfully Jan 30 13:41:58.914124 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 30 13:41:58.921716 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 30 13:41:58.924522 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 30 13:41:58.926830 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 30 13:41:58.926992 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 30 13:41:58.935358 initrd-setup-root-after-ignition[986]: grep: /sysroot/oem/oem-release: No such file or directory Jan 30 13:41:58.938285 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:41:58.938285 initrd-setup-root-after-ignition[988]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:41:58.941419 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:41:58.941009 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:41:58.942850 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 30 13:41:58.954756 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 30 13:41:58.982285 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 30 13:41:58.982415 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 30 13:41:58.983087 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 30 13:41:58.986220 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 30 13:41:58.986615 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 30 13:41:58.987515 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 30 13:41:59.008793 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:41:59.029744 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 30 13:41:59.041937 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:41:59.042271 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:41:59.044949 systemd[1]: Stopped target timers.target - Timer Units. Jan 30 13:41:59.045295 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 30 13:41:59.045399 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:41:59.049434 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 30 13:41:59.049958 systemd[1]: Stopped target basic.target - Basic System. Jan 30 13:41:59.050289 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 30 13:41:59.050644 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:41:59.051140 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 30 13:41:59.051480 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 30 13:41:59.051991 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:41:59.052338 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 30 13:41:59.052687 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 30 13:41:59.053007 systemd[1]: Stopped target swap.target - Swaps. Jan 30 13:41:59.053320 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 30 13:41:59.053430 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:41:59.071845 systemd-networkd[785]: eth0: Gained IPv6LL Jan 30 13:41:59.072331 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:41:59.074378 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:41:59.075089 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 30 13:41:59.075241 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:41:59.077603 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 30 13:41:59.077727 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 30 13:41:59.078436 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 30 13:41:59.078544 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:41:59.083015 systemd[1]: Stopped target paths.target - Path Units. Jan 30 13:41:59.083255 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 30 13:41:59.083421 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:41:59.086904 systemd[1]: Stopped target slices.target - Slice Units. Jan 30 13:41:59.087245 systemd[1]: Stopped target sockets.target - Socket Units. Jan 30 13:41:59.087599 systemd[1]: iscsid.socket: Deactivated successfully. Jan 30 13:41:59.087699 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:41:59.088125 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 30 13:41:59.088213 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:41:59.094177 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 30 13:41:59.094289 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:41:59.096557 systemd[1]: ignition-files.service: Deactivated successfully. Jan 30 13:41:59.096686 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 30 13:41:59.103737 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 30 13:41:59.104111 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 30 13:41:59.104213 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:41:59.105175 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 30 13:41:59.111292 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 30 13:41:59.117775 ignition[1012]: INFO : Ignition 2.19.0 Jan 30 13:41:59.117775 ignition[1012]: INFO : Stage: umount Jan 30 13:41:59.117775 ignition[1012]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:41:59.117775 ignition[1012]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 30 13:41:59.117775 ignition[1012]: INFO : umount: umount passed Jan 30 13:41:59.117775 ignition[1012]: INFO : Ignition finished successfully Jan 30 13:41:59.113604 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:41:59.117882 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 30 13:41:59.118736 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:41:59.129969 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 30 13:41:59.130116 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 30 13:41:59.134305 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 30 13:41:59.137641 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 30 13:41:59.138748 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 30 13:41:59.142044 systemd[1]: Stopped target network.target - Network. Jan 30 13:41:59.143780 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 30 13:41:59.144719 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 30 13:41:59.146673 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 30 13:41:59.146726 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 30 13:41:59.149761 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 30 13:41:59.149813 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 30 13:41:59.152813 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 30 13:41:59.153856 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 30 13:41:59.156369 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 30 13:41:59.159040 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 30 13:41:59.162620 systemd-networkd[785]: eth0: DHCPv6 lease lost Jan 30 13:41:59.164599 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 30 13:41:59.165834 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 30 13:41:59.168679 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 30 13:41:59.169963 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:41:59.181703 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 30 13:41:59.182122 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 30 13:41:59.182179 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:41:59.183129 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:41:59.187146 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 30 13:41:59.187263 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 30 13:41:59.192113 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 13:41:59.192190 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:41:59.193223 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 30 13:41:59.193311 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 30 13:41:59.195123 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 30 13:41:59.195185 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:41:59.214176 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 30 13:41:59.214428 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:41:59.216641 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 30 13:41:59.216728 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 30 13:41:59.218070 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 30 13:41:59.218124 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:41:59.218352 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 30 13:41:59.218410 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:41:59.219223 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 30 13:41:59.219280 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 30 13:41:59.219872 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:41:59.219927 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:41:59.221478 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 30 13:41:59.233493 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 30 13:41:59.233571 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:41:59.234052 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:41:59.234112 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:41:59.240847 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 30 13:41:59.240969 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 30 13:41:59.248385 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 30 13:41:59.248512 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 30 13:41:59.478823 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 30 13:41:59.478978 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 30 13:41:59.480142 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 30 13:41:59.482023 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 30 13:41:59.482083 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 30 13:41:59.488707 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 30 13:41:59.497188 systemd[1]: Switching root. Jan 30 13:41:59.525737 systemd-journald[192]: Journal stopped Jan 30 13:42:00.936249 systemd-journald[192]: Received SIGTERM from PID 1 (systemd). Jan 30 13:42:00.936328 kernel: SELinux: policy capability network_peer_controls=1 Jan 30 13:42:00.936342 kernel: SELinux: policy capability open_perms=1 Jan 30 13:42:00.936353 kernel: SELinux: policy capability extended_socket_class=1 Jan 30 13:42:00.936365 kernel: SELinux: policy capability always_check_network=0 Jan 30 13:42:00.936379 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 30 13:42:00.936391 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 30 13:42:00.936401 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 30 13:42:00.936414 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 30 13:42:00.936426 kernel: audit: type=1403 audit(1738244520.232:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 30 13:42:00.936438 systemd[1]: Successfully loaded SELinux policy in 43.655ms. Jan 30 13:42:00.936461 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.227ms. Jan 30 13:42:00.936476 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 13:42:00.936493 systemd[1]: Detected virtualization kvm. Jan 30 13:42:00.936504 systemd[1]: Detected architecture x86-64. Jan 30 13:42:00.936516 systemd[1]: Detected first boot. Jan 30 13:42:00.936527 systemd[1]: Initializing machine ID from VM UUID. Jan 30 13:42:00.936539 zram_generator::config[1078]: No configuration found. Jan 30 13:42:00.936552 systemd[1]: Populated /etc with preset unit settings. Jan 30 13:42:00.936566 systemd[1]: Queued start job for default target multi-user.target. Jan 30 13:42:00.936783 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 30 13:42:00.936800 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 30 13:42:00.936812 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 30 13:42:00.936824 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 30 13:42:00.936841 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 30 13:42:00.936853 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 30 13:42:00.936866 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 30 13:42:00.936878 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 30 13:42:00.936894 systemd[1]: Created slice user.slice - User and Session Slice. Jan 30 13:42:00.936905 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:42:00.936917 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:42:00.936930 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 30 13:42:00.936942 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 30 13:42:00.936954 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 30 13:42:00.936966 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 13:42:00.936978 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 30 13:42:00.936990 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:42:00.937004 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 30 13:42:00.937016 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:42:00.937027 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 13:42:00.937039 systemd[1]: Reached target slices.target - Slice Units. Jan 30 13:42:00.937051 systemd[1]: Reached target swap.target - Swaps. Jan 30 13:42:00.937063 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 30 13:42:00.937075 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 30 13:42:00.937089 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 30 13:42:00.937101 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 30 13:42:00.937116 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:42:00.937132 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 13:42:00.937144 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:42:00.937157 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 30 13:42:00.937169 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 30 13:42:00.937182 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 30 13:42:00.937193 systemd[1]: Mounting media.mount - External Media Directory... Jan 30 13:42:00.937205 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:00.937220 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 30 13:42:00.937231 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 30 13:42:00.937243 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 30 13:42:00.937254 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 30 13:42:00.937267 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:42:00.937279 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 13:42:00.937291 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 30 13:42:00.937303 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:42:00.937317 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 13:42:00.937329 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:42:00.937341 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 30 13:42:00.937352 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:42:00.937365 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 30 13:42:00.937377 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Jan 30 13:42:00.937389 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Jan 30 13:42:00.937401 kernel: fuse: init (API version 7.39) Jan 30 13:42:00.937416 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 13:42:00.937427 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 13:42:00.937439 kernel: loop: module loaded Jan 30 13:42:00.937451 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 30 13:42:00.937463 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 30 13:42:00.937474 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 13:42:00.937505 systemd-journald[1159]: Collecting audit messages is disabled. Jan 30 13:42:00.937531 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:00.937545 systemd-journald[1159]: Journal started Jan 30 13:42:00.937567 systemd-journald[1159]: Runtime Journal (/run/log/journal/b292f1083fb9452b857790fff709c5b7) is 6.0M, max 48.4M, 42.3M free. Jan 30 13:42:00.940837 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 13:42:00.943017 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 30 13:42:00.944185 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 30 13:42:00.945394 systemd[1]: Mounted media.mount - External Media Directory. Jan 30 13:42:00.946552 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 30 13:42:00.947850 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 30 13:42:00.949256 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 30 13:42:00.950701 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 30 13:42:00.952313 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:42:00.954266 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 30 13:42:00.954490 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 30 13:42:00.956015 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:42:00.956221 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:42:00.957691 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:42:00.957895 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:42:00.959616 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 30 13:42:00.959886 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 30 13:42:00.961595 kernel: ACPI: bus type drm_connector registered Jan 30 13:42:00.961992 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:42:00.962279 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:42:00.964248 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 13:42:00.964464 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 13:42:00.966184 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 13:42:00.967746 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 30 13:42:00.969427 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 30 13:42:00.983569 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 30 13:42:00.993697 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 30 13:42:00.995919 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 30 13:42:00.997068 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 30 13:42:00.998994 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 30 13:42:01.015771 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 30 13:42:01.017257 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 13:42:01.020154 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 30 13:42:01.021365 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 13:42:01.025067 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:42:01.027781 systemd-journald[1159]: Time spent on flushing to /var/log/journal/b292f1083fb9452b857790fff709c5b7 is 13.761ms for 926 entries. Jan 30 13:42:01.027781 systemd-journald[1159]: System Journal (/var/log/journal/b292f1083fb9452b857790fff709c5b7) is 8.0M, max 195.6M, 187.6M free. Jan 30 13:42:01.055685 systemd-journald[1159]: Received client request to flush runtime journal. Jan 30 13:42:01.029293 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 13:42:01.037746 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:42:01.039317 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 30 13:42:01.040993 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 30 13:42:01.045320 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 30 13:42:01.047759 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 30 13:42:01.063780 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 30 13:42:01.065800 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 30 13:42:01.070368 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. Jan 30 13:42:01.070389 systemd-tmpfiles[1211]: ACLs are not supported, ignoring. Jan 30 13:42:01.071486 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:42:01.076058 udevadm[1220]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Jan 30 13:42:01.080032 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:42:01.087859 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 30 13:42:01.113670 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 30 13:42:01.120829 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 13:42:01.142939 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Jan 30 13:42:01.142966 systemd-tmpfiles[1233]: ACLs are not supported, ignoring. Jan 30 13:42:01.150326 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:42:01.593437 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 30 13:42:01.608861 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:42:01.633874 systemd-udevd[1239]: Using default interface naming scheme 'v255'. Jan 30 13:42:01.650209 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:42:01.660797 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 13:42:01.676744 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 30 13:42:01.700539 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1243) Jan 30 13:42:01.699230 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Jan 30 13:42:01.749749 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 30 13:42:01.769860 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 30 13:42:01.773608 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jan 30 13:42:01.778086 kernel: ACPI: button: Power Button [PWRF] Jan 30 13:42:01.801858 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 30 13:42:01.802542 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Jan 30 13:42:01.803045 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 30 13:42:01.807593 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Jan 30 13:42:01.835659 kernel: mousedev: PS/2 mouse device common for all mice Jan 30 13:42:01.845134 systemd-networkd[1244]: lo: Link UP Jan 30 13:42:01.845699 systemd-networkd[1244]: lo: Gained carrier Jan 30 13:42:01.847634 systemd-networkd[1244]: Enumeration completed Jan 30 13:42:01.848133 systemd-networkd[1244]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:42:01.848138 systemd-networkd[1244]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:42:01.849366 systemd-networkd[1244]: eth0: Link UP Jan 30 13:42:01.849428 systemd-networkd[1244]: eth0: Gained carrier Jan 30 13:42:01.849486 systemd-networkd[1244]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:42:01.851827 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:42:01.853305 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 13:42:01.860808 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 30 13:42:01.902881 systemd-networkd[1244]: eth0: DHCPv4 address 10.0.0.28/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 30 13:42:01.933630 kernel: kvm_amd: TSC scaling supported Jan 30 13:42:01.933835 kernel: kvm_amd: Nested Virtualization enabled Jan 30 13:42:01.933874 kernel: kvm_amd: Nested Paging enabled Jan 30 13:42:01.933910 kernel: kvm_amd: LBR virtualization supported Jan 30 13:42:01.933940 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 30 13:42:01.933976 kernel: kvm_amd: Virtual GIF supported Jan 30 13:42:01.951516 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:42:01.958670 kernel: EDAC MC: Ver: 3.0.0 Jan 30 13:42:01.992128 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 30 13:42:02.000753 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 30 13:42:02.010877 lvm[1286]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 13:42:02.043125 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 30 13:42:02.044794 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:42:02.056760 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 30 13:42:02.062014 lvm[1289]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 13:42:02.098428 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 30 13:42:02.100070 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 30 13:42:02.101362 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 30 13:42:02.101381 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 13:42:02.102438 systemd[1]: Reached target machines.target - Containers. Jan 30 13:42:02.104489 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 30 13:42:02.118792 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 30 13:42:02.121571 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 30 13:42:02.122794 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:42:02.123924 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 30 13:42:02.127514 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 30 13:42:02.130186 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 30 13:42:02.132583 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 30 13:42:02.146269 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 30 13:42:02.149712 kernel: loop0: detected capacity change from 0 to 140768 Jan 30 13:42:02.161052 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 30 13:42:02.161840 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 30 13:42:02.173608 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 30 13:42:02.200638 kernel: loop1: detected capacity change from 0 to 142488 Jan 30 13:42:02.238620 kernel: loop2: detected capacity change from 0 to 210664 Jan 30 13:42:02.269619 kernel: loop3: detected capacity change from 0 to 140768 Jan 30 13:42:02.281618 kernel: loop4: detected capacity change from 0 to 142488 Jan 30 13:42:02.292611 kernel: loop5: detected capacity change from 0 to 210664 Jan 30 13:42:02.298713 (sd-merge)[1310]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Jan 30 13:42:02.299296 (sd-merge)[1310]: Merged extensions into '/usr'. Jan 30 13:42:02.303904 systemd[1]: Reloading requested from client PID 1297 ('systemd-sysext') (unit systemd-sysext.service)... Jan 30 13:42:02.303921 systemd[1]: Reloading... Jan 30 13:42:02.366651 zram_generator::config[1341]: No configuration found. Jan 30 13:42:02.398424 ldconfig[1293]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 30 13:42:02.503136 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:42:02.566711 systemd[1]: Reloading finished in 262 ms. Jan 30 13:42:02.586662 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 30 13:42:02.588194 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 30 13:42:02.598728 systemd[1]: Starting ensure-sysext.service... Jan 30 13:42:02.600649 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 13:42:02.605707 systemd[1]: Reloading requested from client PID 1382 ('systemctl') (unit ensure-sysext.service)... Jan 30 13:42:02.605720 systemd[1]: Reloading... Jan 30 13:42:02.624375 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 30 13:42:02.624872 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 30 13:42:02.626120 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 30 13:42:02.626519 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Jan 30 13:42:02.626774 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Jan 30 13:42:02.635679 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 13:42:02.635833 systemd-tmpfiles[1383]: Skipping /boot Jan 30 13:42:02.649608 zram_generator::config[1412]: No configuration found. Jan 30 13:42:02.651973 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 13:42:02.651989 systemd-tmpfiles[1383]: Skipping /boot Jan 30 13:42:02.759769 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:42:02.824470 systemd[1]: Reloading finished in 218 ms. Jan 30 13:42:02.842103 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:42:02.859219 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 30 13:42:02.862132 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 30 13:42:02.864917 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 30 13:42:02.869123 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 13:42:02.872878 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 30 13:42:02.879269 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:02.879493 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:42:02.882881 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:42:02.887547 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:42:02.891635 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:42:02.895725 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:42:02.895878 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:02.897065 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:42:02.897330 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:42:02.900555 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:42:02.901736 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:42:02.907477 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:42:02.907807 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:42:02.912624 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 30 13:42:02.918396 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 13:42:02.919472 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 13:42:02.928889 augenrules[1492]: No rules Jan 30 13:42:02.929877 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 30 13:42:02.932378 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 30 13:42:02.934718 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 30 13:42:02.944436 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 30 13:42:02.949047 systemd[1]: Finished ensure-sysext.service. Jan 30 13:42:02.952550 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:02.952803 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:42:02.958722 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:42:02.963553 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 13:42:02.966729 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:42:02.971648 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:42:02.973070 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:42:02.974965 systemd-networkd[1244]: eth0: Gained IPv6LL Jan 30 13:42:02.976035 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 30 13:42:02.977635 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 30 13:42:02.978751 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 30 13:42:02.980905 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:42:02.981256 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:42:02.983002 systemd-resolved[1461]: Positive Trust Anchors: Jan 30 13:42:02.983018 systemd-resolved[1461]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 13:42:02.983060 systemd-resolved[1461]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 13:42:02.983378 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 30 13:42:02.985404 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 13:42:02.985760 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 13:42:02.987788 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:42:02.987992 systemd-resolved[1461]: Defaulting to hostname 'linux'. Jan 30 13:42:02.988128 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:42:02.990359 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:42:02.990627 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:42:02.992007 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 13:42:02.997450 systemd[1]: Reached target network.target - Network. Jan 30 13:42:02.998597 systemd[1]: Reached target network-online.target - Network is Online. Jan 30 13:42:02.999707 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:42:03.000932 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 13:42:03.000979 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 13:42:03.001003 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 30 13:42:03.061806 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 30 13:42:03.759289 systemd-resolved[1461]: Clock change detected. Flushing caches. Jan 30 13:42:03.759339 systemd-timesyncd[1511]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 30 13:42:03.759380 systemd-timesyncd[1511]: Initial clock synchronization to Thu 2025-01-30 13:42:03.759231 UTC. Jan 30 13:42:03.760216 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 13:42:03.761405 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 30 13:42:03.762674 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 30 13:42:03.763946 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 30 13:42:03.765247 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 30 13:42:03.765275 systemd[1]: Reached target paths.target - Path Units. Jan 30 13:42:03.766219 systemd[1]: Reached target time-set.target - System Time Set. Jan 30 13:42:03.767483 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 30 13:42:03.768666 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 30 13:42:03.769908 systemd[1]: Reached target timers.target - Timer Units. Jan 30 13:42:03.771422 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 30 13:42:03.774337 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 30 13:42:03.776624 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 30 13:42:03.785718 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 30 13:42:03.786960 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 13:42:03.788049 systemd[1]: Reached target basic.target - Basic System. Jan 30 13:42:03.789298 systemd[1]: System is tainted: cgroupsv1 Jan 30 13:42:03.789340 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 30 13:42:03.789369 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 30 13:42:03.790572 systemd[1]: Starting containerd.service - containerd container runtime... Jan 30 13:42:03.792827 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 30 13:42:03.796895 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 30 13:42:03.800345 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 30 13:42:03.803201 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 30 13:42:03.804516 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 30 13:42:03.807285 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:42:03.811860 jq[1530]: false Jan 30 13:42:03.812756 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 30 13:42:03.821020 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 30 13:42:03.827067 extend-filesystems[1531]: Found loop3 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found loop4 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found loop5 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found sr0 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda1 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda2 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda3 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found usr Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda4 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda6 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda7 Jan 30 13:42:03.829441 extend-filesystems[1531]: Found vda9 Jan 30 13:42:03.829441 extend-filesystems[1531]: Checking size of /dev/vda9 Jan 30 13:42:03.872271 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1256) Jan 30 13:42:03.872306 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Jan 30 13:42:03.827269 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 30 13:42:03.828909 dbus-daemon[1527]: [system] SELinux support is enabled Jan 30 13:42:03.872762 extend-filesystems[1531]: Resized partition /dev/vda9 Jan 30 13:42:03.829041 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 30 13:42:03.880136 extend-filesystems[1557]: resize2fs 1.47.1 (20-May-2024) Jan 30 13:42:03.837735 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 30 13:42:03.841027 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 30 13:42:03.843547 systemd[1]: Starting update-engine.service - Update Engine... Jan 30 13:42:03.854187 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 30 13:42:03.884696 jq[1554]: true Jan 30 13:42:03.858063 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 30 13:42:03.873552 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 30 13:42:03.873872 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 30 13:42:03.875710 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 30 13:42:03.875995 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 30 13:42:03.889770 systemd[1]: motdgen.service: Deactivated successfully. Jan 30 13:42:03.890064 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 30 13:42:03.896319 update_engine[1550]: I20250130 13:42:03.894255 1550 main.cc:92] Flatcar Update Engine starting Jan 30 13:42:03.896319 update_engine[1550]: I20250130 13:42:03.895723 1550 update_check_scheduler.cc:74] Next update check in 7m48s Jan 30 13:42:03.896445 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 30 13:42:03.900114 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Jan 30 13:42:03.916339 (ntainerd)[1573]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 30 13:42:03.929653 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 30 13:42:03.936511 jq[1569]: true Jan 30 13:42:03.930033 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 30 13:42:03.937893 extend-filesystems[1557]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 30 13:42:03.937893 extend-filesystems[1557]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 30 13:42:03.937893 extend-filesystems[1557]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Jan 30 13:42:03.941584 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 30 13:42:03.948529 extend-filesystems[1531]: Resized filesystem in /dev/vda9 Jan 30 13:42:03.942000 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 30 13:42:03.962753 systemd[1]: Started update-engine.service - Update Engine. Jan 30 13:42:03.967669 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 30 13:42:03.967812 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 30 13:42:03.967838 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 30 13:42:03.969422 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 30 13:42:03.969444 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 30 13:42:03.971930 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 30 13:42:03.979715 systemd-logind[1545]: Watching system buttons on /dev/input/event1 (Power Button) Jan 30 13:42:03.979749 systemd-logind[1545]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 30 13:42:03.982340 systemd-logind[1545]: New seat seat0. Jan 30 13:42:03.982650 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 30 13:42:03.988696 systemd[1]: Started systemd-logind.service - User Login Management. Jan 30 13:42:04.015251 bash[1606]: Updated "/home/core/.ssh/authorized_keys" Jan 30 13:42:04.017684 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 30 13:42:04.020876 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 30 13:42:04.024303 locksmithd[1600]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 30 13:42:04.137151 containerd[1573]: time="2025-01-30T13:42:04.136991448Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jan 30 13:42:04.155747 sshd_keygen[1559]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 30 13:42:04.161549 containerd[1573]: time="2025-01-30T13:42:04.160441996Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.162662 containerd[1573]: time="2025-01-30T13:42:04.162601576Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:42:04.162662 containerd[1573]: time="2025-01-30T13:42:04.162656118Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 30 13:42:04.162718 containerd[1573]: time="2025-01-30T13:42:04.162681005Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 30 13:42:04.162868 containerd[1573]: time="2025-01-30T13:42:04.162849581Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 30 13:42:04.162889 containerd[1573]: time="2025-01-30T13:42:04.162873546Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.162960 containerd[1573]: time="2025-01-30T13:42:04.162942335Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:42:04.162981 containerd[1573]: time="2025-01-30T13:42:04.162959688Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163292 containerd[1573]: time="2025-01-30T13:42:04.163266503Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163758 containerd[1573]: time="2025-01-30T13:42:04.163376369Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163758 containerd[1573]: time="2025-01-30T13:42:04.163397639Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163758 containerd[1573]: time="2025-01-30T13:42:04.163407928Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163758 containerd[1573]: time="2025-01-30T13:42:04.163500111Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.163758 containerd[1573]: time="2025-01-30T13:42:04.163725634Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:42:04.164050 containerd[1573]: time="2025-01-30T13:42:04.164030997Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:42:04.164131 containerd[1573]: time="2025-01-30T13:42:04.164117369Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 30 13:42:04.164272 containerd[1573]: time="2025-01-30T13:42:04.164258153Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 30 13:42:04.164376 containerd[1573]: time="2025-01-30T13:42:04.164362639Z" level=info msg="metadata content store policy set" policy=shared Jan 30 13:42:04.171040 containerd[1573]: time="2025-01-30T13:42:04.171006538Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 30 13:42:04.171167 containerd[1573]: time="2025-01-30T13:42:04.171153985Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 30 13:42:04.171271 containerd[1573]: time="2025-01-30T13:42:04.171257539Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 30 13:42:04.171329 containerd[1573]: time="2025-01-30T13:42:04.171317832Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 30 13:42:04.171377 containerd[1573]: time="2025-01-30T13:42:04.171366363Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 30 13:42:04.171578 containerd[1573]: time="2025-01-30T13:42:04.171562270Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 30 13:42:04.171974 containerd[1573]: time="2025-01-30T13:42:04.171959225Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 30 13:42:04.172166 containerd[1573]: time="2025-01-30T13:42:04.172150664Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 30 13:42:04.172226 containerd[1573]: time="2025-01-30T13:42:04.172213652Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 30 13:42:04.172272 containerd[1573]: time="2025-01-30T13:42:04.172261832Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172307278Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172322937Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172334499Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172347493Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172361990Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172374363Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172385574Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172397837Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172417154Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172429868Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172443012Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172454674Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172467688Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172694 containerd[1573]: time="2025-01-30T13:42:04.172481304Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172498336Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172511240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172525196Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172543400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172565692Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172578266Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172590148Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172608933Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172628029Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172639421Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.172955 containerd[1573]: time="2025-01-30T13:42:04.172654599Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 30 13:42:04.173196 containerd[1573]: time="2025-01-30T13:42:04.173180265Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 30 13:42:04.173259 containerd[1573]: time="2025-01-30T13:42:04.173244716Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 30 13:42:04.173301 containerd[1573]: time="2025-01-30T13:42:04.173290952Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 30 13:42:04.173345 containerd[1573]: time="2025-01-30T13:42:04.173334234Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 30 13:42:04.173385 containerd[1573]: time="2025-01-30T13:42:04.173375100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.173427 containerd[1573]: time="2025-01-30T13:42:04.173417931Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 30 13:42:04.174597 containerd[1573]: time="2025-01-30T13:42:04.173459058Z" level=info msg="NRI interface is disabled by configuration." Jan 30 13:42:04.174597 containerd[1573]: time="2025-01-30T13:42:04.173471611Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.173695661Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.173745184Z" level=info msg="Connect containerd service" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.173782204Z" level=info msg="using legacy CRI server" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.173789046Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.173884976Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.174390675Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.174533623Z" level=info msg="Start subscribing containerd event" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.174591331Z" level=info msg="Start recovering state" Jan 30 13:42:04.174658 containerd[1573]: time="2025-01-30T13:42:04.174648789Z" level=info msg="Start event monitor" Jan 30 13:42:04.174944 containerd[1573]: time="2025-01-30T13:42:04.174672924Z" level=info msg="Start snapshots syncer" Jan 30 13:42:04.174944 containerd[1573]: time="2025-01-30T13:42:04.174683764Z" level=info msg="Start cni network conf syncer for default" Jan 30 13:42:04.174944 containerd[1573]: time="2025-01-30T13:42:04.174690717Z" level=info msg="Start streaming server" Jan 30 13:42:04.175189 containerd[1573]: time="2025-01-30T13:42:04.175172401Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 30 13:42:04.175281 containerd[1573]: time="2025-01-30T13:42:04.175268812Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 30 13:42:04.176326 containerd[1573]: time="2025-01-30T13:42:04.175364792Z" level=info msg="containerd successfully booted in 0.039824s" Jan 30 13:42:04.175484 systemd[1]: Started containerd.service - containerd container runtime. Jan 30 13:42:04.183968 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 30 13:42:04.200958 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 30 13:42:04.208646 systemd[1]: issuegen.service: Deactivated successfully. Jan 30 13:42:04.209038 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 30 13:42:04.216360 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 30 13:42:04.229078 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 30 13:42:04.237484 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 30 13:42:04.240034 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 30 13:42:04.241525 systemd[1]: Reached target getty.target - Login Prompts. Jan 30 13:42:04.613616 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:42:04.615411 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 30 13:42:04.617347 systemd[1]: Startup finished in 6.233s (kernel) + 3.729s (userspace) = 9.963s. Jan 30 13:42:04.643673 (kubelet)[1650]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:42:05.075550 kubelet[1650]: E0130 13:42:05.075425 1650 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:42:05.079374 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:42:05.079632 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:42:13.268972 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 30 13:42:13.279312 systemd[1]: Started sshd@0-10.0.0.28:22-10.0.0.1:41002.service - OpenSSH per-connection server daemon (10.0.0.1:41002). Jan 30 13:42:13.328151 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 41002 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:13.330347 sshd[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:13.340595 systemd-logind[1545]: New session 1 of user core. Jan 30 13:42:13.341775 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 30 13:42:13.359410 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 30 13:42:13.372656 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 30 13:42:13.375314 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 30 13:42:13.383939 (systemd)[1670]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 30 13:42:13.497635 systemd[1670]: Queued start job for default target default.target. Jan 30 13:42:13.498059 systemd[1670]: Created slice app.slice - User Application Slice. Jan 30 13:42:13.498081 systemd[1670]: Reached target paths.target - Paths. Jan 30 13:42:13.498108 systemd[1670]: Reached target timers.target - Timers. Jan 30 13:42:13.513240 systemd[1670]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 30 13:42:13.520148 systemd[1670]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 30 13:42:13.520239 systemd[1670]: Reached target sockets.target - Sockets. Jan 30 13:42:13.520263 systemd[1670]: Reached target basic.target - Basic System. Jan 30 13:42:13.520315 systemd[1670]: Reached target default.target - Main User Target. Jan 30 13:42:13.520360 systemd[1670]: Startup finished in 128ms. Jan 30 13:42:13.520904 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 30 13:42:13.522440 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 30 13:42:13.586420 systemd[1]: Started sshd@1-10.0.0.28:22-10.0.0.1:41010.service - OpenSSH per-connection server daemon (10.0.0.1:41010). Jan 30 13:42:13.619518 sshd[1682]: Accepted publickey for core from 10.0.0.1 port 41010 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:13.621007 sshd[1682]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:13.625146 systemd-logind[1545]: New session 2 of user core. Jan 30 13:42:13.645632 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 30 13:42:13.698570 sshd[1682]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:13.712337 systemd[1]: Started sshd@2-10.0.0.28:22-10.0.0.1:41024.service - OpenSSH per-connection server daemon (10.0.0.1:41024). Jan 30 13:42:13.712947 systemd[1]: sshd@1-10.0.0.28:22-10.0.0.1:41010.service: Deactivated successfully. Jan 30 13:42:13.715656 systemd[1]: session-2.scope: Deactivated successfully. Jan 30 13:42:13.716344 systemd-logind[1545]: Session 2 logged out. Waiting for processes to exit. Jan 30 13:42:13.717843 systemd-logind[1545]: Removed session 2. Jan 30 13:42:13.746043 sshd[1687]: Accepted publickey for core from 10.0.0.1 port 41024 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:13.747396 sshd[1687]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:13.751407 systemd-logind[1545]: New session 3 of user core. Jan 30 13:42:13.764345 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 30 13:42:13.814717 sshd[1687]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:13.823344 systemd[1]: Started sshd@3-10.0.0.28:22-10.0.0.1:41036.service - OpenSSH per-connection server daemon (10.0.0.1:41036). Jan 30 13:42:13.823945 systemd[1]: sshd@2-10.0.0.28:22-10.0.0.1:41024.service: Deactivated successfully. Jan 30 13:42:13.826432 systemd-logind[1545]: Session 3 logged out. Waiting for processes to exit. Jan 30 13:42:13.827476 systemd[1]: session-3.scope: Deactivated successfully. Jan 30 13:42:13.828548 systemd-logind[1545]: Removed session 3. Jan 30 13:42:13.857659 sshd[1695]: Accepted publickey for core from 10.0.0.1 port 41036 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:13.859386 sshd[1695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:13.863161 systemd-logind[1545]: New session 4 of user core. Jan 30 13:42:13.874339 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 30 13:42:13.927637 sshd[1695]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:13.937314 systemd[1]: Started sshd@4-10.0.0.28:22-10.0.0.1:41052.service - OpenSSH per-connection server daemon (10.0.0.1:41052). Jan 30 13:42:13.937905 systemd[1]: sshd@3-10.0.0.28:22-10.0.0.1:41036.service: Deactivated successfully. Jan 30 13:42:13.940124 systemd-logind[1545]: Session 4 logged out. Waiting for processes to exit. Jan 30 13:42:13.941779 systemd[1]: session-4.scope: Deactivated successfully. Jan 30 13:42:13.942869 systemd-logind[1545]: Removed session 4. Jan 30 13:42:13.970965 sshd[1703]: Accepted publickey for core from 10.0.0.1 port 41052 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:13.972587 sshd[1703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:13.976579 systemd-logind[1545]: New session 5 of user core. Jan 30 13:42:13.987328 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 30 13:42:14.046019 sudo[1710]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 30 13:42:14.046371 sudo[1710]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:42:14.062267 sudo[1710]: pam_unix(sudo:session): session closed for user root Jan 30 13:42:14.064312 sshd[1703]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:14.074335 systemd[1]: Started sshd@5-10.0.0.28:22-10.0.0.1:41068.service - OpenSSH per-connection server daemon (10.0.0.1:41068). Jan 30 13:42:14.074926 systemd[1]: sshd@4-10.0.0.28:22-10.0.0.1:41052.service: Deactivated successfully. Jan 30 13:42:14.077314 systemd-logind[1545]: Session 5 logged out. Waiting for processes to exit. Jan 30 13:42:14.078507 systemd[1]: session-5.scope: Deactivated successfully. Jan 30 13:42:14.079640 systemd-logind[1545]: Removed session 5. Jan 30 13:42:14.110386 sshd[1712]: Accepted publickey for core from 10.0.0.1 port 41068 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:14.112077 sshd[1712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:14.115920 systemd-logind[1545]: New session 6 of user core. Jan 30 13:42:14.125329 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 30 13:42:14.178720 sudo[1720]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 30 13:42:14.179077 sudo[1720]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:42:14.182875 sudo[1720]: pam_unix(sudo:session): session closed for user root Jan 30 13:42:14.189211 sudo[1719]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jan 30 13:42:14.189553 sudo[1719]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:42:14.213376 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jan 30 13:42:14.215206 auditctl[1723]: No rules Jan 30 13:42:14.216586 systemd[1]: audit-rules.service: Deactivated successfully. Jan 30 13:42:14.216952 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jan 30 13:42:14.218932 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 30 13:42:14.250117 augenrules[1742]: No rules Jan 30 13:42:14.252044 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 30 13:42:14.253549 sudo[1719]: pam_unix(sudo:session): session closed for user root Jan 30 13:42:14.255407 sshd[1712]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:14.262393 systemd[1]: Started sshd@6-10.0.0.28:22-10.0.0.1:41076.service - OpenSSH per-connection server daemon (10.0.0.1:41076). Jan 30 13:42:14.262952 systemd[1]: sshd@5-10.0.0.28:22-10.0.0.1:41068.service: Deactivated successfully. Jan 30 13:42:14.265706 systemd-logind[1545]: Session 6 logged out. Waiting for processes to exit. Jan 30 13:42:14.266719 systemd[1]: session-6.scope: Deactivated successfully. Jan 30 13:42:14.267684 systemd-logind[1545]: Removed session 6. Jan 30 13:42:14.297757 sshd[1748]: Accepted publickey for core from 10.0.0.1 port 41076 ssh2: RSA SHA256:5CVmNz7KcUi5XiFI6hIHcAt9PUhPYHR+qHQIWL4Xluc Jan 30 13:42:14.299584 sshd[1748]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:42:14.304172 systemd-logind[1545]: New session 7 of user core. Jan 30 13:42:14.313508 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 30 13:42:14.369817 sudo[1755]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 30 13:42:14.370246 sudo[1755]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:42:14.392521 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 30 13:42:14.415326 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 30 13:42:14.415675 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 30 13:42:14.885821 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:42:14.899363 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:42:14.917545 systemd[1]: Reloading requested from client PID 1807 ('systemctl') (unit session-7.scope)... Jan 30 13:42:14.917564 systemd[1]: Reloading... Jan 30 13:42:14.990283 zram_generator::config[1845]: No configuration found. Jan 30 13:42:15.372700 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:42:15.444299 systemd[1]: Reloading finished in 526 ms. Jan 30 13:42:15.497087 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 30 13:42:15.497234 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 30 13:42:15.497633 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:42:15.514073 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:42:15.651058 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:42:15.655292 (kubelet)[1905]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 30 13:42:15.700193 kubelet[1905]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:42:15.700193 kubelet[1905]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 30 13:42:15.700193 kubelet[1905]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:42:15.701348 kubelet[1905]: I0130 13:42:15.701291 1905 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 30 13:42:15.889711 kubelet[1905]: I0130 13:42:15.889654 1905 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Jan 30 13:42:15.889711 kubelet[1905]: I0130 13:42:15.889685 1905 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 30 13:42:15.889937 kubelet[1905]: I0130 13:42:15.889912 1905 server.go:927] "Client rotation is on, will bootstrap in background" Jan 30 13:42:15.904732 kubelet[1905]: I0130 13:42:15.904603 1905 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 13:42:15.915379 kubelet[1905]: I0130 13:42:15.915342 1905 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 30 13:42:15.917159 kubelet[1905]: I0130 13:42:15.917116 1905 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 30 13:42:15.917324 kubelet[1905]: I0130 13:42:15.917150 1905 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"10.0.0.28","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 30 13:42:15.917690 kubelet[1905]: I0130 13:42:15.917665 1905 topology_manager.go:138] "Creating topology manager with none policy" Jan 30 13:42:15.917690 kubelet[1905]: I0130 13:42:15.917681 1905 container_manager_linux.go:301] "Creating device plugin manager" Jan 30 13:42:15.917844 kubelet[1905]: I0130 13:42:15.917820 1905 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:42:15.918413 kubelet[1905]: I0130 13:42:15.918382 1905 kubelet.go:400] "Attempting to sync node with API server" Jan 30 13:42:15.918413 kubelet[1905]: I0130 13:42:15.918398 1905 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 30 13:42:15.918413 kubelet[1905]: I0130 13:42:15.918416 1905 kubelet.go:312] "Adding apiserver pod source" Jan 30 13:42:15.918487 kubelet[1905]: I0130 13:42:15.918432 1905 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 30 13:42:15.918703 kubelet[1905]: E0130 13:42:15.918510 1905 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:15.918703 kubelet[1905]: E0130 13:42:15.918564 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:15.922046 kubelet[1905]: I0130 13:42:15.921804 1905 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 30 13:42:15.922046 kubelet[1905]: W0130 13:42:15.922025 1905 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 30 13:42:15.922122 kubelet[1905]: E0130 13:42:15.922049 1905 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 30 13:42:15.922122 kubelet[1905]: W0130 13:42:15.922024 1905 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "10.0.0.28" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 30 13:42:15.922122 kubelet[1905]: E0130 13:42:15.922066 1905 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes "10.0.0.28" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 30 13:42:15.922908 kubelet[1905]: I0130 13:42:15.922879 1905 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 30 13:42:15.922942 kubelet[1905]: W0130 13:42:15.922932 1905 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 30 13:42:15.923576 kubelet[1905]: I0130 13:42:15.923560 1905 server.go:1264] "Started kubelet" Jan 30 13:42:15.923638 kubelet[1905]: I0130 13:42:15.923615 1905 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 30 13:42:15.924257 kubelet[1905]: I0130 13:42:15.923720 1905 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 30 13:42:15.924257 kubelet[1905]: I0130 13:42:15.924145 1905 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 30 13:42:15.924545 kubelet[1905]: I0130 13:42:15.924513 1905 server.go:455] "Adding debug handlers to kubelet server" Jan 30 13:42:15.924857 kubelet[1905]: I0130 13:42:15.924836 1905 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 30 13:42:15.926518 kubelet[1905]: E0130 13:42:15.926500 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:15.926554 kubelet[1905]: I0130 13:42:15.926537 1905 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 30 13:42:15.927331 kubelet[1905]: I0130 13:42:15.926611 1905 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 30 13:42:15.927331 kubelet[1905]: I0130 13:42:15.926676 1905 reconciler.go:26] "Reconciler: start to sync state" Jan 30 13:42:15.930365 kubelet[1905]: I0130 13:42:15.928419 1905 factory.go:221] Registration of the systemd container factory successfully Jan 30 13:42:15.930365 kubelet[1905]: I0130 13:42:15.928499 1905 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 30 13:42:15.932221 kubelet[1905]: E0130 13:42:15.931968 1905 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 30 13:42:15.932770 kubelet[1905]: I0130 13:42:15.932740 1905 factory.go:221] Registration of the containerd container factory successfully Jan 30 13:42:15.932961 kubelet[1905]: E0130 13:42:15.932850 1905 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32207c7737 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.923537719 +0000 UTC m=+0.264568501,LastTimestamp:2025-01-30 13:42:15.923537719 +0000 UTC m=+0.264568501,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:15.937919 kubelet[1905]: E0130 13:42:15.937853 1905 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c3220fcedda default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.931956698 +0000 UTC m=+0.272987480,LastTimestamp:2025-01-30 13:42:15.931956698 +0000 UTC m=+0.272987480,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:15.940934 kubelet[1905]: E0130 13:42:15.940916 1905 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"10.0.0.28\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Jan 30 13:42:15.941077 kubelet[1905]: W0130 13:42:15.941064 1905 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 30 13:42:15.941150 kubelet[1905]: E0130 13:42:15.941139 1905 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 30 13:42:15.951429 kubelet[1905]: I0130 13:42:15.951408 1905 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 30 13:42:15.951429 kubelet[1905]: I0130 13:42:15.951425 1905 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 30 13:42:15.951513 kubelet[1905]: I0130 13:42:15.951448 1905 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:42:15.951637 kubelet[1905]: E0130 13:42:15.951538 1905 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cd5db default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.28 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950824923 +0000 UTC m=+0.291855695,LastTimestamp:2025-01-30 13:42:15.950824923 +0000 UTC m=+0.291855695,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:15.954674 kubelet[1905]: E0130 13:42:15.954619 1905 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cff6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.28 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950835563 +0000 UTC m=+0.291866345,LastTimestamp:2025-01-30 13:42:15.950835563 +0000 UTC m=+0.291866345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:15.957547 kubelet[1905]: E0130 13:42:15.957458 1905 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221d0b5a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.28 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950838618 +0000 UTC m=+0.291869400,LastTimestamp:2025-01-30 13:42:15.950838618 +0000 UTC m=+0.291869400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.027408 kubelet[1905]: I0130 13:42:16.027385 1905 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.28" Jan 30 13:42:16.030771 kubelet[1905]: E0130 13:42:16.030732 1905 kubelet_node_status.go:96] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="10.0.0.28" Jan 30 13:42:16.030838 kubelet[1905]: E0130 13:42:16.030778 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221cd5db\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cd5db default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.28 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950824923 +0000 UTC m=+0.291855695,LastTimestamp:2025-01-30 13:42:16.02735288 +0000 UTC m=+0.368383662,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.033529 kubelet[1905]: E0130 13:42:16.033450 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221cff6b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cff6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.28 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950835563 +0000 UTC m=+0.291866345,LastTimestamp:2025-01-30 13:42:16.027360083 +0000 UTC m=+0.368390865,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.036469 kubelet[1905]: E0130 13:42:16.036374 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221d0b5a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221d0b5a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.28 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950838618 +0000 UTC m=+0.291869400,LastTimestamp:2025-01-30 13:42:16.027362788 +0000 UTC m=+0.368393570,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.144829 kubelet[1905]: E0130 13:42:16.144800 1905 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"10.0.0.28\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="400ms" Jan 30 13:42:16.232491 kubelet[1905]: I0130 13:42:16.232387 1905 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.28" Jan 30 13:42:16.235875 kubelet[1905]: E0130 13:42:16.235839 1905 kubelet_node_status.go:96] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="10.0.0.28" Jan 30 13:42:16.236132 kubelet[1905]: E0130 13:42:16.235992 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221cd5db\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cd5db default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 10.0.0.28 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950824923 +0000 UTC m=+0.291855695,LastTimestamp:2025-01-30 13:42:16.232319586 +0000 UTC m=+0.573350368,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.239348 kubelet[1905]: E0130 13:42:16.239265 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221cff6b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221cff6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.28 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950835563 +0000 UTC m=+0.291866345,LastTimestamp:2025-01-30 13:42:16.232333482 +0000 UTC m=+0.573364264,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.242337 kubelet[1905]: E0130 13:42:16.242250 1905 event.go:359] "Server rejected event (will not retry!)" err="events \"10.0.0.28.181f7c32221d0b5a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.28.181f7c32221d0b5a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.28,UID:10.0.0.28,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node 10.0.0.28 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:10.0.0.28,},FirstTimestamp:2025-01-30 13:42:15.950838618 +0000 UTC m=+0.291869400,LastTimestamp:2025-01-30 13:42:16.232337309 +0000 UTC m=+0.573368091,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.28,}" Jan 30 13:42:16.385379 kubelet[1905]: I0130 13:42:16.385312 1905 policy_none.go:49] "None policy: Start" Jan 30 13:42:16.386244 kubelet[1905]: I0130 13:42:16.386202 1905 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 30 13:42:16.386244 kubelet[1905]: I0130 13:42:16.386241 1905 state_mem.go:35] "Initializing new in-memory state store" Jan 30 13:42:16.393627 kubelet[1905]: I0130 13:42:16.393595 1905 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 30 13:42:16.393867 kubelet[1905]: I0130 13:42:16.393824 1905 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 30 13:42:16.393960 kubelet[1905]: I0130 13:42:16.393946 1905 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 30 13:42:16.395638 kubelet[1905]: E0130 13:42:16.395621 1905 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"10.0.0.28\" not found" Jan 30 13:42:16.414694 kubelet[1905]: I0130 13:42:16.414657 1905 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 30 13:42:16.415945 kubelet[1905]: I0130 13:42:16.415915 1905 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 30 13:42:16.415945 kubelet[1905]: I0130 13:42:16.415939 1905 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 30 13:42:16.415999 kubelet[1905]: I0130 13:42:16.415959 1905 kubelet.go:2337] "Starting kubelet main sync loop" Jan 30 13:42:16.416051 kubelet[1905]: E0130 13:42:16.416019 1905 kubelet.go:2361] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 30 13:42:16.549227 kubelet[1905]: E0130 13:42:16.549079 1905 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"10.0.0.28\" not found" node="10.0.0.28" Jan 30 13:42:16.637327 kubelet[1905]: I0130 13:42:16.637288 1905 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.28" Jan 30 13:42:16.643327 kubelet[1905]: I0130 13:42:16.643298 1905 kubelet_node_status.go:76] "Successfully registered node" node="10.0.0.28" Jan 30 13:42:16.651235 kubelet[1905]: E0130 13:42:16.651208 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:16.751756 kubelet[1905]: E0130 13:42:16.751686 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:16.852514 kubelet[1905]: E0130 13:42:16.852369 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:16.892703 kubelet[1905]: I0130 13:42:16.892672 1905 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 30 13:42:16.892942 kubelet[1905]: W0130 13:42:16.892895 1905 reflector.go:470] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Jan 30 13:42:16.919256 kubelet[1905]: E0130 13:42:16.919199 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:16.953408 kubelet[1905]: E0130 13:42:16.953341 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.054203 kubelet[1905]: E0130 13:42:17.054144 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.155000 kubelet[1905]: E0130 13:42:17.154810 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.170807 sudo[1755]: pam_unix(sudo:session): session closed for user root Jan 30 13:42:17.172590 sshd[1748]: pam_unix(sshd:session): session closed for user core Jan 30 13:42:17.176582 systemd[1]: sshd@6-10.0.0.28:22-10.0.0.1:41076.service: Deactivated successfully. Jan 30 13:42:17.178828 systemd-logind[1545]: Session 7 logged out. Waiting for processes to exit. Jan 30 13:42:17.178958 systemd[1]: session-7.scope: Deactivated successfully. Jan 30 13:42:17.179962 systemd-logind[1545]: Removed session 7. Jan 30 13:42:17.255417 kubelet[1905]: E0130 13:42:17.255367 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.356511 kubelet[1905]: E0130 13:42:17.356462 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.456931 kubelet[1905]: E0130 13:42:17.456832 1905 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.28\" not found" Jan 30 13:42:17.558341 kubelet[1905]: I0130 13:42:17.558304 1905 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Jan 30 13:42:17.558642 containerd[1573]: time="2025-01-30T13:42:17.558591389Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 30 13:42:17.559144 kubelet[1905]: I0130 13:42:17.558769 1905 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Jan 30 13:42:17.919995 kubelet[1905]: I0130 13:42:17.919851 1905 apiserver.go:52] "Watching apiserver" Jan 30 13:42:17.919995 kubelet[1905]: E0130 13:42:17.919909 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:17.923676 kubelet[1905]: I0130 13:42:17.923628 1905 topology_manager.go:215] "Topology Admit Handler" podUID="c8de6409-82b8-494d-920d-41141f7a284a" podNamespace="kube-system" podName="cilium-b622d" Jan 30 13:42:17.923817 kubelet[1905]: I0130 13:42:17.923794 1905 topology_manager.go:215] "Topology Admit Handler" podUID="cc839e17-3de5-4d0b-8fdb-264e461ece18" podNamespace="kube-system" podName="kube-proxy-ts5qv" Jan 30 13:42:17.928790 kubelet[1905]: I0130 13:42:17.928753 1905 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 30 13:42:17.939159 kubelet[1905]: I0130 13:42:17.939071 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-run\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939159 kubelet[1905]: I0130 13:42:17.939158 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-etc-cni-netd\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939309 kubelet[1905]: I0130 13:42:17.939186 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-lib-modules\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939309 kubelet[1905]: I0130 13:42:17.939206 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-net\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939309 kubelet[1905]: I0130 13:42:17.939225 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-hubble-tls\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939309 kubelet[1905]: I0130 13:42:17.939245 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/cc839e17-3de5-4d0b-8fdb-264e461ece18-kube-proxy\") pod \"kube-proxy-ts5qv\" (UID: \"cc839e17-3de5-4d0b-8fdb-264e461ece18\") " pod="kube-system/kube-proxy-ts5qv" Jan 30 13:42:17.939439 kubelet[1905]: I0130 13:42:17.939265 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cstrk\" (UniqueName: \"kubernetes.io/projected/cc839e17-3de5-4d0b-8fdb-264e461ece18-kube-api-access-cstrk\") pod \"kube-proxy-ts5qv\" (UID: \"cc839e17-3de5-4d0b-8fdb-264e461ece18\") " pod="kube-system/kube-proxy-ts5qv" Jan 30 13:42:17.939439 kubelet[1905]: I0130 13:42:17.939369 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-cgroup\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939439 kubelet[1905]: I0130 13:42:17.939419 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c8de6409-82b8-494d-920d-41141f7a284a-clustermesh-secrets\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939523 kubelet[1905]: I0130 13:42:17.939449 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-kernel\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939523 kubelet[1905]: I0130 13:42:17.939478 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc839e17-3de5-4d0b-8fdb-264e461ece18-lib-modules\") pod \"kube-proxy-ts5qv\" (UID: \"cc839e17-3de5-4d0b-8fdb-264e461ece18\") " pod="kube-system/kube-proxy-ts5qv" Jan 30 13:42:17.939523 kubelet[1905]: I0130 13:42:17.939494 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cni-path\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939523 kubelet[1905]: I0130 13:42:17.939511 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c8de6409-82b8-494d-920d-41141f7a284a-cilium-config-path\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939523 kubelet[1905]: I0130 13:42:17.939526 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjgg\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-kube-api-access-jmjgg\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939662 kubelet[1905]: I0130 13:42:17.939541 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cc839e17-3de5-4d0b-8fdb-264e461ece18-xtables-lock\") pod \"kube-proxy-ts5qv\" (UID: \"cc839e17-3de5-4d0b-8fdb-264e461ece18\") " pod="kube-system/kube-proxy-ts5qv" Jan 30 13:42:17.939662 kubelet[1905]: I0130 13:42:17.939557 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-bpf-maps\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939662 kubelet[1905]: I0130 13:42:17.939574 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-hostproc\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:17.939753 kubelet[1905]: I0130 13:42:17.939628 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-xtables-lock\") pod \"cilium-b622d\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " pod="kube-system/cilium-b622d" Jan 30 13:42:18.229395 kubelet[1905]: E0130 13:42:18.229262 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:18.229871 kubelet[1905]: E0130 13:42:18.229844 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:18.230433 containerd[1573]: time="2025-01-30T13:42:18.230117390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ts5qv,Uid:cc839e17-3de5-4d0b-8fdb-264e461ece18,Namespace:kube-system,Attempt:0,}" Jan 30 13:42:18.230477 containerd[1573]: time="2025-01-30T13:42:18.230435065Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b622d,Uid:c8de6409-82b8-494d-920d-41141f7a284a,Namespace:kube-system,Attempt:0,}" Jan 30 13:42:18.790033 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3886177423.mount: Deactivated successfully. Jan 30 13:42:18.796462 containerd[1573]: time="2025-01-30T13:42:18.796413868Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:42:18.797450 containerd[1573]: time="2025-01-30T13:42:18.797403824Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:42:18.798172 containerd[1573]: time="2025-01-30T13:42:18.798141878Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jan 30 13:42:18.799084 containerd[1573]: time="2025-01-30T13:42:18.799052947Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 30 13:42:18.800051 containerd[1573]: time="2025-01-30T13:42:18.800001696Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:42:18.802994 containerd[1573]: time="2025-01-30T13:42:18.802971916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:42:18.803901 containerd[1573]: time="2025-01-30T13:42:18.803869940Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 573.668242ms" Jan 30 13:42:18.805844 containerd[1573]: time="2025-01-30T13:42:18.805812213Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 575.288241ms" Jan 30 13:42:18.909562 containerd[1573]: time="2025-01-30T13:42:18.909466802Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:42:18.909562 containerd[1573]: time="2025-01-30T13:42:18.909533257Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:42:18.909729 containerd[1573]: time="2025-01-30T13:42:18.909548816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:18.909729 containerd[1573]: time="2025-01-30T13:42:18.909649615Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:18.910163 containerd[1573]: time="2025-01-30T13:42:18.909975416Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:42:18.910163 containerd[1573]: time="2025-01-30T13:42:18.910042432Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:42:18.910163 containerd[1573]: time="2025-01-30T13:42:18.910057039Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:18.910732 containerd[1573]: time="2025-01-30T13:42:18.910677803Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:18.920553 kubelet[1905]: E0130 13:42:18.920517 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:19.015008 containerd[1573]: time="2025-01-30T13:42:19.014880050Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b622d,Uid:c8de6409-82b8-494d-920d-41141f7a284a,Namespace:kube-system,Attempt:0,} returns sandbox id \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\"" Jan 30 13:42:19.016148 containerd[1573]: time="2025-01-30T13:42:19.016038091Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ts5qv,Uid:cc839e17-3de5-4d0b-8fdb-264e461ece18,Namespace:kube-system,Attempt:0,} returns sandbox id \"cf59e4681f14726b6a5d87a0351ae6a95aec9892b997934710c9efcf3786baa4\"" Jan 30 13:42:19.016364 kubelet[1905]: E0130 13:42:19.016199 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:19.016952 kubelet[1905]: E0130 13:42:19.016897 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:19.017509 containerd[1573]: time="2025-01-30T13:42:19.017444399Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jan 30 13:42:19.920693 kubelet[1905]: E0130 13:42:19.920626 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:20.920842 kubelet[1905]: E0130 13:42:20.920802 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:21.921446 kubelet[1905]: E0130 13:42:21.921378 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:22.921602 kubelet[1905]: E0130 13:42:22.921538 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:23.093669 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2010439247.mount: Deactivated successfully. Jan 30 13:42:23.922677 kubelet[1905]: E0130 13:42:23.922573 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:24.923546 kubelet[1905]: E0130 13:42:24.923484 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:25.924403 kubelet[1905]: E0130 13:42:25.924367 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:26.925442 kubelet[1905]: E0130 13:42:26.925407 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:27.143395 containerd[1573]: time="2025-01-30T13:42:27.143316129Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:27.144024 containerd[1573]: time="2025-01-30T13:42:27.143976948Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Jan 30 13:42:27.145339 containerd[1573]: time="2025-01-30T13:42:27.145309116Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:27.146825 containerd[1573]: time="2025-01-30T13:42:27.146782209Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 8.129294669s" Jan 30 13:42:27.146893 containerd[1573]: time="2025-01-30T13:42:27.146824849Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Jan 30 13:42:27.147871 containerd[1573]: time="2025-01-30T13:42:27.147827629Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\"" Jan 30 13:42:27.149173 containerd[1573]: time="2025-01-30T13:42:27.149143998Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 30 13:42:27.597631 containerd[1573]: time="2025-01-30T13:42:27.597589463Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\"" Jan 30 13:42:27.598318 containerd[1573]: time="2025-01-30T13:42:27.598282874Z" level=info msg="StartContainer for \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\"" Jan 30 13:42:27.648757 containerd[1573]: time="2025-01-30T13:42:27.648722616Z" level=info msg="StartContainer for \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\" returns successfully" Jan 30 13:42:27.926293 kubelet[1905]: E0130 13:42:27.926158 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:28.379554 containerd[1573]: time="2025-01-30T13:42:28.379475809Z" level=info msg="shim disconnected" id=befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3 namespace=k8s.io Jan 30 13:42:28.379554 containerd[1573]: time="2025-01-30T13:42:28.379522757Z" level=warning msg="cleaning up after shim disconnected" id=befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3 namespace=k8s.io Jan 30 13:42:28.379554 containerd[1573]: time="2025-01-30T13:42:28.379530773Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:42:28.436698 kubelet[1905]: E0130 13:42:28.436664 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:28.438179 containerd[1573]: time="2025-01-30T13:42:28.438138012Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 30 13:42:28.457485 containerd[1573]: time="2025-01-30T13:42:28.457447875Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\"" Jan 30 13:42:28.458076 containerd[1573]: time="2025-01-30T13:42:28.457870037Z" level=info msg="StartContainer for \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\"" Jan 30 13:42:28.461119 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3-rootfs.mount: Deactivated successfully. Jan 30 13:42:28.508699 containerd[1573]: time="2025-01-30T13:42:28.508602097Z" level=info msg="StartContainer for \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\" returns successfully" Jan 30 13:42:28.517180 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 13:42:28.517512 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:42:28.517581 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:42:28.527677 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:42:28.537139 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665-rootfs.mount: Deactivated successfully. Jan 30 13:42:28.544636 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:42:28.553385 containerd[1573]: time="2025-01-30T13:42:28.553339616Z" level=info msg="shim disconnected" id=43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665 namespace=k8s.io Jan 30 13:42:28.553588 containerd[1573]: time="2025-01-30T13:42:28.553385582Z" level=warning msg="cleaning up after shim disconnected" id=43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665 namespace=k8s.io Jan 30 13:42:28.553588 containerd[1573]: time="2025-01-30T13:42:28.553396312Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:42:28.926551 kubelet[1905]: E0130 13:42:28.926525 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:29.442922 kubelet[1905]: E0130 13:42:29.442836 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:29.446342 containerd[1573]: time="2025-01-30T13:42:29.446275419Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 30 13:42:29.461371 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4284438949.mount: Deactivated successfully. Jan 30 13:42:29.467974 containerd[1573]: time="2025-01-30T13:42:29.467910162Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\"" Jan 30 13:42:29.468480 containerd[1573]: time="2025-01-30T13:42:29.468454523Z" level=info msg="StartContainer for \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\"" Jan 30 13:42:29.534829 containerd[1573]: time="2025-01-30T13:42:29.534711529Z" level=info msg="StartContainer for \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\" returns successfully" Jan 30 13:42:29.556769 containerd[1573]: time="2025-01-30T13:42:29.556734039Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:29.556891 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825-rootfs.mount: Deactivated successfully. Jan 30 13:42:29.574236 containerd[1573]: time="2025-01-30T13:42:29.574191137Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.9: active requests=0, bytes read=29058337" Jan 30 13:42:29.674871 containerd[1573]: time="2025-01-30T13:42:29.674773074Z" level=info msg="ImageCreate event name:\"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:29.927337 kubelet[1905]: E0130 13:42:29.927287 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:30.137116 containerd[1573]: time="2025-01-30T13:42:30.137044107Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:30.137553 containerd[1573]: time="2025-01-30T13:42:30.137502387Z" level=info msg="shim disconnected" id=c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825 namespace=k8s.io Jan 30 13:42:30.137553 containerd[1573]: time="2025-01-30T13:42:30.137551519Z" level=warning msg="cleaning up after shim disconnected" id=c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825 namespace=k8s.io Jan 30 13:42:30.137635 containerd[1573]: time="2025-01-30T13:42:30.137560436Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:42:30.137905 containerd[1573]: time="2025-01-30T13:42:30.137872010Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.9\" with image id \"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\", repo tag \"registry.k8s.io/kube-proxy:v1.30.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\", size \"29057356\" in 2.989993926s" Jan 30 13:42:30.137905 containerd[1573]: time="2025-01-30T13:42:30.137910662Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\" returns image reference \"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\"" Jan 30 13:42:30.139811 containerd[1573]: time="2025-01-30T13:42:30.139785358Z" level=info msg="CreateContainer within sandbox \"cf59e4681f14726b6a5d87a0351ae6a95aec9892b997934710c9efcf3786baa4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 30 13:42:30.160745 containerd[1573]: time="2025-01-30T13:42:30.160699409Z" level=info msg="CreateContainer within sandbox \"cf59e4681f14726b6a5d87a0351ae6a95aec9892b997934710c9efcf3786baa4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"7a9bcfe9b691939e11ee8058ab1851c1169b2fd6d4ab7bad6edb4e78f141b995\"" Jan 30 13:42:30.161302 containerd[1573]: time="2025-01-30T13:42:30.161261143Z" level=info msg="StartContainer for \"7a9bcfe9b691939e11ee8058ab1851c1169b2fd6d4ab7bad6edb4e78f141b995\"" Jan 30 13:42:30.220104 containerd[1573]: time="2025-01-30T13:42:30.219979872Z" level=info msg="StartContainer for \"7a9bcfe9b691939e11ee8058ab1851c1169b2fd6d4ab7bad6edb4e78f141b995\" returns successfully" Jan 30 13:42:30.445729 kubelet[1905]: E0130 13:42:30.445644 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:30.447561 kubelet[1905]: E0130 13:42:30.447538 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:30.449470 containerd[1573]: time="2025-01-30T13:42:30.449434641Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 30 13:42:30.529146 kubelet[1905]: I0130 13:42:30.529001 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-ts5qv" podStartSLOduration=3.407987416 podStartE2EDuration="14.528983894s" podCreationTimestamp="2025-01-30 13:42:16 +0000 UTC" firstStartedPulling="2025-01-30 13:42:19.017560236 +0000 UTC m=+3.358591018" lastFinishedPulling="2025-01-30 13:42:30.138556704 +0000 UTC m=+14.479587496" observedRunningTime="2025-01-30 13:42:30.528913633 +0000 UTC m=+14.869944445" watchObservedRunningTime="2025-01-30 13:42:30.528983894 +0000 UTC m=+14.870014676" Jan 30 13:42:30.541656 containerd[1573]: time="2025-01-30T13:42:30.541618238Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\"" Jan 30 13:42:30.542077 containerd[1573]: time="2025-01-30T13:42:30.542053164Z" level=info msg="StartContainer for \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\"" Jan 30 13:42:30.592281 containerd[1573]: time="2025-01-30T13:42:30.592242447Z" level=info msg="StartContainer for \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\" returns successfully" Jan 30 13:42:30.613223 containerd[1573]: time="2025-01-30T13:42:30.613166156Z" level=info msg="shim disconnected" id=cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf namespace=k8s.io Jan 30 13:42:30.613223 containerd[1573]: time="2025-01-30T13:42:30.613216450Z" level=warning msg="cleaning up after shim disconnected" id=cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf namespace=k8s.io Jan 30 13:42:30.613223 containerd[1573]: time="2025-01-30T13:42:30.613225017Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:42:30.928016 kubelet[1905]: E0130 13:42:30.927986 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:31.451136 kubelet[1905]: E0130 13:42:31.451107 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:31.451136 kubelet[1905]: E0130 13:42:31.451107 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:31.452912 containerd[1573]: time="2025-01-30T13:42:31.452879478Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 30 13:42:31.461037 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf-rootfs.mount: Deactivated successfully. Jan 30 13:42:31.470564 containerd[1573]: time="2025-01-30T13:42:31.470512517Z" level=info msg="CreateContainer within sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\"" Jan 30 13:42:31.470963 containerd[1573]: time="2025-01-30T13:42:31.470923287Z" level=info msg="StartContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\"" Jan 30 13:42:31.522935 containerd[1573]: time="2025-01-30T13:42:31.522899291Z" level=info msg="StartContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" returns successfully" Jan 30 13:42:31.620343 kubelet[1905]: I0130 13:42:31.620054 1905 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Jan 30 13:42:31.928381 kubelet[1905]: E0130 13:42:31.928340 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:31.956124 kernel: Initializing XFRM netlink socket Jan 30 13:42:32.455388 kubelet[1905]: E0130 13:42:32.455352 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:32.465918 kubelet[1905]: I0130 13:42:32.465874 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-b622d" podStartSLOduration=8.335195319 podStartE2EDuration="16.465863134s" podCreationTimestamp="2025-01-30 13:42:16 +0000 UTC" firstStartedPulling="2025-01-30 13:42:19.016988263 +0000 UTC m=+3.358019045" lastFinishedPulling="2025-01-30 13:42:27.147656078 +0000 UTC m=+11.488686860" observedRunningTime="2025-01-30 13:42:32.465694608 +0000 UTC m=+16.806725410" watchObservedRunningTime="2025-01-30 13:42:32.465863134 +0000 UTC m=+16.806893916" Jan 30 13:42:32.929079 kubelet[1905]: E0130 13:42:32.929017 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:33.128584 kubelet[1905]: I0130 13:42:33.128520 1905 topology_manager.go:215] "Topology Admit Handler" podUID="532e67d9-ef41-4920-9274-adc6f6429d4f" podNamespace="default" podName="nginx-deployment-85f456d6dd-4hwxz" Jan 30 13:42:33.233741 kubelet[1905]: I0130 13:42:33.233604 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcgwf\" (UniqueName: \"kubernetes.io/projected/532e67d9-ef41-4920-9274-adc6f6429d4f-kube-api-access-zcgwf\") pod \"nginx-deployment-85f456d6dd-4hwxz\" (UID: \"532e67d9-ef41-4920-9274-adc6f6429d4f\") " pod="default/nginx-deployment-85f456d6dd-4hwxz" Jan 30 13:42:33.432556 containerd[1573]: time="2025-01-30T13:42:33.432498711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-4hwxz,Uid:532e67d9-ef41-4920-9274-adc6f6429d4f,Namespace:default,Attempt:0,}" Jan 30 13:42:33.456542 kubelet[1905]: E0130 13:42:33.456510 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:33.783202 systemd-networkd[1244]: cilium_host: Link UP Jan 30 13:42:33.783851 systemd-networkd[1244]: cilium_net: Link UP Jan 30 13:42:33.784149 systemd-networkd[1244]: cilium_net: Gained carrier Jan 30 13:42:33.784384 systemd-networkd[1244]: cilium_host: Gained carrier Jan 30 13:42:33.880277 systemd-networkd[1244]: cilium_vxlan: Link UP Jan 30 13:42:33.880287 systemd-networkd[1244]: cilium_vxlan: Gained carrier Jan 30 13:42:33.929497 kubelet[1905]: E0130 13:42:33.929454 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:34.098121 kernel: NET: Registered PF_ALG protocol family Jan 30 13:42:34.119267 systemd-networkd[1244]: cilium_host: Gained IPv6LL Jan 30 13:42:34.457272 systemd-networkd[1244]: cilium_net: Gained IPv6LL Jan 30 13:42:34.458045 kubelet[1905]: E0130 13:42:34.457665 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:34.686048 systemd-networkd[1244]: lxc_health: Link UP Jan 30 13:42:34.697264 systemd-networkd[1244]: lxc_health: Gained carrier Jan 30 13:42:34.788855 systemd-networkd[1244]: lxc270b2c004362: Link UP Jan 30 13:42:34.800125 kernel: eth0: renamed from tmp85257 Jan 30 13:42:34.808680 systemd-networkd[1244]: lxc270b2c004362: Gained carrier Jan 30 13:42:34.930068 kubelet[1905]: E0130 13:42:34.929987 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:35.607335 systemd-networkd[1244]: cilium_vxlan: Gained IPv6LL Jan 30 13:42:35.918853 kubelet[1905]: E0130 13:42:35.918717 1905 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:35.930300 kubelet[1905]: E0130 13:42:35.930251 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:35.991381 systemd-networkd[1244]: lxc_health: Gained IPv6LL Jan 30 13:42:36.232390 kubelet[1905]: E0130 13:42:36.232269 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:36.460001 kubelet[1905]: E0130 13:42:36.459962 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:42:36.823367 systemd-networkd[1244]: lxc270b2c004362: Gained IPv6LL Jan 30 13:42:36.930615 kubelet[1905]: E0130 13:42:36.930539 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:37.931080 kubelet[1905]: E0130 13:42:37.931016 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:38.364071 containerd[1573]: time="2025-01-30T13:42:38.363575567Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:42:38.364071 containerd[1573]: time="2025-01-30T13:42:38.364036211Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:42:38.364071 containerd[1573]: time="2025-01-30T13:42:38.364047553Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:38.364612 containerd[1573]: time="2025-01-30T13:42:38.364154919Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:38.385930 systemd-resolved[1461]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 30 13:42:38.411605 containerd[1573]: time="2025-01-30T13:42:38.411554276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-4hwxz,Uid:532e67d9-ef41-4920-9274-adc6f6429d4f,Namespace:default,Attempt:0,} returns sandbox id \"85257a41d175ccb824a98d6f1d035a57831dc3a84bd91b298de15f379604d3e3\"" Jan 30 13:42:38.412872 containerd[1573]: time="2025-01-30T13:42:38.412848950Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Jan 30 13:42:38.931711 kubelet[1905]: E0130 13:42:38.931640 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:39.932565 kubelet[1905]: E0130 13:42:39.932527 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:40.933370 kubelet[1905]: E0130 13:42:40.933323 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:41.495356 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3855446065.mount: Deactivated successfully. Jan 30 13:42:41.935229 kubelet[1905]: E0130 13:42:41.935178 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:42.667359 containerd[1573]: time="2025-01-30T13:42:42.667283551Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:42.667976 containerd[1573]: time="2025-01-30T13:42:42.667916970Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=71015561" Jan 30 13:42:42.669111 containerd[1573]: time="2025-01-30T13:42:42.669060813Z" level=info msg="ImageCreate event name:\"sha256:0dcfd986e814f68db775fba6b61fbaec3761562dc2ae3043d38dbff123e1bb1e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:42.671509 containerd[1573]: time="2025-01-30T13:42:42.671470393Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:2ffeb5a7ca68f2017f0bc48251750a6e40fcd3c341b94a22fc7812dcabbb84db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:42.672409 containerd[1573]: time="2025-01-30T13:42:42.672378225Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:0dcfd986e814f68db775fba6b61fbaec3761562dc2ae3043d38dbff123e1bb1e\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:2ffeb5a7ca68f2017f0bc48251750a6e40fcd3c341b94a22fc7812dcabbb84db\", size \"71015439\" in 4.259494108s" Jan 30 13:42:42.672474 containerd[1573]: time="2025-01-30T13:42:42.672408733Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:0dcfd986e814f68db775fba6b61fbaec3761562dc2ae3043d38dbff123e1bb1e\"" Jan 30 13:42:42.674700 containerd[1573]: time="2025-01-30T13:42:42.674650653Z" level=info msg="CreateContainer within sandbox \"85257a41d175ccb824a98d6f1d035a57831dc3a84bd91b298de15f379604d3e3\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Jan 30 13:42:42.692056 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount229737476.mount: Deactivated successfully. Jan 30 13:42:42.836116 containerd[1573]: time="2025-01-30T13:42:42.836045428Z" level=info msg="CreateContainer within sandbox \"85257a41d175ccb824a98d6f1d035a57831dc3a84bd91b298de15f379604d3e3\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"5784fb64e84f730179d1f781ee35fae24fe9e02c519091ad55548ff6991a0092\"" Jan 30 13:42:42.836685 containerd[1573]: time="2025-01-30T13:42:42.836649561Z" level=info msg="StartContainer for \"5784fb64e84f730179d1f781ee35fae24fe9e02c519091ad55548ff6991a0092\"" Jan 30 13:42:42.935999 kubelet[1905]: E0130 13:42:42.935837 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:42.983814 containerd[1573]: time="2025-01-30T13:42:42.983745725Z" level=info msg="StartContainer for \"5784fb64e84f730179d1f781ee35fae24fe9e02c519091ad55548ff6991a0092\" returns successfully" Jan 30 13:42:43.538551 kubelet[1905]: I0130 13:42:43.538469 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nginx-deployment-85f456d6dd-4hwxz" podStartSLOduration=6.277794986 podStartE2EDuration="10.538451686s" podCreationTimestamp="2025-01-30 13:42:33 +0000 UTC" firstStartedPulling="2025-01-30 13:42:38.412633446 +0000 UTC m=+22.753664228" lastFinishedPulling="2025-01-30 13:42:42.673290146 +0000 UTC m=+27.014320928" observedRunningTime="2025-01-30 13:42:43.538232759 +0000 UTC m=+27.879263541" watchObservedRunningTime="2025-01-30 13:42:43.538451686 +0000 UTC m=+27.879482468" Jan 30 13:42:43.936723 kubelet[1905]: E0130 13:42:43.936583 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:44.936814 kubelet[1905]: E0130 13:42:44.936748 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:45.428689 kubelet[1905]: I0130 13:42:45.428646 1905 topology_manager.go:215] "Topology Admit Handler" podUID="16e827a0-d5b3-46b7-a451-4837db863e1c" podNamespace="default" podName="nfs-server-provisioner-0" Jan 30 13:42:45.497895 kubelet[1905]: I0130 13:42:45.497829 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/16e827a0-d5b3-46b7-a451-4837db863e1c-data\") pod \"nfs-server-provisioner-0\" (UID: \"16e827a0-d5b3-46b7-a451-4837db863e1c\") " pod="default/nfs-server-provisioner-0" Jan 30 13:42:45.497895 kubelet[1905]: I0130 13:42:45.497879 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n679\" (UniqueName: \"kubernetes.io/projected/16e827a0-d5b3-46b7-a451-4837db863e1c-kube-api-access-8n679\") pod \"nfs-server-provisioner-0\" (UID: \"16e827a0-d5b3-46b7-a451-4837db863e1c\") " pod="default/nfs-server-provisioner-0" Jan 30 13:42:45.732643 containerd[1573]: time="2025-01-30T13:42:45.732499074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:16e827a0-d5b3-46b7-a451-4837db863e1c,Namespace:default,Attempt:0,}" Jan 30 13:42:45.760870 systemd-networkd[1244]: lxc51c598e7469c: Link UP Jan 30 13:42:45.770212 kernel: eth0: renamed from tmpaffd3 Jan 30 13:42:45.776702 systemd-networkd[1244]: lxc51c598e7469c: Gained carrier Jan 30 13:42:45.937198 kubelet[1905]: E0130 13:42:45.937151 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:46.054157 containerd[1573]: time="2025-01-30T13:42:46.054012251Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:42:46.054157 containerd[1573]: time="2025-01-30T13:42:46.054078036Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:42:46.054157 containerd[1573]: time="2025-01-30T13:42:46.054108083Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:46.054343 containerd[1573]: time="2025-01-30T13:42:46.054200859Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:42:46.087529 systemd-resolved[1461]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 30 13:42:46.111489 containerd[1573]: time="2025-01-30T13:42:46.111446260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:16e827a0-d5b3-46b7-a451-4837db863e1c,Namespace:default,Attempt:0,} returns sandbox id \"affd3a40c0abc96a7fb4129e3ca197a5e18fc0b37c093deb31fdb30278acd8b6\"" Jan 30 13:42:46.113120 containerd[1573]: time="2025-01-30T13:42:46.113078272Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Jan 30 13:42:46.935275 systemd-networkd[1244]: lxc51c598e7469c: Gained IPv6LL Jan 30 13:42:46.937544 kubelet[1905]: E0130 13:42:46.937515 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:47.938629 kubelet[1905]: E0130 13:42:47.938586 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:48.685203 update_engine[1550]: I20250130 13:42:48.685132 1550 update_attempter.cc:509] Updating boot flags... Jan 30 13:42:48.766118 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3118) Jan 30 13:42:48.869439 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3118) Jan 30 13:42:48.915253 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3118) Jan 30 13:42:48.939319 kubelet[1905]: E0130 13:42:48.939221 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:49.187982 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3809939181.mount: Deactivated successfully. Jan 30 13:42:49.939918 kubelet[1905]: E0130 13:42:49.939869 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:50.940391 kubelet[1905]: E0130 13:42:50.940315 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:51.705536 containerd[1573]: time="2025-01-30T13:42:51.705461656Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:51.706560 containerd[1573]: time="2025-01-30T13:42:51.706519891Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=91039406" Jan 30 13:42:51.708027 containerd[1573]: time="2025-01-30T13:42:51.707980597Z" level=info msg="ImageCreate event name:\"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:51.712167 containerd[1573]: time="2025-01-30T13:42:51.712130177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:42:51.712842 containerd[1573]: time="2025-01-30T13:42:51.712811949Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"91036984\" in 5.599682208s" Jan 30 13:42:51.712876 containerd[1573]: time="2025-01-30T13:42:51.712847094Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Jan 30 13:42:51.714791 containerd[1573]: time="2025-01-30T13:42:51.714762583Z" level=info msg="CreateContainer within sandbox \"affd3a40c0abc96a7fb4129e3ca197a5e18fc0b37c093deb31fdb30278acd8b6\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Jan 30 13:42:51.727685 containerd[1573]: time="2025-01-30T13:42:51.727644421Z" level=info msg="CreateContainer within sandbox \"affd3a40c0abc96a7fb4129e3ca197a5e18fc0b37c093deb31fdb30278acd8b6\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"f8b523df4e44383978438b1c50f9b6da673f7fd9b5758d82a7d77d9ce435c628\"" Jan 30 13:42:51.728141 containerd[1573]: time="2025-01-30T13:42:51.728085376Z" level=info msg="StartContainer for \"f8b523df4e44383978438b1c50f9b6da673f7fd9b5758d82a7d77d9ce435c628\"" Jan 30 13:42:51.814956 containerd[1573]: time="2025-01-30T13:42:51.814908500Z" level=info msg="StartContainer for \"f8b523df4e44383978438b1c50f9b6da673f7fd9b5758d82a7d77d9ce435c628\" returns successfully" Jan 30 13:42:51.941311 kubelet[1905]: E0130 13:42:51.941243 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:52.941437 kubelet[1905]: E0130 13:42:52.941370 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:53.941925 kubelet[1905]: E0130 13:42:53.941882 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:54.942878 kubelet[1905]: E0130 13:42:54.942811 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:55.919184 kubelet[1905]: E0130 13:42:55.919064 1905 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:55.943738 kubelet[1905]: E0130 13:42:55.943683 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:56.944340 kubelet[1905]: E0130 13:42:56.944271 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:57.944470 kubelet[1905]: E0130 13:42:57.944415 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:58.945243 kubelet[1905]: E0130 13:42:58.945173 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:42:59.945874 kubelet[1905]: E0130 13:42:59.945805 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:00.946983 kubelet[1905]: E0130 13:43:00.946909 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:01.191911 kubelet[1905]: I0130 13:43:01.191820 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=10.59103723 podStartE2EDuration="16.191801981s" podCreationTimestamp="2025-01-30 13:42:45 +0000 UTC" firstStartedPulling="2025-01-30 13:42:46.112789614 +0000 UTC m=+30.453820396" lastFinishedPulling="2025-01-30 13:42:51.713554365 +0000 UTC m=+36.054585147" observedRunningTime="2025-01-30 13:42:52.497233329 +0000 UTC m=+36.838264131" watchObservedRunningTime="2025-01-30 13:43:01.191801981 +0000 UTC m=+45.532832763" Jan 30 13:43:01.192120 kubelet[1905]: I0130 13:43:01.192067 1905 topology_manager.go:215] "Topology Admit Handler" podUID="c334ec21-e6c1-4ef7-8fb8-b193b69350f2" podNamespace="default" podName="test-pod-1" Jan 30 13:43:01.385308 kubelet[1905]: I0130 13:43:01.385244 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-745b02d9-bc6b-4066-b3cb-f58e61415b8b\" (UniqueName: \"kubernetes.io/nfs/c334ec21-e6c1-4ef7-8fb8-b193b69350f2-pvc-745b02d9-bc6b-4066-b3cb-f58e61415b8b\") pod \"test-pod-1\" (UID: \"c334ec21-e6c1-4ef7-8fb8-b193b69350f2\") " pod="default/test-pod-1" Jan 30 13:43:01.385308 kubelet[1905]: I0130 13:43:01.385297 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6zk\" (UniqueName: \"kubernetes.io/projected/c334ec21-e6c1-4ef7-8fb8-b193b69350f2-kube-api-access-5n6zk\") pod \"test-pod-1\" (UID: \"c334ec21-e6c1-4ef7-8fb8-b193b69350f2\") " pod="default/test-pod-1" Jan 30 13:43:01.517124 kernel: FS-Cache: Loaded Jan 30 13:43:01.587391 kernel: RPC: Registered named UNIX socket transport module. Jan 30 13:43:01.587504 kernel: RPC: Registered udp transport module. Jan 30 13:43:01.587527 kernel: RPC: Registered tcp transport module. Jan 30 13:43:01.587542 kernel: RPC: Registered tcp-with-tls transport module. Jan 30 13:43:01.588862 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Jan 30 13:43:01.883437 kernel: NFS: Registering the id_resolver key type Jan 30 13:43:01.883582 kernel: Key type id_resolver registered Jan 30 13:43:01.883604 kernel: Key type id_legacy registered Jan 30 13:43:01.911331 nfsidmap[3292]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Jan 30 13:43:01.916064 nfsidmap[3295]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Jan 30 13:43:01.948039 kubelet[1905]: E0130 13:43:01.947992 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:02.096627 containerd[1573]: time="2025-01-30T13:43:02.096570257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:c334ec21-e6c1-4ef7-8fb8-b193b69350f2,Namespace:default,Attempt:0,}" Jan 30 13:43:02.124738 systemd-networkd[1244]: lxc8a5d56ab1f55: Link UP Jan 30 13:43:02.136138 kernel: eth0: renamed from tmp23990 Jan 30 13:43:02.144688 systemd-networkd[1244]: lxc8a5d56ab1f55: Gained carrier Jan 30 13:43:02.336287 containerd[1573]: time="2025-01-30T13:43:02.336167144Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:43:02.336287 containerd[1573]: time="2025-01-30T13:43:02.336240863Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:43:02.336287 containerd[1573]: time="2025-01-30T13:43:02.336259147Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:02.336445 containerd[1573]: time="2025-01-30T13:43:02.336378012Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:02.371299 systemd-resolved[1461]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 30 13:43:02.397706 containerd[1573]: time="2025-01-30T13:43:02.397422121Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:c334ec21-e6c1-4ef7-8fb8-b193b69350f2,Namespace:default,Attempt:0,} returns sandbox id \"239907206db4572011481c3fd24e3e9058e929ef633992f4108dc861b7fb4bf7\"" Jan 30 13:43:02.398984 containerd[1573]: time="2025-01-30T13:43:02.398953046Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Jan 30 13:43:02.907669 containerd[1573]: time="2025-01-30T13:43:02.907623732Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:43:02.908500 containerd[1573]: time="2025-01-30T13:43:02.908462052Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Jan 30 13:43:02.910925 containerd[1573]: time="2025-01-30T13:43:02.910894896Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:0dcfd986e814f68db775fba6b61fbaec3761562dc2ae3043d38dbff123e1bb1e\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:2ffeb5a7ca68f2017f0bc48251750a6e40fcd3c341b94a22fc7812dcabbb84db\", size \"71015439\" in 511.906844ms" Jan 30 13:43:02.910925 containerd[1573]: time="2025-01-30T13:43:02.910924692Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:0dcfd986e814f68db775fba6b61fbaec3761562dc2ae3043d38dbff123e1bb1e\"" Jan 30 13:43:02.912645 containerd[1573]: time="2025-01-30T13:43:02.912610379Z" level=info msg="CreateContainer within sandbox \"239907206db4572011481c3fd24e3e9058e929ef633992f4108dc861b7fb4bf7\" for container &ContainerMetadata{Name:test,Attempt:0,}" Jan 30 13:43:02.925660 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1226313319.mount: Deactivated successfully. Jan 30 13:43:02.927018 containerd[1573]: time="2025-01-30T13:43:02.926966083Z" level=info msg="CreateContainer within sandbox \"239907206db4572011481c3fd24e3e9058e929ef633992f4108dc861b7fb4bf7\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"a17172707dd7bcdb24b6c4cb74660774de27aa612b1f79a907a87cc6e617b80d\"" Jan 30 13:43:02.927426 containerd[1573]: time="2025-01-30T13:43:02.927401944Z" level=info msg="StartContainer for \"a17172707dd7bcdb24b6c4cb74660774de27aa612b1f79a907a87cc6e617b80d\"" Jan 30 13:43:02.949184 kubelet[1905]: E0130 13:43:02.949140 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:02.975026 containerd[1573]: time="2025-01-30T13:43:02.974847647Z" level=info msg="StartContainer for \"a17172707dd7bcdb24b6c4cb74660774de27aa612b1f79a907a87cc6e617b80d\" returns successfully" Jan 30 13:43:03.518680 kubelet[1905]: I0130 13:43:03.518635 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=18.005840367 podStartE2EDuration="18.518618863s" podCreationTimestamp="2025-01-30 13:42:45 +0000 UTC" firstStartedPulling="2025-01-30 13:43:02.398737239 +0000 UTC m=+46.739768022" lastFinishedPulling="2025-01-30 13:43:02.911515736 +0000 UTC m=+47.252546518" observedRunningTime="2025-01-30 13:43:03.518221595 +0000 UTC m=+47.859252378" watchObservedRunningTime="2025-01-30 13:43:03.518618863 +0000 UTC m=+47.859649646" Jan 30 13:43:03.949954 kubelet[1905]: E0130 13:43:03.949882 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:03.959360 systemd-networkd[1244]: lxc8a5d56ab1f55: Gained IPv6LL Jan 30 13:43:04.950610 kubelet[1905]: E0130 13:43:04.950551 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:05.951074 kubelet[1905]: E0130 13:43:05.950992 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:06.951261 kubelet[1905]: E0130 13:43:06.951184 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:07.952175 kubelet[1905]: E0130 13:43:07.952081 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:08.237290 systemd[1]: run-containerd-runc-k8s.io-e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945-runc.xH4nX2.mount: Deactivated successfully. Jan 30 13:43:08.260444 containerd[1573]: time="2025-01-30T13:43:08.260392788Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 13:43:08.267985 containerd[1573]: time="2025-01-30T13:43:08.267952312Z" level=info msg="StopContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" with timeout 2 (s)" Jan 30 13:43:08.268192 containerd[1573]: time="2025-01-30T13:43:08.268167717Z" level=info msg="Stop container \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" with signal terminated" Jan 30 13:43:08.274794 systemd-networkd[1244]: lxc_health: Link DOWN Jan 30 13:43:08.274805 systemd-networkd[1244]: lxc_health: Lost carrier Jan 30 13:43:08.322889 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945-rootfs.mount: Deactivated successfully. Jan 30 13:43:08.340083 containerd[1573]: time="2025-01-30T13:43:08.340020624Z" level=info msg="shim disconnected" id=e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945 namespace=k8s.io Jan 30 13:43:08.340083 containerd[1573]: time="2025-01-30T13:43:08.340075186Z" level=warning msg="cleaning up after shim disconnected" id=e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945 namespace=k8s.io Jan 30 13:43:08.340083 containerd[1573]: time="2025-01-30T13:43:08.340085766Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:08.357149 containerd[1573]: time="2025-01-30T13:43:08.357082016Z" level=info msg="StopContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" returns successfully" Jan 30 13:43:08.357722 containerd[1573]: time="2025-01-30T13:43:08.357696322Z" level=info msg="StopPodSandbox for \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\"" Jan 30 13:43:08.357776 containerd[1573]: time="2025-01-30T13:43:08.357730997Z" level=info msg="Container to stop \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:43:08.357776 containerd[1573]: time="2025-01-30T13:43:08.357748690Z" level=info msg="Container to stop \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:43:08.357776 containerd[1573]: time="2025-01-30T13:43:08.357762236Z" level=info msg="Container to stop \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:43:08.357904 containerd[1573]: time="2025-01-30T13:43:08.357774178Z" level=info msg="Container to stop \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:43:08.357904 containerd[1573]: time="2025-01-30T13:43:08.357786241Z" level=info msg="Container to stop \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:43:08.359942 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b-shm.mount: Deactivated successfully. Jan 30 13:43:08.382497 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b-rootfs.mount: Deactivated successfully. Jan 30 13:43:08.385980 containerd[1573]: time="2025-01-30T13:43:08.385941377Z" level=info msg="shim disconnected" id=8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b namespace=k8s.io Jan 30 13:43:08.385980 containerd[1573]: time="2025-01-30T13:43:08.385978887Z" level=warning msg="cleaning up after shim disconnected" id=8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b namespace=k8s.io Jan 30 13:43:08.385980 containerd[1573]: time="2025-01-30T13:43:08.385988856Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:08.398388 containerd[1573]: time="2025-01-30T13:43:08.398329757Z" level=warning msg="cleanup warnings time=\"2025-01-30T13:43:08Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jan 30 13:43:08.399618 containerd[1573]: time="2025-01-30T13:43:08.399583516Z" level=info msg="TearDown network for sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" successfully" Jan 30 13:43:08.399618 containerd[1573]: time="2025-01-30T13:43:08.399605828Z" level=info msg="StopPodSandbox for \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" returns successfully" Jan 30 13:43:08.520535 kubelet[1905]: I0130 13:43:08.520419 1905 scope.go:117] "RemoveContainer" containerID="e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945" Jan 30 13:43:08.521592 containerd[1573]: time="2025-01-30T13:43:08.521561858Z" level=info msg="RemoveContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\"" Jan 30 13:43:08.525023 containerd[1573]: time="2025-01-30T13:43:08.524991522Z" level=info msg="RemoveContainer for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" returns successfully" Jan 30 13:43:08.525211 kubelet[1905]: I0130 13:43:08.525178 1905 scope.go:117] "RemoveContainer" containerID="cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf" Jan 30 13:43:08.525253 kubelet[1905]: I0130 13:43:08.525197 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-hubble-tls\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525287 kubelet[1905]: I0130 13:43:08.525259 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-etc-cni-netd\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525287 kubelet[1905]: I0130 13:43:08.525274 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-lib-modules\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525333 kubelet[1905]: I0130 13:43:08.525294 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c8de6409-82b8-494d-920d-41141f7a284a-cilium-config-path\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525333 kubelet[1905]: I0130 13:43:08.525299 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525333 kubelet[1905]: I0130 13:43:08.525309 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-run\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525333 kubelet[1905]: I0130 13:43:08.525323 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525425 kubelet[1905]: I0130 13:43:08.525340 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525425 kubelet[1905]: I0130 13:43:08.525345 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-cgroup\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525425 kubelet[1905]: I0130 13:43:08.525369 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cni-path\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525425 kubelet[1905]: I0130 13:43:08.525390 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjgg\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-kube-api-access-jmjgg\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525425 kubelet[1905]: I0130 13:43:08.525408 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-bpf-maps\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525541 kubelet[1905]: I0130 13:43:08.525426 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cni-path" (OuterVolumeSpecName: "cni-path") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525541 kubelet[1905]: I0130 13:43:08.525427 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-net\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525541 kubelet[1905]: I0130 13:43:08.525447 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525541 kubelet[1905]: I0130 13:43:08.525453 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-hostproc\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525541 kubelet[1905]: I0130 13:43:08.525462 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-hostproc" (OuterVolumeSpecName: "hostproc") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525476 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525481 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c8de6409-82b8-494d-920d-41141f7a284a-clustermesh-secrets\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525502 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-kernel\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525519 1905 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-xtables-lock\") pod \"c8de6409-82b8-494d-920d-41141f7a284a\" (UID: \"c8de6409-82b8-494d-920d-41141f7a284a\") " Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525549 1905 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-run\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525561 1905 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cilium-cgroup\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525657 kubelet[1905]: I0130 13:43:08.525572 1905 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-cni-path\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525581 1905 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-net\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525593 1905 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-hostproc\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525602 1905 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-etc-cni-netd\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525612 1905 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-lib-modules\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525631 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.525803 kubelet[1905]: I0130 13:43:08.525653 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.528656 kubelet[1905]: I0130 13:43:08.528434 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 30 13:43:08.528656 kubelet[1905]: I0130 13:43:08.528469 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-kube-api-access-jmjgg" (OuterVolumeSpecName: "kube-api-access-jmjgg") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "kube-api-access-jmjgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 30 13:43:08.528656 kubelet[1905]: I0130 13:43:08.528507 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:43:08.528821 kubelet[1905]: I0130 13:43:08.528699 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8de6409-82b8-494d-920d-41141f7a284a-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 30 13:43:08.529218 containerd[1573]: time="2025-01-30T13:43:08.529188830Z" level=info msg="RemoveContainer for \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\"" Jan 30 13:43:08.529574 kubelet[1905]: I0130 13:43:08.529540 1905 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8de6409-82b8-494d-920d-41141f7a284a-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "c8de6409-82b8-494d-920d-41141f7a284a" (UID: "c8de6409-82b8-494d-920d-41141f7a284a"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 30 13:43:08.532167 containerd[1573]: time="2025-01-30T13:43:08.532142837Z" level=info msg="RemoveContainer for \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\" returns successfully" Jan 30 13:43:08.532284 kubelet[1905]: I0130 13:43:08.532266 1905 scope.go:117] "RemoveContainer" containerID="c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825" Jan 30 13:43:08.532960 containerd[1573]: time="2025-01-30T13:43:08.532942752Z" level=info msg="RemoveContainer for \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\"" Jan 30 13:43:08.535508 containerd[1573]: time="2025-01-30T13:43:08.535485978Z" level=info msg="RemoveContainer for \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\" returns successfully" Jan 30 13:43:08.535608 kubelet[1905]: I0130 13:43:08.535593 1905 scope.go:117] "RemoveContainer" containerID="43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665" Jan 30 13:43:08.536337 containerd[1573]: time="2025-01-30T13:43:08.536316891Z" level=info msg="RemoveContainer for \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\"" Jan 30 13:43:08.538859 containerd[1573]: time="2025-01-30T13:43:08.538835370Z" level=info msg="RemoveContainer for \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\" returns successfully" Jan 30 13:43:08.538956 kubelet[1905]: I0130 13:43:08.538943 1905 scope.go:117] "RemoveContainer" containerID="befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3" Jan 30 13:43:08.539713 containerd[1573]: time="2025-01-30T13:43:08.539680340Z" level=info msg="RemoveContainer for \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\"" Jan 30 13:43:08.542211 containerd[1573]: time="2025-01-30T13:43:08.542186786Z" level=info msg="RemoveContainer for \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\" returns successfully" Jan 30 13:43:08.542361 kubelet[1905]: I0130 13:43:08.542293 1905 scope.go:117] "RemoveContainer" containerID="e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945" Jan 30 13:43:08.542467 containerd[1573]: time="2025-01-30T13:43:08.542437128Z" level=error msg="ContainerStatus for \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\": not found" Jan 30 13:43:08.542543 kubelet[1905]: E0130 13:43:08.542522 1905 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\": not found" containerID="e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945" Jan 30 13:43:08.542615 kubelet[1905]: I0130 13:43:08.542548 1905 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945"} err="failed to get container status \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\": rpc error: code = NotFound desc = an error occurred when try to find container \"e442c8ca7921a44c8b10d7f0a7405d25cd74bdb2df931d45a9ca43c4f23d0945\": not found" Jan 30 13:43:08.542615 kubelet[1905]: I0130 13:43:08.542607 1905 scope.go:117] "RemoveContainer" containerID="cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf" Jan 30 13:43:08.542788 containerd[1573]: time="2025-01-30T13:43:08.542748083Z" level=error msg="ContainerStatus for \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\": not found" Jan 30 13:43:08.542889 kubelet[1905]: E0130 13:43:08.542869 1905 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\": not found" containerID="cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf" Jan 30 13:43:08.542959 kubelet[1905]: I0130 13:43:08.542891 1905 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf"} err="failed to get container status \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\": rpc error: code = NotFound desc = an error occurred when try to find container \"cbb0c6d1f1df5d7f8acf7502edffaccf0503508b3462184da5e2d174607ac3bf\": not found" Jan 30 13:43:08.542959 kubelet[1905]: I0130 13:43:08.542914 1905 scope.go:117] "RemoveContainer" containerID="c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825" Jan 30 13:43:08.543124 containerd[1573]: time="2025-01-30T13:43:08.543076441Z" level=error msg="ContainerStatus for \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\": not found" Jan 30 13:43:08.543214 kubelet[1905]: E0130 13:43:08.543199 1905 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\": not found" containerID="c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825" Jan 30 13:43:08.543250 kubelet[1905]: I0130 13:43:08.543216 1905 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825"} err="failed to get container status \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\": rpc error: code = NotFound desc = an error occurred when try to find container \"c3841edcea759080c8a9a578b8d5d8990328bc10757c1f843bbc0d267579d825\": not found" Jan 30 13:43:08.543250 kubelet[1905]: I0130 13:43:08.543227 1905 scope.go:117] "RemoveContainer" containerID="43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665" Jan 30 13:43:08.543384 containerd[1573]: time="2025-01-30T13:43:08.543357909Z" level=error msg="ContainerStatus for \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\": not found" Jan 30 13:43:08.543490 kubelet[1905]: E0130 13:43:08.543461 1905 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\": not found" containerID="43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665" Jan 30 13:43:08.543523 kubelet[1905]: I0130 13:43:08.543484 1905 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665"} err="failed to get container status \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\": rpc error: code = NotFound desc = an error occurred when try to find container \"43ea4f031c40d1c63bc3bbaa7088da718ce38b1b8bf547db45be5e860629a665\": not found" Jan 30 13:43:08.543523 kubelet[1905]: I0130 13:43:08.543501 1905 scope.go:117] "RemoveContainer" containerID="befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3" Jan 30 13:43:08.543653 containerd[1573]: time="2025-01-30T13:43:08.543630292Z" level=error msg="ContainerStatus for \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\": not found" Jan 30 13:43:08.543745 kubelet[1905]: E0130 13:43:08.543725 1905 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\": not found" containerID="befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3" Jan 30 13:43:08.543774 kubelet[1905]: I0130 13:43:08.543748 1905 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3"} err="failed to get container status \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\": rpc error: code = NotFound desc = an error occurred when try to find container \"befcbfa2a2c172e94138c2bde752b9461c2efc014ef49ce7bc8ba5277e8422f3\": not found" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626041 1905 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-host-proc-sys-kernel\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626075 1905 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-xtables-lock\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626083 1905 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c8de6409-82b8-494d-920d-41141f7a284a-clustermesh-secrets\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626109 1905 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c8de6409-82b8-494d-920d-41141f7a284a-cilium-config-path\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626119 1905 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-hubble-tls\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626112 kubelet[1905]: I0130 13:43:08.626126 1905 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-jmjgg\" (UniqueName: \"kubernetes.io/projected/c8de6409-82b8-494d-920d-41141f7a284a-kube-api-access-jmjgg\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.626337 kubelet[1905]: I0130 13:43:08.626136 1905 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c8de6409-82b8-494d-920d-41141f7a284a-bpf-maps\") on node \"10.0.0.28\" DevicePath \"\"" Jan 30 13:43:08.952962 kubelet[1905]: E0130 13:43:08.952913 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:09.232837 systemd[1]: var-lib-kubelet-pods-c8de6409\x2d82b8\x2d494d\x2d920d\x2d41141f7a284a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djmjgg.mount: Deactivated successfully. Jan 30 13:43:09.233066 systemd[1]: var-lib-kubelet-pods-c8de6409\x2d82b8\x2d494d\x2d920d\x2d41141f7a284a-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jan 30 13:43:09.233250 systemd[1]: var-lib-kubelet-pods-c8de6409\x2d82b8\x2d494d\x2d920d\x2d41141f7a284a-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jan 30 13:43:09.953122 kubelet[1905]: E0130 13:43:09.953032 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:10.418360 kubelet[1905]: I0130 13:43:10.418311 1905 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8de6409-82b8-494d-920d-41141f7a284a" path="/var/lib/kubelet/pods/c8de6409-82b8-494d-920d-41141f7a284a/volumes" Jan 30 13:43:10.953334 kubelet[1905]: E0130 13:43:10.953276 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:11.034771 kubelet[1905]: I0130 13:43:11.034707 1905 topology_manager.go:215] "Topology Admit Handler" podUID="fcd1188a-401d-48c5-8b6b-839c9ca668e0" podNamespace="kube-system" podName="cilium-operator-599987898-5cgkj" Jan 30 13:43:11.034771 kubelet[1905]: E0130 13:43:11.034776 1905 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="apply-sysctl-overwrites" Jan 30 13:43:11.034771 kubelet[1905]: E0130 13:43:11.034789 1905 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="mount-bpf-fs" Jan 30 13:43:11.035004 kubelet[1905]: E0130 13:43:11.034798 1905 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="cilium-agent" Jan 30 13:43:11.035004 kubelet[1905]: E0130 13:43:11.034807 1905 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="mount-cgroup" Jan 30 13:43:11.035004 kubelet[1905]: E0130 13:43:11.034815 1905 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="clean-cilium-state" Jan 30 13:43:11.035004 kubelet[1905]: I0130 13:43:11.034838 1905 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8de6409-82b8-494d-920d-41141f7a284a" containerName="cilium-agent" Jan 30 13:43:11.038515 kubelet[1905]: I0130 13:43:11.038456 1905 topology_manager.go:215] "Topology Admit Handler" podUID="09b36350-2080-4852-9b85-035f3b23a98c" podNamespace="kube-system" podName="cilium-tllsj" Jan 30 13:43:11.139769 kubelet[1905]: I0130 13:43:11.139698 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/09b36350-2080-4852-9b85-035f3b23a98c-cilium-config-path\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139769 kubelet[1905]: I0130 13:43:11.139754 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/09b36350-2080-4852-9b85-035f3b23a98c-cilium-ipsec-secrets\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139769 kubelet[1905]: I0130 13:43:11.139777 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fcd1188a-401d-48c5-8b6b-839c9ca668e0-cilium-config-path\") pod \"cilium-operator-599987898-5cgkj\" (UID: \"fcd1188a-401d-48c5-8b6b-839c9ca668e0\") " pod="kube-system/cilium-operator-599987898-5cgkj" Jan 30 13:43:11.139992 kubelet[1905]: I0130 13:43:11.139795 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-xtables-lock\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139992 kubelet[1905]: I0130 13:43:11.139810 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/09b36350-2080-4852-9b85-035f3b23a98c-clustermesh-secrets\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139992 kubelet[1905]: I0130 13:43:11.139828 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-host-proc-sys-kernel\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139992 kubelet[1905]: I0130 13:43:11.139847 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/09b36350-2080-4852-9b85-035f3b23a98c-hubble-tls\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.139992 kubelet[1905]: I0130 13:43:11.139926 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-hostproc\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140126 kubelet[1905]: I0130 13:43:11.139991 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-cni-path\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140126 kubelet[1905]: I0130 13:43:11.140061 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-etc-cni-netd\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140173 kubelet[1905]: I0130 13:43:11.140139 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-host-proc-sys-net\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140173 kubelet[1905]: I0130 13:43:11.140164 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2q4h\" (UniqueName: \"kubernetes.io/projected/fcd1188a-401d-48c5-8b6b-839c9ca668e0-kube-api-access-h2q4h\") pod \"cilium-operator-599987898-5cgkj\" (UID: \"fcd1188a-401d-48c5-8b6b-839c9ca668e0\") " pod="kube-system/cilium-operator-599987898-5cgkj" Jan 30 13:43:11.140221 kubelet[1905]: I0130 13:43:11.140182 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-cilium-run\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140221 kubelet[1905]: I0130 13:43:11.140204 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-bpf-maps\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140221 kubelet[1905]: I0130 13:43:11.140219 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-cilium-cgroup\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140280 kubelet[1905]: I0130 13:43:11.140235 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09b36350-2080-4852-9b85-035f3b23a98c-lib-modules\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.140280 kubelet[1905]: I0130 13:43:11.140250 1905 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dg64\" (UniqueName: \"kubernetes.io/projected/09b36350-2080-4852-9b85-035f3b23a98c-kube-api-access-8dg64\") pod \"cilium-tllsj\" (UID: \"09b36350-2080-4852-9b85-035f3b23a98c\") " pod="kube-system/cilium-tllsj" Jan 30 13:43:11.339334 kubelet[1905]: E0130 13:43:11.339278 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:11.339979 containerd[1573]: time="2025-01-30T13:43:11.339905465Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-5cgkj,Uid:fcd1188a-401d-48c5-8b6b-839c9ca668e0,Namespace:kube-system,Attempt:0,}" Jan 30 13:43:11.342521 kubelet[1905]: E0130 13:43:11.342474 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:11.343059 containerd[1573]: time="2025-01-30T13:43:11.342988423Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tllsj,Uid:09b36350-2080-4852-9b85-035f3b23a98c,Namespace:kube-system,Attempt:0,}" Jan 30 13:43:11.363864 containerd[1573]: time="2025-01-30T13:43:11.362231008Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:43:11.363864 containerd[1573]: time="2025-01-30T13:43:11.363833902Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:43:11.363864 containerd[1573]: time="2025-01-30T13:43:11.363851615Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:11.364068 containerd[1573]: time="2025-01-30T13:43:11.363967493Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:11.368470 containerd[1573]: time="2025-01-30T13:43:11.368285623Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:43:11.368470 containerd[1573]: time="2025-01-30T13:43:11.368352469Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:43:11.368470 containerd[1573]: time="2025-01-30T13:43:11.368371665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:11.369327 containerd[1573]: time="2025-01-30T13:43:11.368458869Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:43:11.401775 containerd[1573]: time="2025-01-30T13:43:11.401666805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tllsj,Uid:09b36350-2080-4852-9b85-035f3b23a98c,Namespace:kube-system,Attempt:0,} returns sandbox id \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\"" Jan 30 13:43:11.402424 kubelet[1905]: E0130 13:43:11.402402 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:11.404657 kubelet[1905]: E0130 13:43:11.404468 1905 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jan 30 13:43:11.405311 containerd[1573]: time="2025-01-30T13:43:11.405277676Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 30 13:43:11.422849 containerd[1573]: time="2025-01-30T13:43:11.422750033Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-5cgkj,Uid:fcd1188a-401d-48c5-8b6b-839c9ca668e0,Namespace:kube-system,Attempt:0,} returns sandbox id \"091828b6b0a189b7654d026037cac51e3d0ea92bf557915b67465cd212fc3634\"" Jan 30 13:43:11.423580 kubelet[1905]: E0130 13:43:11.423545 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:11.424464 containerd[1573]: time="2025-01-30T13:43:11.424429019Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jan 30 13:43:11.439751 containerd[1573]: time="2025-01-30T13:43:11.439673547Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9b7709d3c755fa99fd5cee767181e6f208e1d877e293231685b108328076aeae\"" Jan 30 13:43:11.440340 containerd[1573]: time="2025-01-30T13:43:11.440291379Z" level=info msg="StartContainer for \"9b7709d3c755fa99fd5cee767181e6f208e1d877e293231685b108328076aeae\"" Jan 30 13:43:11.498824 containerd[1573]: time="2025-01-30T13:43:11.498763315Z" level=info msg="StartContainer for \"9b7709d3c755fa99fd5cee767181e6f208e1d877e293231685b108328076aeae\" returns successfully" Jan 30 13:43:11.528395 kubelet[1905]: E0130 13:43:11.528059 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:11.544262 containerd[1573]: time="2025-01-30T13:43:11.544195509Z" level=info msg="shim disconnected" id=9b7709d3c755fa99fd5cee767181e6f208e1d877e293231685b108328076aeae namespace=k8s.io Jan 30 13:43:11.544262 containerd[1573]: time="2025-01-30T13:43:11.544259380Z" level=warning msg="cleaning up after shim disconnected" id=9b7709d3c755fa99fd5cee767181e6f208e1d877e293231685b108328076aeae namespace=k8s.io Jan 30 13:43:11.544262 containerd[1573]: time="2025-01-30T13:43:11.544270831Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:11.954355 kubelet[1905]: E0130 13:43:11.954298 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:12.532224 kubelet[1905]: E0130 13:43:12.532199 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:12.534234 containerd[1573]: time="2025-01-30T13:43:12.534194100Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 30 13:43:12.547884 containerd[1573]: time="2025-01-30T13:43:12.547823756Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a\"" Jan 30 13:43:12.548456 containerd[1573]: time="2025-01-30T13:43:12.548414547Z" level=info msg="StartContainer for \"d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a\"" Jan 30 13:43:12.605665 containerd[1573]: time="2025-01-30T13:43:12.605618750Z" level=info msg="StartContainer for \"d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a\" returns successfully" Jan 30 13:43:12.635678 containerd[1573]: time="2025-01-30T13:43:12.635609077Z" level=info msg="shim disconnected" id=d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a namespace=k8s.io Jan 30 13:43:12.635678 containerd[1573]: time="2025-01-30T13:43:12.635670833Z" level=warning msg="cleaning up after shim disconnected" id=d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a namespace=k8s.io Jan 30 13:43:12.635678 containerd[1573]: time="2025-01-30T13:43:12.635679519Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:12.955169 kubelet[1905]: E0130 13:43:12.955119 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:13.246478 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d639acc3348e38a4f428f25d87275be3dbc752b7e5494855c46577f6db7f4a2a-rootfs.mount: Deactivated successfully. Jan 30 13:43:13.535226 kubelet[1905]: E0130 13:43:13.535199 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:13.536551 containerd[1573]: time="2025-01-30T13:43:13.536520130Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 30 13:43:13.956388 kubelet[1905]: E0130 13:43:13.956253 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:14.021004 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2765422457.mount: Deactivated successfully. Jan 30 13:43:14.024384 containerd[1573]: time="2025-01-30T13:43:14.024328592Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:43:14.025300 containerd[1573]: time="2025-01-30T13:43:14.025265743Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94\"" Jan 30 13:43:14.025824 containerd[1573]: time="2025-01-30T13:43:14.025789047Z" level=info msg="StartContainer for \"47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94\"" Jan 30 13:43:14.026383 containerd[1573]: time="2025-01-30T13:43:14.026344861Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Jan 30 13:43:14.027804 containerd[1573]: time="2025-01-30T13:43:14.027764621Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:43:14.029356 containerd[1573]: time="2025-01-30T13:43:14.029309013Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.604833345s" Jan 30 13:43:14.029356 containerd[1573]: time="2025-01-30T13:43:14.029348838Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Jan 30 13:43:14.032120 containerd[1573]: time="2025-01-30T13:43:14.031873644Z" level=info msg="CreateContainer within sandbox \"091828b6b0a189b7654d026037cac51e3d0ea92bf557915b67465cd212fc3634\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jan 30 13:43:14.277134 containerd[1573]: time="2025-01-30T13:43:14.276845951Z" level=info msg="StartContainer for \"47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94\" returns successfully" Jan 30 13:43:14.291195 containerd[1573]: time="2025-01-30T13:43:14.291136061Z" level=info msg="CreateContainer within sandbox \"091828b6b0a189b7654d026037cac51e3d0ea92bf557915b67465cd212fc3634\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"37b1aba271d1f639e7f2e97b3524c6c9cf7b051a162a8b634221c0057db460f2\"" Jan 30 13:43:14.291995 containerd[1573]: time="2025-01-30T13:43:14.291949970Z" level=info msg="StartContainer for \"37b1aba271d1f639e7f2e97b3524c6c9cf7b051a162a8b634221c0057db460f2\"" Jan 30 13:43:14.295471 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94-rootfs.mount: Deactivated successfully. Jan 30 13:43:14.298565 containerd[1573]: time="2025-01-30T13:43:14.298358958Z" level=info msg="shim disconnected" id=47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94 namespace=k8s.io Jan 30 13:43:14.298565 containerd[1573]: time="2025-01-30T13:43:14.298417307Z" level=warning msg="cleaning up after shim disconnected" id=47fd0993f732fdaaded8dd0edb9ef22190a3ade652d61c4e348197b3c2f1bd94 namespace=k8s.io Jan 30 13:43:14.298565 containerd[1573]: time="2025-01-30T13:43:14.298428368Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:14.343437 containerd[1573]: time="2025-01-30T13:43:14.343390157Z" level=info msg="StartContainer for \"37b1aba271d1f639e7f2e97b3524c6c9cf7b051a162a8b634221c0057db460f2\" returns successfully" Jan 30 13:43:14.538847 kubelet[1905]: E0130 13:43:14.538817 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:14.541292 kubelet[1905]: E0130 13:43:14.540659 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:14.542642 containerd[1573]: time="2025-01-30T13:43:14.542592585Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 30 13:43:14.660816 kubelet[1905]: I0130 13:43:14.660717 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-5cgkj" podStartSLOduration=1.054745425 podStartE2EDuration="3.660696873s" podCreationTimestamp="2025-01-30 13:43:11 +0000 UTC" firstStartedPulling="2025-01-30 13:43:11.424155445 +0000 UTC m=+55.765186227" lastFinishedPulling="2025-01-30 13:43:14.030106893 +0000 UTC m=+58.371137675" observedRunningTime="2025-01-30 13:43:14.660625408 +0000 UTC m=+59.001656190" watchObservedRunningTime="2025-01-30 13:43:14.660696873 +0000 UTC m=+59.001727655" Jan 30 13:43:14.854500 containerd[1573]: time="2025-01-30T13:43:14.854342916Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"816de0b832e6fffb8e313fdf385e7af1aa3ff7527ee2c7d6462ab24d25cf3aa3\"" Jan 30 13:43:14.855220 containerd[1573]: time="2025-01-30T13:43:14.855181603Z" level=info msg="StartContainer for \"816de0b832e6fffb8e313fdf385e7af1aa3ff7527ee2c7d6462ab24d25cf3aa3\"" Jan 30 13:43:14.912013 containerd[1573]: time="2025-01-30T13:43:14.911960984Z" level=info msg="StartContainer for \"816de0b832e6fffb8e313fdf385e7af1aa3ff7527ee2c7d6462ab24d25cf3aa3\" returns successfully" Jan 30 13:43:14.930414 containerd[1573]: time="2025-01-30T13:43:14.930350990Z" level=info msg="shim disconnected" id=816de0b832e6fffb8e313fdf385e7af1aa3ff7527ee2c7d6462ab24d25cf3aa3 namespace=k8s.io Jan 30 13:43:14.930414 containerd[1573]: time="2025-01-30T13:43:14.930403960Z" level=warning msg="cleaning up after shim disconnected" id=816de0b832e6fffb8e313fdf385e7af1aa3ff7527ee2c7d6462ab24d25cf3aa3 namespace=k8s.io Jan 30 13:43:14.930414 containerd[1573]: time="2025-01-30T13:43:14.930413328Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:43:14.957308 kubelet[1905]: E0130 13:43:14.957250 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:15.544541 kubelet[1905]: E0130 13:43:15.544499 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:15.544710 kubelet[1905]: E0130 13:43:15.544561 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:15.546345 containerd[1573]: time="2025-01-30T13:43:15.546316549Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 30 13:43:15.816297 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2106424442.mount: Deactivated successfully. Jan 30 13:43:15.919477 kubelet[1905]: E0130 13:43:15.919415 1905 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:15.932970 containerd[1573]: time="2025-01-30T13:43:15.932940038Z" level=info msg="StopPodSandbox for \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\"" Jan 30 13:43:15.933075 containerd[1573]: time="2025-01-30T13:43:15.933019667Z" level=info msg="TearDown network for sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" successfully" Jan 30 13:43:15.933075 containerd[1573]: time="2025-01-30T13:43:15.933031930Z" level=info msg="StopPodSandbox for \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" returns successfully" Jan 30 13:43:15.933376 containerd[1573]: time="2025-01-30T13:43:15.933352373Z" level=info msg="RemovePodSandbox for \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\"" Jan 30 13:43:15.933443 containerd[1573]: time="2025-01-30T13:43:15.933378763Z" level=info msg="Forcibly stopping sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\"" Jan 30 13:43:15.933478 containerd[1573]: time="2025-01-30T13:43:15.933443374Z" level=info msg="TearDown network for sandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" successfully" Jan 30 13:43:15.957734 kubelet[1905]: E0130 13:43:15.957682 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:16.038898 containerd[1573]: time="2025-01-30T13:43:16.038825443Z" level=info msg="CreateContainer within sandbox \"eb044d7f0905bddac0ff2a7c74676e35dee8bcde67d9f19d4d9e7fd383118a0c\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b21c2f5247ece4b6db7fdd0aacfd7284d014fcdfcdef480491fd5de6c8fe5ff2\"" Jan 30 13:43:16.091618 containerd[1573]: time="2025-01-30T13:43:16.039436211Z" level=info msg="StartContainer for \"b21c2f5247ece4b6db7fdd0aacfd7284d014fcdfcdef480491fd5de6c8fe5ff2\"" Jan 30 13:43:16.405594 kubelet[1905]: E0130 13:43:16.405477 1905 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jan 30 13:43:16.538714 containerd[1573]: time="2025-01-30T13:43:16.538411692Z" level=info msg="StartContainer for \"b21c2f5247ece4b6db7fdd0aacfd7284d014fcdfcdef480491fd5de6c8fe5ff2\" returns successfully" Jan 30 13:43:16.647249 containerd[1573]: time="2025-01-30T13:43:16.647024434Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:43:16.647249 containerd[1573]: time="2025-01-30T13:43:16.647109323Z" level=info msg="RemovePodSandbox \"8af18c0baad8c1f420729f617a8eb40e44594843c7e9c70d5a00f88ccf5eb49b\" returns successfully" Jan 30 13:43:16.690126 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Jan 30 13:43:16.958558 kubelet[1905]: E0130 13:43:16.958488 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:17.551455 kubelet[1905]: E0130 13:43:17.551417 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:17.745123 kubelet[1905]: I0130 13:43:17.742971 1905 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-tllsj" podStartSLOduration=6.7429486579999995 podStartE2EDuration="6.742948658s" podCreationTimestamp="2025-01-30 13:43:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:43:17.742706473 +0000 UTC m=+62.083737255" watchObservedRunningTime="2025-01-30 13:43:17.742948658 +0000 UTC m=+62.083979440" Jan 30 13:43:17.958977 kubelet[1905]: E0130 13:43:17.958814 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:18.662971 kubelet[1905]: I0130 13:43:18.662930 1905 setters.go:580] "Node became not ready" node="10.0.0.28" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-01-30T13:43:18Z","lastTransitionTime":"2025-01-30T13:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jan 30 13:43:18.959819 kubelet[1905]: E0130 13:43:18.959630 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:19.344075 kubelet[1905]: E0130 13:43:19.344038 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:19.838691 systemd[1]: run-containerd-runc-k8s.io-b21c2f5247ece4b6db7fdd0aacfd7284d014fcdfcdef480491fd5de6c8fe5ff2-runc.4GV7Jz.mount: Deactivated successfully. Jan 30 13:43:19.960045 kubelet[1905]: E0130 13:43:19.959996 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:20.114804 systemd-networkd[1244]: lxc_health: Link UP Jan 30 13:43:20.128317 systemd-networkd[1244]: lxc_health: Gained carrier Jan 30 13:43:20.960928 kubelet[1905]: E0130 13:43:20.960850 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:21.344973 kubelet[1905]: E0130 13:43:21.344932 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:21.437175 systemd-networkd[1244]: lxc_health: Gained IPv6LL Jan 30 13:43:21.559482 kubelet[1905]: E0130 13:43:21.559433 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:21.961478 kubelet[1905]: E0130 13:43:21.961425 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:22.561268 kubelet[1905]: E0130 13:43:22.561204 1905 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 30 13:43:22.962066 kubelet[1905]: E0130 13:43:22.961900 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:23.962833 kubelet[1905]: E0130 13:43:23.962760 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:24.963238 kubelet[1905]: E0130 13:43:24.963159 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:25.963532 kubelet[1905]: E0130 13:43:25.963468 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:26.963654 kubelet[1905]: E0130 13:43:26.963591 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Jan 30 13:43:27.964300 kubelet[1905]: E0130 13:43:27.964225 1905 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"