Oct 8 20:03:18.902493 kernel: Linux version 6.6.54-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Oct 8 18:24:27 -00 2024 Oct 8 20:03:18.902520 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 8 20:03:18.902535 kernel: BIOS-provided physical RAM map: Oct 8 20:03:18.902544 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Oct 8 20:03:18.902553 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Oct 8 20:03:18.902561 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Oct 8 20:03:18.902572 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Oct 8 20:03:18.902581 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Oct 8 20:03:18.902590 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Oct 8 20:03:18.902602 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Oct 8 20:03:18.902611 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 8 20:03:18.902620 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Oct 8 20:03:18.902629 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Oct 8 20:03:18.902638 kernel: NX (Execute Disable) protection: active Oct 8 20:03:18.902649 kernel: APIC: Static calls initialized Oct 8 20:03:18.902662 kernel: SMBIOS 2.8 present. Oct 8 20:03:18.902672 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Oct 8 20:03:18.902682 kernel: Hypervisor detected: KVM Oct 8 20:03:18.902693 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 8 20:03:18.902704 kernel: kvm-clock: using sched offset of 2223767137 cycles Oct 8 20:03:18.902731 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 8 20:03:18.902752 kernel: tsc: Detected 2794.748 MHz processor Oct 8 20:03:18.902762 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 8 20:03:18.902772 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 8 20:03:18.902799 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Oct 8 20:03:18.902822 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Oct 8 20:03:18.902849 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 8 20:03:18.902882 kernel: Using GB pages for direct mapping Oct 8 20:03:18.902901 kernel: ACPI: Early table checksum verification disabled Oct 8 20:03:18.902918 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Oct 8 20:03:18.902938 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.902948 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.902958 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.902973 kernel: ACPI: FACS 0x000000009CFE0000 000040 Oct 8 20:03:18.902983 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.902993 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.903003 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.903013 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 8 20:03:18.903023 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db] Oct 8 20:03:18.903033 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7] Oct 8 20:03:18.903047 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Oct 8 20:03:18.903060 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b] Oct 8 20:03:18.903071 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3] Oct 8 20:03:18.903081 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df] Oct 8 20:03:18.903092 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407] Oct 8 20:03:18.903102 kernel: No NUMA configuration found Oct 8 20:03:18.903113 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Oct 8 20:03:18.903123 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Oct 8 20:03:18.903137 kernel: Zone ranges: Oct 8 20:03:18.903147 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 8 20:03:18.903157 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Oct 8 20:03:18.903168 kernel: Normal empty Oct 8 20:03:18.903178 kernel: Movable zone start for each node Oct 8 20:03:18.903188 kernel: Early memory node ranges Oct 8 20:03:18.903199 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Oct 8 20:03:18.903209 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Oct 8 20:03:18.903219 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Oct 8 20:03:18.903233 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 8 20:03:18.903243 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Oct 8 20:03:18.903254 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Oct 8 20:03:18.903264 kernel: ACPI: PM-Timer IO Port: 0x608 Oct 8 20:03:18.903274 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 8 20:03:18.903285 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Oct 8 20:03:18.903294 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Oct 8 20:03:18.903304 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 8 20:03:18.903315 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 8 20:03:18.903328 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 8 20:03:18.903338 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 8 20:03:18.903349 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 8 20:03:18.903359 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Oct 8 20:03:18.903369 kernel: TSC deadline timer available Oct 8 20:03:18.903380 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Oct 8 20:03:18.903390 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 8 20:03:18.903400 kernel: kvm-guest: KVM setup pv remote TLB flush Oct 8 20:03:18.903411 kernel: kvm-guest: setup PV sched yield Oct 8 20:03:18.903424 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Oct 8 20:03:18.903434 kernel: Booting paravirtualized kernel on KVM Oct 8 20:03:18.903445 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 8 20:03:18.903456 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Oct 8 20:03:18.903466 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u524288 Oct 8 20:03:18.903476 kernel: pcpu-alloc: s196904 r8192 d32472 u524288 alloc=1*2097152 Oct 8 20:03:18.903486 kernel: pcpu-alloc: [0] 0 1 2 3 Oct 8 20:03:18.903497 kernel: kvm-guest: PV spinlocks enabled Oct 8 20:03:18.903507 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Oct 8 20:03:18.903519 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 8 20:03:18.903533 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Oct 8 20:03:18.903543 kernel: random: crng init done Oct 8 20:03:18.903552 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Oct 8 20:03:18.903563 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 8 20:03:18.903574 kernel: Fallback order for Node 0: 0 Oct 8 20:03:18.903584 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Oct 8 20:03:18.903595 kernel: Policy zone: DMA32 Oct 8 20:03:18.903605 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 8 20:03:18.903619 kernel: Memory: 2434584K/2571752K available (12288K kernel code, 2305K rwdata, 22716K rodata, 42828K init, 2360K bss, 136908K reserved, 0K cma-reserved) Oct 8 20:03:18.903630 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Oct 8 20:03:18.903641 kernel: ftrace: allocating 37784 entries in 148 pages Oct 8 20:03:18.903651 kernel: ftrace: allocated 148 pages with 3 groups Oct 8 20:03:18.903662 kernel: Dynamic Preempt: voluntary Oct 8 20:03:18.903672 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 8 20:03:18.903687 kernel: rcu: RCU event tracing is enabled. Oct 8 20:03:18.903698 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Oct 8 20:03:18.903722 kernel: Trampoline variant of Tasks RCU enabled. Oct 8 20:03:18.903737 kernel: Rude variant of Tasks RCU enabled. Oct 8 20:03:18.903747 kernel: Tracing variant of Tasks RCU enabled. Oct 8 20:03:18.903758 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 8 20:03:18.903768 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Oct 8 20:03:18.903779 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Oct 8 20:03:18.903790 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 8 20:03:18.903800 kernel: Console: colour VGA+ 80x25 Oct 8 20:03:18.903810 kernel: printk: console [ttyS0] enabled Oct 8 20:03:18.903820 kernel: ACPI: Core revision 20230628 Oct 8 20:03:18.903833 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Oct 8 20:03:18.903843 kernel: APIC: Switch to symmetric I/O mode setup Oct 8 20:03:18.903861 kernel: x2apic enabled Oct 8 20:03:18.903871 kernel: APIC: Switched APIC routing to: physical x2apic Oct 8 20:03:18.903881 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Oct 8 20:03:18.903891 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Oct 8 20:03:18.903902 kernel: kvm-guest: setup PV IPIs Oct 8 20:03:18.903926 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Oct 8 20:03:18.903936 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Oct 8 20:03:18.903947 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Oct 8 20:03:18.903958 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Oct 8 20:03:18.903969 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Oct 8 20:03:18.903982 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Oct 8 20:03:18.903993 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 8 20:03:18.904004 kernel: Spectre V2 : Mitigation: Retpolines Oct 8 20:03:18.904015 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Oct 8 20:03:18.904029 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Oct 8 20:03:18.904040 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Oct 8 20:03:18.904051 kernel: RETBleed: Mitigation: untrained return thunk Oct 8 20:03:18.904062 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Oct 8 20:03:18.904073 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Oct 8 20:03:18.904083 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Oct 8 20:03:18.904094 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Oct 8 20:03:18.904105 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Oct 8 20:03:18.904125 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 8 20:03:18.904153 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 8 20:03:18.904175 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 8 20:03:18.904193 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 8 20:03:18.904215 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Oct 8 20:03:18.904234 kernel: Freeing SMP alternatives memory: 32K Oct 8 20:03:18.904255 kernel: pid_max: default: 32768 minimum: 301 Oct 8 20:03:18.904277 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Oct 8 20:03:18.904295 kernel: landlock: Up and running. Oct 8 20:03:18.904313 kernel: SELinux: Initializing. Oct 8 20:03:18.904332 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 8 20:03:18.904343 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Oct 8 20:03:18.904354 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Oct 8 20:03:18.904365 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Oct 8 20:03:18.904376 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Oct 8 20:03:18.904386 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Oct 8 20:03:18.904397 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Oct 8 20:03:18.904408 kernel: ... version: 0 Oct 8 20:03:18.904419 kernel: ... bit width: 48 Oct 8 20:03:18.904433 kernel: ... generic registers: 6 Oct 8 20:03:18.904444 kernel: ... value mask: 0000ffffffffffff Oct 8 20:03:18.904454 kernel: ... max period: 00007fffffffffff Oct 8 20:03:18.904465 kernel: ... fixed-purpose events: 0 Oct 8 20:03:18.904476 kernel: ... event mask: 000000000000003f Oct 8 20:03:18.904487 kernel: signal: max sigframe size: 1776 Oct 8 20:03:18.904498 kernel: rcu: Hierarchical SRCU implementation. Oct 8 20:03:18.904509 kernel: rcu: Max phase no-delay instances is 400. Oct 8 20:03:18.904520 kernel: smp: Bringing up secondary CPUs ... Oct 8 20:03:18.904533 kernel: smpboot: x86: Booting SMP configuration: Oct 8 20:03:18.904544 kernel: .... node #0, CPUs: #1 #2 #3 Oct 8 20:03:18.904555 kernel: smp: Brought up 1 node, 4 CPUs Oct 8 20:03:18.904566 kernel: smpboot: Max logical packages: 1 Oct 8 20:03:18.904576 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Oct 8 20:03:18.904587 kernel: devtmpfs: initialized Oct 8 20:03:18.904598 kernel: x86/mm: Memory block size: 128MB Oct 8 20:03:18.904609 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 8 20:03:18.904630 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Oct 8 20:03:18.904654 kernel: pinctrl core: initialized pinctrl subsystem Oct 8 20:03:18.904666 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 8 20:03:18.904677 kernel: audit: initializing netlink subsys (disabled) Oct 8 20:03:18.904687 kernel: audit: type=2000 audit(1728417798.992:1): state=initialized audit_enabled=0 res=1 Oct 8 20:03:18.904698 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 8 20:03:18.904721 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 8 20:03:18.904748 kernel: cpuidle: using governor menu Oct 8 20:03:18.904759 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 8 20:03:18.904770 kernel: dca service started, version 1.12.1 Oct 8 20:03:18.904785 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Oct 8 20:03:18.904796 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Oct 8 20:03:18.904807 kernel: PCI: Using configuration type 1 for base access Oct 8 20:03:18.904818 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 8 20:03:18.904829 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Oct 8 20:03:18.904839 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Oct 8 20:03:18.904850 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 8 20:03:18.904870 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 8 20:03:18.904881 kernel: ACPI: Added _OSI(Module Device) Oct 8 20:03:18.904894 kernel: ACPI: Added _OSI(Processor Device) Oct 8 20:03:18.904905 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Oct 8 20:03:18.904916 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 8 20:03:18.904927 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 8 20:03:18.904938 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Oct 8 20:03:18.904949 kernel: ACPI: Interpreter enabled Oct 8 20:03:18.904959 kernel: ACPI: PM: (supports S0 S3 S5) Oct 8 20:03:18.904970 kernel: ACPI: Using IOAPIC for interrupt routing Oct 8 20:03:18.904981 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 8 20:03:18.904995 kernel: PCI: Using E820 reservations for host bridge windows Oct 8 20:03:18.905006 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Oct 8 20:03:18.905017 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 8 20:03:18.905230 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Oct 8 20:03:18.905391 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Oct 8 20:03:18.905546 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Oct 8 20:03:18.905560 kernel: PCI host bridge to bus 0000:00 Oct 8 20:03:18.905748 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 8 20:03:18.905910 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 8 20:03:18.906050 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 8 20:03:18.906233 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Oct 8 20:03:18.906431 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Oct 8 20:03:18.906574 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Oct 8 20:03:18.906727 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 8 20:03:18.906938 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Oct 8 20:03:18.907104 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Oct 8 20:03:18.907259 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Oct 8 20:03:18.907446 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Oct 8 20:03:18.907648 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Oct 8 20:03:18.907822 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 8 20:03:18.908010 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Oct 8 20:03:18.908172 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Oct 8 20:03:18.908330 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Oct 8 20:03:18.908486 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Oct 8 20:03:18.908649 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Oct 8 20:03:18.908845 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Oct 8 20:03:18.909015 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Oct 8 20:03:18.909174 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Oct 8 20:03:18.909340 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Oct 8 20:03:18.909497 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Oct 8 20:03:18.909654 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Oct 8 20:03:18.909826 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Oct 8 20:03:18.909993 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Oct 8 20:03:18.910159 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Oct 8 20:03:18.910320 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Oct 8 20:03:18.910486 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Oct 8 20:03:18.910638 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Oct 8 20:03:18.910826 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Oct 8 20:03:18.911004 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Oct 8 20:03:18.911160 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Oct 8 20:03:18.911174 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 8 20:03:18.911190 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 8 20:03:18.911201 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 8 20:03:18.911212 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 8 20:03:18.911223 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Oct 8 20:03:18.911234 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Oct 8 20:03:18.911245 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Oct 8 20:03:18.911256 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Oct 8 20:03:18.911267 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Oct 8 20:03:18.911277 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Oct 8 20:03:18.911291 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Oct 8 20:03:18.911302 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Oct 8 20:03:18.911313 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Oct 8 20:03:18.911324 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Oct 8 20:03:18.911335 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Oct 8 20:03:18.911346 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Oct 8 20:03:18.911356 kernel: iommu: Default domain type: Translated Oct 8 20:03:18.911367 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 8 20:03:18.911378 kernel: PCI: Using ACPI for IRQ routing Oct 8 20:03:18.911392 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 8 20:03:18.911403 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Oct 8 20:03:18.911414 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Oct 8 20:03:18.911572 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Oct 8 20:03:18.911772 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Oct 8 20:03:18.911938 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 8 20:03:18.911954 kernel: vgaarb: loaded Oct 8 20:03:18.911965 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Oct 8 20:03:18.911981 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Oct 8 20:03:18.911992 kernel: clocksource: Switched to clocksource kvm-clock Oct 8 20:03:18.912003 kernel: VFS: Disk quotas dquot_6.6.0 Oct 8 20:03:18.912015 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 8 20:03:18.912026 kernel: pnp: PnP ACPI init Oct 8 20:03:18.912184 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Oct 8 20:03:18.912200 kernel: pnp: PnP ACPI: found 6 devices Oct 8 20:03:18.912211 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 8 20:03:18.912226 kernel: NET: Registered PF_INET protocol family Oct 8 20:03:18.912237 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Oct 8 20:03:18.912248 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Oct 8 20:03:18.912259 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 8 20:03:18.912271 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 8 20:03:18.912282 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Oct 8 20:03:18.912293 kernel: TCP: Hash tables configured (established 32768 bind 32768) Oct 8 20:03:18.912303 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 8 20:03:18.912315 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Oct 8 20:03:18.912329 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 8 20:03:18.912340 kernel: NET: Registered PF_XDP protocol family Oct 8 20:03:18.912481 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 8 20:03:18.912624 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 8 20:03:18.912794 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 8 20:03:18.912945 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Oct 8 20:03:18.913082 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Oct 8 20:03:18.913220 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Oct 8 20:03:18.913239 kernel: PCI: CLS 0 bytes, default 64 Oct 8 20:03:18.913250 kernel: Initialise system trusted keyrings Oct 8 20:03:18.913261 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Oct 8 20:03:18.913272 kernel: Key type asymmetric registered Oct 8 20:03:18.913283 kernel: Asymmetric key parser 'x509' registered Oct 8 20:03:18.913294 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Oct 8 20:03:18.913304 kernel: io scheduler mq-deadline registered Oct 8 20:03:18.913315 kernel: io scheduler kyber registered Oct 8 20:03:18.913326 kernel: io scheduler bfq registered Oct 8 20:03:18.913340 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 8 20:03:18.913352 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Oct 8 20:03:18.913363 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Oct 8 20:03:18.913374 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Oct 8 20:03:18.913385 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 8 20:03:18.913396 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 8 20:03:18.913407 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 8 20:03:18.913418 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 8 20:03:18.913429 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 8 20:03:18.913592 kernel: rtc_cmos 00:04: RTC can wake from S4 Oct 8 20:03:18.913863 kernel: rtc_cmos 00:04: registered as rtc0 Oct 8 20:03:18.913881 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 8 20:03:18.914025 kernel: rtc_cmos 00:04: setting system clock to 2024-10-08T20:03:18 UTC (1728417798) Oct 8 20:03:18.914166 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Oct 8 20:03:18.914181 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Oct 8 20:03:18.914192 kernel: NET: Registered PF_INET6 protocol family Oct 8 20:03:18.914203 kernel: Segment Routing with IPv6 Oct 8 20:03:18.914219 kernel: In-situ OAM (IOAM) with IPv6 Oct 8 20:03:18.914230 kernel: NET: Registered PF_PACKET protocol family Oct 8 20:03:18.914241 kernel: Key type dns_resolver registered Oct 8 20:03:18.914252 kernel: IPI shorthand broadcast: enabled Oct 8 20:03:18.914263 kernel: sched_clock: Marking stable (564003522, 117076403)->(738897647, -57817722) Oct 8 20:03:18.914274 kernel: registered taskstats version 1 Oct 8 20:03:18.914285 kernel: Loading compiled-in X.509 certificates Oct 8 20:03:18.914296 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.54-flatcar: 14ce23fc5070d0471461f1dd6e298a5588e7ba8f' Oct 8 20:03:18.914307 kernel: Key type .fscrypt registered Oct 8 20:03:18.914321 kernel: Key type fscrypt-provisioning registered Oct 8 20:03:18.914332 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 8 20:03:18.914343 kernel: ima: Allocated hash algorithm: sha1 Oct 8 20:03:18.914354 kernel: ima: No architecture policies found Oct 8 20:03:18.914365 kernel: clk: Disabling unused clocks Oct 8 20:03:18.914376 kernel: Freeing unused kernel image (initmem) memory: 42828K Oct 8 20:03:18.914387 kernel: Write protecting the kernel read-only data: 36864k Oct 8 20:03:18.914398 kernel: Freeing unused kernel image (rodata/data gap) memory: 1860K Oct 8 20:03:18.914409 kernel: Run /init as init process Oct 8 20:03:18.914423 kernel: with arguments: Oct 8 20:03:18.914433 kernel: /init Oct 8 20:03:18.914444 kernel: with environment: Oct 8 20:03:18.914455 kernel: HOME=/ Oct 8 20:03:18.914466 kernel: TERM=linux Oct 8 20:03:18.914476 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Oct 8 20:03:18.914489 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 8 20:03:18.914502 systemd[1]: Detected virtualization kvm. Oct 8 20:03:18.914517 systemd[1]: Detected architecture x86-64. Oct 8 20:03:18.914529 systemd[1]: Running in initrd. Oct 8 20:03:18.914540 systemd[1]: No hostname configured, using default hostname. Oct 8 20:03:18.914551 systemd[1]: Hostname set to . Oct 8 20:03:18.914564 systemd[1]: Initializing machine ID from VM UUID. Oct 8 20:03:18.914575 systemd[1]: Queued start job for default target initrd.target. Oct 8 20:03:18.914587 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 8 20:03:18.914599 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 8 20:03:18.914615 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 8 20:03:18.914642 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 8 20:03:18.914657 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 8 20:03:18.914669 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 8 20:03:18.914684 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Oct 8 20:03:18.914699 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Oct 8 20:03:18.914725 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 8 20:03:18.914752 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 8 20:03:18.914764 systemd[1]: Reached target paths.target - Path Units. Oct 8 20:03:18.914776 systemd[1]: Reached target slices.target - Slice Units. Oct 8 20:03:18.914788 systemd[1]: Reached target swap.target - Swaps. Oct 8 20:03:18.914800 systemd[1]: Reached target timers.target - Timer Units. Oct 8 20:03:18.914813 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 8 20:03:18.914828 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 8 20:03:18.914840 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 8 20:03:18.914864 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Oct 8 20:03:18.914877 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 8 20:03:18.914889 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 8 20:03:18.914901 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 8 20:03:18.914913 systemd[1]: Reached target sockets.target - Socket Units. Oct 8 20:03:18.914925 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 8 20:03:18.914938 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 8 20:03:18.914953 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 8 20:03:18.914965 systemd[1]: Starting systemd-fsck-usr.service... Oct 8 20:03:18.914979 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 8 20:03:18.914992 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 8 20:03:18.915004 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 8 20:03:18.915016 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 8 20:03:18.915028 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 8 20:03:18.915040 systemd[1]: Finished systemd-fsck-usr.service. Oct 8 20:03:18.915056 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 8 20:03:18.915091 systemd-journald[193]: Collecting audit messages is disabled. Oct 8 20:03:18.915121 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 8 20:03:18.915133 systemd-journald[193]: Journal started Oct 8 20:03:18.915161 systemd-journald[193]: Runtime Journal (/run/log/journal/3ef00aaf0b6d472397d6ddd41391d4a0) is 6.0M, max 48.4M, 42.3M free. Oct 8 20:03:18.904697 systemd-modules-load[194]: Inserted module 'overlay' Oct 8 20:03:18.939518 systemd[1]: Started systemd-journald.service - Journal Service. Oct 8 20:03:18.940275 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 8 20:03:18.944633 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 8 20:03:18.946733 kernel: Bridge firewalling registered Oct 8 20:03:18.946829 systemd-modules-load[194]: Inserted module 'br_netfilter' Oct 8 20:03:18.959981 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 8 20:03:18.963489 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 8 20:03:18.966780 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 8 20:03:18.968374 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 8 20:03:18.975910 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 8 20:03:18.979483 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 8 20:03:18.981364 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 8 20:03:18.985454 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 8 20:03:18.987827 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 8 20:03:18.997534 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 8 20:03:19.002988 dracut-cmdline[225]: dracut-dracut-053 Oct 8 20:03:19.005986 dracut-cmdline[225]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 8 20:03:19.005896 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 8 20:03:19.042620 systemd-resolved[233]: Positive Trust Anchors: Oct 8 20:03:19.042639 systemd-resolved[233]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 8 20:03:19.042684 systemd-resolved[233]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 8 20:03:19.045826 systemd-resolved[233]: Defaulting to hostname 'linux'. Oct 8 20:03:19.047234 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 8 20:03:19.053670 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 8 20:03:19.082738 kernel: SCSI subsystem initialized Oct 8 20:03:19.091728 kernel: Loading iSCSI transport class v2.0-870. Oct 8 20:03:19.102740 kernel: iscsi: registered transport (tcp) Oct 8 20:03:19.122733 kernel: iscsi: registered transport (qla4xxx) Oct 8 20:03:19.122754 kernel: QLogic iSCSI HBA Driver Oct 8 20:03:19.167652 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 8 20:03:19.180865 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 8 20:03:19.205543 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 8 20:03:19.205639 kernel: device-mapper: uevent: version 1.0.3 Oct 8 20:03:19.205654 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Oct 8 20:03:19.246749 kernel: raid6: avx2x4 gen() 23270 MB/s Oct 8 20:03:19.263735 kernel: raid6: avx2x2 gen() 26724 MB/s Oct 8 20:03:19.280862 kernel: raid6: avx2x1 gen() 23891 MB/s Oct 8 20:03:19.280893 kernel: raid6: using algorithm avx2x2 gen() 26724 MB/s Oct 8 20:03:19.298862 kernel: raid6: .... xor() 19697 MB/s, rmw enabled Oct 8 20:03:19.298939 kernel: raid6: using avx2x2 recovery algorithm Oct 8 20:03:19.319751 kernel: xor: automatically using best checksumming function avx Oct 8 20:03:19.472763 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 8 20:03:19.486090 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 8 20:03:19.502942 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 8 20:03:19.515151 systemd-udevd[413]: Using default interface naming scheme 'v255'. Oct 8 20:03:19.519927 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 8 20:03:19.532925 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 8 20:03:19.548548 dracut-pre-trigger[421]: rd.md=0: removing MD RAID activation Oct 8 20:03:19.581537 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 8 20:03:19.593990 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 8 20:03:19.656607 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 8 20:03:19.665887 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 8 20:03:19.679598 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 8 20:03:19.682927 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 8 20:03:19.686364 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 8 20:03:19.689188 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 8 20:03:19.692731 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Oct 8 20:03:19.696864 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 8 20:03:19.705726 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Oct 8 20:03:19.708964 kernel: cryptd: max_cpu_qlen set to 1000 Oct 8 20:03:19.709605 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 8 20:03:19.716166 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 8 20:03:19.716271 kernel: GPT:9289727 != 19775487 Oct 8 20:03:19.716296 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 8 20:03:19.716324 kernel: GPT:9289727 != 19775487 Oct 8 20:03:19.716348 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 8 20:03:19.716369 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 8 20:03:19.731743 kernel: libata version 3.00 loaded. Oct 8 20:03:19.734102 kernel: AVX2 version of gcm_enc/dec engaged. Oct 8 20:03:19.734157 kernel: AES CTR mode by8 optimization enabled Oct 8 20:03:19.736882 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 8 20:03:19.737171 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 8 20:03:19.744428 kernel: ahci 0000:00:1f.2: version 3.0 Oct 8 20:03:19.744724 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Oct 8 20:03:19.740213 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 8 20:03:19.751344 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Oct 8 20:03:19.752223 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Oct 8 20:03:19.754179 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (462) Oct 8 20:03:19.754197 kernel: BTRFS: device fsid a8680da2-059a-4648-a8e8-f62925ab33ec devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (460) Oct 8 20:03:19.742434 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 8 20:03:19.742619 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 8 20:03:19.744578 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 8 20:03:19.761169 kernel: scsi host0: ahci Oct 8 20:03:19.762468 kernel: scsi host1: ahci Oct 8 20:03:19.757381 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 8 20:03:19.766729 kernel: scsi host2: ahci Oct 8 20:03:19.768867 kernel: scsi host3: ahci Oct 8 20:03:19.769768 kernel: scsi host4: ahci Oct 8 20:03:19.772530 kernel: scsi host5: ahci Oct 8 20:03:19.772818 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Oct 8 20:03:19.772838 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Oct 8 20:03:19.774286 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Oct 8 20:03:19.774315 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Oct 8 20:03:19.774326 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Oct 8 20:03:19.774343 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Oct 8 20:03:19.781658 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Oct 8 20:03:19.815646 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Oct 8 20:03:19.817171 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 8 20:03:19.825538 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 8 20:03:19.830815 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Oct 8 20:03:19.832388 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Oct 8 20:03:19.846870 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 8 20:03:19.849252 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 8 20:03:19.858142 disk-uuid[554]: Primary Header is updated. Oct 8 20:03:19.858142 disk-uuid[554]: Secondary Entries is updated. Oct 8 20:03:19.858142 disk-uuid[554]: Secondary Header is updated. Oct 8 20:03:19.862322 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 8 20:03:19.865747 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 8 20:03:19.879258 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 8 20:03:20.078741 kernel: ata1: SATA link down (SStatus 0 SControl 300) Oct 8 20:03:20.078810 kernel: ata2: SATA link down (SStatus 0 SControl 300) Oct 8 20:03:20.079737 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Oct 8 20:03:20.081490 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Oct 8 20:03:20.081573 kernel: ata3.00: applying bridge limits Oct 8 20:03:20.082339 kernel: ata3.00: configured for UDMA/100 Oct 8 20:03:20.082746 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Oct 8 20:03:20.087750 kernel: ata4: SATA link down (SStatus 0 SControl 300) Oct 8 20:03:20.087779 kernel: ata5: SATA link down (SStatus 0 SControl 300) Oct 8 20:03:20.088754 kernel: ata6: SATA link down (SStatus 0 SControl 300) Oct 8 20:03:20.139764 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Oct 8 20:03:20.140174 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Oct 8 20:03:20.153741 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Oct 8 20:03:20.866770 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 8 20:03:20.866845 disk-uuid[556]: The operation has completed successfully. Oct 8 20:03:20.898633 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 8 20:03:20.898776 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 8 20:03:20.923065 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Oct 8 20:03:20.927401 sh[592]: Success Oct 8 20:03:20.941735 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Oct 8 20:03:20.973890 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Oct 8 20:03:20.989232 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Oct 8 20:03:20.992131 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Oct 8 20:03:21.005923 kernel: BTRFS info (device dm-0): first mount of filesystem a8680da2-059a-4648-a8e8-f62925ab33ec Oct 8 20:03:21.005954 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 8 20:03:21.005965 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Oct 8 20:03:21.007116 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 8 20:03:21.007997 kernel: BTRFS info (device dm-0): using free space tree Oct 8 20:03:21.013759 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Oct 8 20:03:21.015664 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 8 20:03:21.026865 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 8 20:03:21.029578 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 8 20:03:21.037920 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 8 20:03:21.037959 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 8 20:03:21.037974 kernel: BTRFS info (device vda6): using free space tree Oct 8 20:03:21.040753 kernel: BTRFS info (device vda6): auto enabling async discard Oct 8 20:03:21.050158 systemd[1]: mnt-oem.mount: Deactivated successfully. Oct 8 20:03:21.052106 kernel: BTRFS info (device vda6): last unmount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 8 20:03:21.061915 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 8 20:03:21.072048 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 8 20:03:21.131729 ignition[682]: Ignition 2.19.0 Oct 8 20:03:21.131741 ignition[682]: Stage: fetch-offline Oct 8 20:03:21.131783 ignition[682]: no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:21.131793 ignition[682]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:21.131892 ignition[682]: parsed url from cmdline: "" Oct 8 20:03:21.131896 ignition[682]: no config URL provided Oct 8 20:03:21.131901 ignition[682]: reading system config file "/usr/lib/ignition/user.ign" Oct 8 20:03:21.131912 ignition[682]: no config at "/usr/lib/ignition/user.ign" Oct 8 20:03:21.131936 ignition[682]: op(1): [started] loading QEMU firmware config module Oct 8 20:03:21.131942 ignition[682]: op(1): executing: "modprobe" "qemu_fw_cfg" Oct 8 20:03:21.142304 ignition[682]: op(1): [finished] loading QEMU firmware config module Oct 8 20:03:21.157054 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 8 20:03:21.177926 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 8 20:03:21.185031 ignition[682]: parsing config with SHA512: 4a9b08c075b0684025de4b404ecb1a8663f37b5969a8bdd4ef437bab48222311478c79253f2f4bb65d9f81e056cb25033b1543453c0a4c568e754036f40b38fc Oct 8 20:03:21.188947 unknown[682]: fetched base config from "system" Oct 8 20:03:21.189544 ignition[682]: fetch-offline: fetch-offline passed Oct 8 20:03:21.188962 unknown[682]: fetched user config from "qemu" Oct 8 20:03:21.189619 ignition[682]: Ignition finished successfully Oct 8 20:03:21.195224 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 8 20:03:21.209507 systemd-networkd[782]: lo: Link UP Oct 8 20:03:21.209522 systemd-networkd[782]: lo: Gained carrier Oct 8 20:03:21.211679 systemd-networkd[782]: Enumeration completed Oct 8 20:03:21.211905 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 8 20:03:21.212260 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 8 20:03:21.212266 systemd-networkd[782]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 8 20:03:21.213300 systemd-networkd[782]: eth0: Link UP Oct 8 20:03:21.213306 systemd-networkd[782]: eth0: Gained carrier Oct 8 20:03:21.213314 systemd-networkd[782]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 8 20:03:21.220369 systemd[1]: Reached target network.target - Network. Oct 8 20:03:21.224767 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Oct 8 20:03:21.226875 systemd-networkd[782]: eth0: DHCPv4 address 10.0.0.126/16, gateway 10.0.0.1 acquired from 10.0.0.1 Oct 8 20:03:21.239887 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 8 20:03:21.253672 ignition[786]: Ignition 2.19.0 Oct 8 20:03:21.253684 ignition[786]: Stage: kargs Oct 8 20:03:21.253870 ignition[786]: no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:21.253881 ignition[786]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:21.254861 ignition[786]: kargs: kargs passed Oct 8 20:03:21.254917 ignition[786]: Ignition finished successfully Oct 8 20:03:21.262397 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 8 20:03:21.275890 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 8 20:03:21.289134 ignition[794]: Ignition 2.19.0 Oct 8 20:03:21.289146 ignition[794]: Stage: disks Oct 8 20:03:21.289337 ignition[794]: no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:21.289349 ignition[794]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:21.290177 ignition[794]: disks: disks passed Oct 8 20:03:21.292868 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 8 20:03:21.290222 ignition[794]: Ignition finished successfully Oct 8 20:03:21.294805 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 8 20:03:21.296698 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 8 20:03:21.296955 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 8 20:03:21.297342 systemd[1]: Reached target sysinit.target - System Initialization. Oct 8 20:03:21.297697 systemd[1]: Reached target basic.target - Basic System. Oct 8 20:03:21.318921 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 8 20:03:21.331479 systemd-fsck[804]: ROOT: clean, 14/553520 files, 52654/553472 blocks Oct 8 20:03:21.338465 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 8 20:03:21.349975 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 8 20:03:21.441734 kernel: EXT4-fs (vda9): mounted filesystem 1df90f14-3ad0-4280-9b7d-a34f65d70e4d r/w with ordered data mode. Quota mode: none. Oct 8 20:03:21.442307 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 8 20:03:21.442986 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 8 20:03:21.450948 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 8 20:03:21.453297 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 8 20:03:21.453642 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Oct 8 20:03:21.453686 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 8 20:03:21.463193 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (812) Oct 8 20:03:21.453727 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 8 20:03:21.467375 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 8 20:03:21.467402 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 8 20:03:21.467414 kernel: BTRFS info (device vda6): using free space tree Oct 8 20:03:21.469740 kernel: BTRFS info (device vda6): auto enabling async discard Oct 8 20:03:21.471651 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 8 20:03:21.491109 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 8 20:03:21.493558 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 8 20:03:21.539604 initrd-setup-root[836]: cut: /sysroot/etc/passwd: No such file or directory Oct 8 20:03:21.545393 initrd-setup-root[843]: cut: /sysroot/etc/group: No such file or directory Oct 8 20:03:21.551134 initrd-setup-root[850]: cut: /sysroot/etc/shadow: No such file or directory Oct 8 20:03:21.556617 initrd-setup-root[857]: cut: /sysroot/etc/gshadow: No such file or directory Oct 8 20:03:21.645425 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 8 20:03:21.657851 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 8 20:03:21.659888 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 8 20:03:21.666773 kernel: BTRFS info (device vda6): last unmount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 8 20:03:21.686598 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 8 20:03:21.689026 ignition[924]: INFO : Ignition 2.19.0 Oct 8 20:03:21.689026 ignition[924]: INFO : Stage: mount Oct 8 20:03:21.690766 ignition[924]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:21.690766 ignition[924]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:21.693883 ignition[924]: INFO : mount: mount passed Oct 8 20:03:21.694662 ignition[924]: INFO : Ignition finished successfully Oct 8 20:03:21.697461 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 8 20:03:21.709906 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 8 20:03:22.005238 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 8 20:03:22.014041 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 8 20:03:22.021745 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (938) Oct 8 20:03:22.021806 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 8 20:03:22.025948 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 8 20:03:22.025972 kernel: BTRFS info (device vda6): using free space tree Oct 8 20:03:22.031305 kernel: BTRFS info (device vda6): auto enabling async discard Oct 8 20:03:22.032156 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 8 20:03:22.052912 ignition[955]: INFO : Ignition 2.19.0 Oct 8 20:03:22.052912 ignition[955]: INFO : Stage: files Oct 8 20:03:22.055020 ignition[955]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:22.055020 ignition[955]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:22.058463 ignition[955]: DEBUG : files: compiled without relabeling support, skipping Oct 8 20:03:22.060117 ignition[955]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 8 20:03:22.060117 ignition[955]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 8 20:03:22.067067 ignition[955]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 8 20:03:22.068897 ignition[955]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 8 20:03:22.071019 unknown[955]: wrote ssh authorized keys file for user: core Oct 8 20:03:22.072476 ignition[955]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 8 20:03:22.074088 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Oct 8 20:03:22.074088 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Oct 8 20:03:22.126222 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Oct 8 20:03:22.261746 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Oct 8 20:03:22.261746 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Oct 8 20:03:22.261746 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Oct 8 20:03:22.614958 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Oct 8 20:03:22.695079 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Oct 8 20:03:22.697240 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Oct 8 20:03:22.880062 systemd-networkd[782]: eth0: Gained IPv6LL Oct 8 20:03:23.115097 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Oct 8 20:03:23.380569 ignition[955]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Oct 8 20:03:23.383252 ignition[955]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Oct 8 20:03:23.403947 ignition[955]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Oct 8 20:03:23.409249 ignition[955]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Oct 8 20:03:23.410861 ignition[955]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Oct 8 20:03:23.410861 ignition[955]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Oct 8 20:03:23.410861 ignition[955]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Oct 8 20:03:23.410861 ignition[955]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 8 20:03:23.410861 ignition[955]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 8 20:03:23.410861 ignition[955]: INFO : files: files passed Oct 8 20:03:23.410861 ignition[955]: INFO : Ignition finished successfully Oct 8 20:03:23.411823 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 8 20:03:23.437908 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 8 20:03:23.440931 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 8 20:03:23.442775 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 8 20:03:23.442887 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 8 20:03:23.451006 initrd-setup-root-after-ignition[984]: grep: /sysroot/oem/oem-release: No such file or directory Oct 8 20:03:23.453971 initrd-setup-root-after-ignition[986]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 8 20:03:23.453971 initrd-setup-root-after-ignition[986]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 8 20:03:23.464526 initrd-setup-root-after-ignition[990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 8 20:03:23.456725 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 8 20:03:23.462177 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 8 20:03:23.475887 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 8 20:03:23.499501 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 8 20:03:23.499632 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 8 20:03:23.502017 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 8 20:03:23.504196 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 8 20:03:23.505285 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 8 20:03:23.516919 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 8 20:03:23.532049 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 8 20:03:23.541026 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 8 20:03:23.550847 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 8 20:03:23.552130 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 8 20:03:23.554452 systemd[1]: Stopped target timers.target - Timer Units. Oct 8 20:03:23.556516 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 8 20:03:23.556633 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 8 20:03:23.559053 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 8 20:03:23.560637 systemd[1]: Stopped target basic.target - Basic System. Oct 8 20:03:23.562752 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 8 20:03:23.564820 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 8 20:03:23.566993 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 8 20:03:23.569210 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 8 20:03:23.571490 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 8 20:03:23.573684 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 8 20:03:23.575667 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 8 20:03:23.577887 systemd[1]: Stopped target swap.target - Swaps. Oct 8 20:03:23.579748 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 8 20:03:23.579887 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 8 20:03:23.582019 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 8 20:03:23.583598 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 8 20:03:23.585702 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 8 20:03:23.585834 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 8 20:03:23.587971 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 8 20:03:23.588082 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 8 20:03:23.590469 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 8 20:03:23.590579 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 8 20:03:23.592415 systemd[1]: Stopped target paths.target - Path Units. Oct 8 20:03:23.594212 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 8 20:03:23.598800 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 8 20:03:23.601318 systemd[1]: Stopped target slices.target - Slice Units. Oct 8 20:03:23.603011 systemd[1]: Stopped target sockets.target - Socket Units. Oct 8 20:03:23.605029 systemd[1]: iscsid.socket: Deactivated successfully. Oct 8 20:03:23.605140 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 8 20:03:23.607468 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 8 20:03:23.607570 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 8 20:03:23.609480 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 8 20:03:23.609655 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 8 20:03:23.611588 systemd[1]: ignition-files.service: Deactivated successfully. Oct 8 20:03:23.611753 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 8 20:03:23.620864 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 8 20:03:23.622634 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 8 20:03:23.623763 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 8 20:03:23.623895 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 8 20:03:23.626132 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 8 20:03:23.626367 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 8 20:03:23.631433 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 8 20:03:23.631585 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 8 20:03:23.634894 ignition[1010]: INFO : Ignition 2.19.0 Oct 8 20:03:23.634894 ignition[1010]: INFO : Stage: umount Oct 8 20:03:23.634894 ignition[1010]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 8 20:03:23.634894 ignition[1010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Oct 8 20:03:23.634894 ignition[1010]: INFO : umount: umount passed Oct 8 20:03:23.634894 ignition[1010]: INFO : Ignition finished successfully Oct 8 20:03:23.637084 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 8 20:03:23.637219 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 8 20:03:23.639399 systemd[1]: Stopped target network.target - Network. Oct 8 20:03:23.640871 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 8 20:03:23.640945 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 8 20:03:23.642693 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 8 20:03:23.642771 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 8 20:03:23.644606 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 8 20:03:23.644656 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 8 20:03:23.646791 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 8 20:03:23.646842 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 8 20:03:23.649000 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 8 20:03:23.650920 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 8 20:03:23.653973 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 8 20:03:23.655796 systemd-networkd[782]: eth0: DHCPv6 lease lost Oct 8 20:03:23.659569 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 8 20:03:23.659778 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 8 20:03:23.662320 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 8 20:03:23.662473 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 8 20:03:23.666029 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 8 20:03:23.666100 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 8 20:03:23.674831 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 8 20:03:23.676101 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 8 20:03:23.676168 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 8 20:03:23.679095 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 8 20:03:23.679158 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 8 20:03:23.681277 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 8 20:03:23.681339 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 8 20:03:23.683837 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 8 20:03:23.683912 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 8 20:03:23.685587 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 8 20:03:23.700592 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 8 20:03:23.700788 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 8 20:03:23.703069 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 8 20:03:23.703279 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 8 20:03:23.705781 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 8 20:03:23.705887 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 8 20:03:23.707293 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 8 20:03:23.707345 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 8 20:03:23.709651 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 8 20:03:23.709729 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 8 20:03:23.712169 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 8 20:03:23.712232 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 8 20:03:23.714010 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 8 20:03:23.714073 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 8 20:03:23.736887 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 8 20:03:23.739221 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 8 20:03:23.739283 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 8 20:03:23.741620 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Oct 8 20:03:23.742926 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 8 20:03:23.746761 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 8 20:03:23.747772 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 8 20:03:23.750244 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 8 20:03:23.750300 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 8 20:03:23.754099 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 8 20:03:23.755253 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 8 20:03:24.002877 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 8 20:03:24.003957 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 8 20:03:24.006431 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 8 20:03:24.008466 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 8 20:03:24.009436 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 8 20:03:24.020856 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 8 20:03:24.030624 systemd[1]: Switching root. Oct 8 20:03:24.063662 systemd-journald[193]: Journal stopped Oct 8 20:03:25.459628 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Oct 8 20:03:25.459911 kernel: SELinux: policy capability network_peer_controls=1 Oct 8 20:03:25.459939 kernel: SELinux: policy capability open_perms=1 Oct 8 20:03:25.459954 kernel: SELinux: policy capability extended_socket_class=1 Oct 8 20:03:25.459971 kernel: SELinux: policy capability always_check_network=0 Oct 8 20:03:25.459987 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 8 20:03:25.460003 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 8 20:03:25.460019 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 8 20:03:25.460034 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 8 20:03:25.460057 kernel: audit: type=1403 audit(1728417804.554:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 8 20:03:25.460083 systemd[1]: Successfully loaded SELinux policy in 40.882ms. Oct 8 20:03:25.460126 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.822ms. Oct 8 20:03:25.460145 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 8 20:03:25.460162 systemd[1]: Detected virtualization kvm. Oct 8 20:03:25.460181 systemd[1]: Detected architecture x86-64. Oct 8 20:03:25.460198 systemd[1]: Detected first boot. Oct 8 20:03:25.460214 systemd[1]: Initializing machine ID from VM UUID. Oct 8 20:03:25.460232 zram_generator::config[1055]: No configuration found. Oct 8 20:03:25.460250 systemd[1]: Populated /etc with preset unit settings. Oct 8 20:03:25.460270 systemd[1]: initrd-switch-root.service: Deactivated successfully. Oct 8 20:03:25.460287 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Oct 8 20:03:25.460304 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Oct 8 20:03:25.460322 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 8 20:03:25.460339 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 8 20:03:25.460356 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 8 20:03:25.460373 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 8 20:03:25.460390 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 8 20:03:25.460407 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 8 20:03:25.460428 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 8 20:03:25.460444 systemd[1]: Created slice user.slice - User and Session Slice. Oct 8 20:03:25.460461 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 8 20:03:25.460477 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 8 20:03:25.460493 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 8 20:03:25.460509 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 8 20:03:25.460526 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 8 20:03:25.460543 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 8 20:03:25.460567 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 8 20:03:25.460595 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 8 20:03:25.460611 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Oct 8 20:03:25.460628 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Oct 8 20:03:25.460645 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Oct 8 20:03:25.460662 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 8 20:03:25.460679 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 8 20:03:25.460704 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 8 20:03:25.460742 systemd[1]: Reached target slices.target - Slice Units. Oct 8 20:03:25.460759 systemd[1]: Reached target swap.target - Swaps. Oct 8 20:03:25.460775 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 8 20:03:25.460792 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 8 20:03:25.460809 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 8 20:03:25.460826 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 8 20:03:25.460842 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 8 20:03:25.460858 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 8 20:03:25.460874 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 8 20:03:25.460891 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 8 20:03:25.460911 systemd[1]: Mounting media.mount - External Media Directory... Oct 8 20:03:25.460927 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:25.460946 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 8 20:03:25.460962 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 8 20:03:25.460978 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 8 20:03:25.460995 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 8 20:03:25.461011 systemd[1]: Reached target machines.target - Containers. Oct 8 20:03:25.461027 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 8 20:03:25.461047 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 8 20:03:25.461064 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 8 20:03:25.461081 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 8 20:03:25.461098 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 8 20:03:25.461114 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 8 20:03:25.461131 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 8 20:03:25.461147 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 8 20:03:25.461164 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 8 20:03:25.461184 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 8 20:03:25.461200 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Oct 8 20:03:25.461216 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Oct 8 20:03:25.461232 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Oct 8 20:03:25.461248 systemd[1]: Stopped systemd-fsck-usr.service. Oct 8 20:03:25.461264 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 8 20:03:25.461279 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 8 20:03:25.461297 kernel: fuse: init (API version 7.39) Oct 8 20:03:25.461313 kernel: loop: module loaded Oct 8 20:03:25.461333 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 8 20:03:25.461371 systemd-journald[1118]: Collecting audit messages is disabled. Oct 8 20:03:25.461401 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 8 20:03:25.461418 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 8 20:03:25.461435 systemd-journald[1118]: Journal started Oct 8 20:03:25.461465 systemd-journald[1118]: Runtime Journal (/run/log/journal/3ef00aaf0b6d472397d6ddd41391d4a0) is 6.0M, max 48.4M, 42.3M free. Oct 8 20:03:25.461515 systemd[1]: verity-setup.service: Deactivated successfully. Oct 8 20:03:25.104266 systemd[1]: Queued start job for default target multi-user.target. Oct 8 20:03:25.119430 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Oct 8 20:03:25.119910 systemd[1]: systemd-journald.service: Deactivated successfully. Oct 8 20:03:25.462856 systemd[1]: Stopped verity-setup.service. Oct 8 20:03:25.464386 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:25.469062 systemd[1]: Started systemd-journald.service - Journal Service. Oct 8 20:03:25.471484 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 8 20:03:25.472729 kernel: ACPI: bus type drm_connector registered Oct 8 20:03:25.473316 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 8 20:03:25.474607 systemd[1]: Mounted media.mount - External Media Directory. Oct 8 20:03:25.475778 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 8 20:03:25.476991 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 8 20:03:25.478237 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 8 20:03:25.479482 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 8 20:03:25.481074 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 8 20:03:25.481276 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 8 20:03:25.482804 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 8 20:03:25.483043 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 8 20:03:25.484488 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 8 20:03:25.484671 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 8 20:03:25.486056 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 8 20:03:25.486239 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 8 20:03:25.488163 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 8 20:03:25.488426 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 8 20:03:25.489985 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 8 20:03:25.490225 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 8 20:03:25.491785 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 8 20:03:25.493409 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 8 20:03:25.495334 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 8 20:03:25.510787 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 8 20:03:25.523887 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 8 20:03:25.526590 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 8 20:03:25.527774 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 8 20:03:25.527805 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 8 20:03:25.529816 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Oct 8 20:03:25.535558 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 8 20:03:25.537998 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 8 20:03:25.539226 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 8 20:03:25.569973 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 8 20:03:25.577098 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 8 20:03:25.579921 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 8 20:03:25.586551 systemd-journald[1118]: Time spent on flushing to /var/log/journal/3ef00aaf0b6d472397d6ddd41391d4a0 is 34.803ms for 950 entries. Oct 8 20:03:25.586551 systemd-journald[1118]: System Journal (/var/log/journal/3ef00aaf0b6d472397d6ddd41391d4a0) is 8.0M, max 195.6M, 187.6M free. Oct 8 20:03:25.908524 systemd-journald[1118]: Received client request to flush runtime journal. Oct 8 20:03:25.908568 kernel: loop0: detected capacity change from 0 to 142488 Oct 8 20:03:25.908592 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 8 20:03:25.908607 kernel: loop1: detected capacity change from 0 to 205544 Oct 8 20:03:25.908622 kernel: loop2: detected capacity change from 0 to 140768 Oct 8 20:03:25.908637 kernel: loop3: detected capacity change from 0 to 142488 Oct 8 20:03:25.908652 kernel: loop4: detected capacity change from 0 to 205544 Oct 8 20:03:25.908667 kernel: loop5: detected capacity change from 0 to 140768 Oct 8 20:03:25.908693 zram_generator::config[1201]: No configuration found. Oct 8 20:03:25.584777 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 8 20:03:25.586005 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 8 20:03:25.587411 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 8 20:03:25.591891 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 8 20:03:25.598216 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 8 20:03:25.601202 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 8 20:03:25.602997 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 8 20:03:25.604334 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 8 20:03:25.607209 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 8 20:03:25.625838 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Oct 8 20:03:25.627417 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 8 20:03:25.635196 udevadm[1163]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Oct 8 20:03:25.636418 systemd-tmpfiles[1155]: ACLs are not supported, ignoring. Oct 8 20:03:25.636433 systemd-tmpfiles[1155]: ACLs are not supported, ignoring. Oct 8 20:03:25.642198 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 8 20:03:25.800356 (sd-merge)[1176]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Oct 8 20:03:25.801976 (sd-merge)[1176]: Merged extensions into '/usr'. Oct 8 20:03:25.844819 systemd[1]: Reloading requested from client PID 1154 ('systemd-sysext') (unit systemd-sysext.service)... Oct 8 20:03:25.844830 systemd[1]: Reloading... Oct 8 20:03:26.028590 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 8 20:03:26.058968 ldconfig[1149]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 8 20:03:26.079737 systemd[1]: Reloading finished in 234 ms. Oct 8 20:03:26.114150 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 8 20:03:26.125100 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 8 20:03:26.126704 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 8 20:03:26.128281 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 8 20:03:26.135631 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 8 20:03:26.152883 systemd[1]: Starting ensure-sysext.service... Oct 8 20:03:26.154843 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Oct 8 20:03:26.157063 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 8 20:03:26.174496 systemd[1]: Reloading requested from client PID 1251 ('systemctl') (unit ensure-sysext.service)... Oct 8 20:03:26.174510 systemd[1]: Reloading... Oct 8 20:03:26.228791 zram_generator::config[1285]: No configuration found. Oct 8 20:03:26.327026 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 8 20:03:26.376199 systemd[1]: Reloading finished in 201 ms. Oct 8 20:03:26.392849 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 8 20:03:26.399232 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 8 20:03:26.405926 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 8 20:03:26.429341 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 8 20:03:26.434035 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.434280 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 8 20:03:26.435889 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 8 20:03:26.439768 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 8 20:03:26.446103 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 8 20:03:26.447929 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 8 20:03:26.448094 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.449449 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 8 20:03:26.449689 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 8 20:03:26.451529 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 8 20:03:26.451779 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 8 20:03:26.451944 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. Oct 8 20:03:26.451967 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. Oct 8 20:03:26.458385 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 8 20:03:26.458704 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 8 20:03:26.459633 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 8 20:03:26.460034 systemd-tmpfiles[1322]: ACLs are not supported, ignoring. Oct 8 20:03:26.460112 systemd-tmpfiles[1322]: ACLs are not supported, ignoring. Oct 8 20:03:26.463252 systemd-tmpfiles[1322]: Detected autofs mount point /boot during canonicalization of boot. Oct 8 20:03:26.463260 systemd-tmpfiles[1322]: Skipping /boot Oct 8 20:03:26.470212 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 8 20:03:26.473394 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.473750 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 8 20:03:26.474010 systemd-tmpfiles[1322]: Detected autofs mount point /boot during canonicalization of boot. Oct 8 20:03:26.474025 systemd-tmpfiles[1322]: Skipping /boot Oct 8 20:03:26.477104 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 8 20:03:26.499238 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 8 20:03:26.500616 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 8 20:03:26.500819 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.501943 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 8 20:03:26.504014 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 8 20:03:26.504231 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 8 20:03:26.508526 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 8 20:03:26.508722 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 8 20:03:26.510457 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 8 20:03:26.510630 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 8 20:03:26.516767 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.527934 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 8 20:03:26.530754 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 8 20:03:26.532079 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 8 20:03:26.533110 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 8 20:03:26.537883 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 8 20:03:26.542658 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 8 20:03:26.545866 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 8 20:03:26.547098 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 8 20:03:26.549347 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 8 20:03:26.551731 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 8 20:03:26.552920 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 8 20:03:26.553753 systemd[1]: Finished ensure-sysext.service. Oct 8 20:03:26.555064 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 8 20:03:26.555267 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 8 20:03:26.556831 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 8 20:03:26.557003 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 8 20:03:26.562535 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 8 20:03:26.564593 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 8 20:03:26.647204 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 8 20:03:26.785685 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 8 20:03:26.809829 augenrules[1370]: No rules Oct 8 20:03:26.812225 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 8 20:03:26.814053 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 8 20:03:26.915283 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 8 20:03:26.916946 systemd[1]: Reached target time-set.target - System Time Set. Oct 8 20:03:26.920278 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 8 20:03:26.926139 systemd-resolved[1354]: Positive Trust Anchors: Oct 8 20:03:26.926157 systemd-resolved[1354]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 8 20:03:26.926189 systemd-resolved[1354]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 8 20:03:26.968900 systemd-resolved[1354]: Defaulting to hostname 'linux'. Oct 8 20:03:26.970556 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 8 20:03:26.972037 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 8 20:03:26.988949 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 8 20:03:26.994730 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 8 20:03:27.172852 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 8 20:03:27.189945 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 8 20:03:27.208280 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 8 20:03:27.223198 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 8 20:03:27.234340 systemd-udevd[1382]: Using default interface naming scheme 'v255'. Oct 8 20:03:27.252999 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 8 20:03:27.268040 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 8 20:03:27.288257 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Oct 8 20:03:27.291730 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1388) Oct 8 20:03:27.321829 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1396) Oct 8 20:03:27.333736 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1396) Oct 8 20:03:27.334662 systemd-networkd[1389]: lo: Link UP Oct 8 20:03:27.334674 systemd-networkd[1389]: lo: Gained carrier Oct 8 20:03:27.336706 systemd-networkd[1389]: Enumeration completed Oct 8 20:03:27.336995 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 8 20:03:27.338502 systemd[1]: Reached target network.target - Network. Oct 8 20:03:27.341951 systemd-networkd[1389]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 8 20:03:27.341962 systemd-networkd[1389]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 8 20:03:27.342897 systemd-networkd[1389]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 8 20:03:27.342936 systemd-networkd[1389]: eth0: Link UP Oct 8 20:03:27.342941 systemd-networkd[1389]: eth0: Gained carrier Oct 8 20:03:27.342954 systemd-networkd[1389]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 8 20:03:27.347899 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 8 20:03:27.352586 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 8 20:03:27.354479 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Oct 8 20:03:27.370734 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Oct 8 20:03:27.375882 systemd-networkd[1389]: eth0: DHCPv4 address 10.0.0.126/16, gateway 10.0.0.1 acquired from 10.0.0.1 Oct 8 20:03:27.378789 systemd-timesyncd[1359]: Network configuration changed, trying to establish connection. Oct 8 20:03:27.379059 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 8 20:03:28.493158 systemd-timesyncd[1359]: Contacted time server 10.0.0.1:123 (10.0.0.1). Oct 8 20:03:28.493211 systemd-timesyncd[1359]: Initial clock synchronization to Tue 2024-10-08 20:03:28.493056 UTC. Oct 8 20:03:28.493249 systemd-resolved[1354]: Clock change detected. Flushing caches. Oct 8 20:03:28.518663 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Oct 8 20:03:28.518828 kernel: ACPI: button: Power Button [PWRF] Oct 8 20:03:28.524878 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Oct 8 20:03:28.525136 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Oct 8 20:03:28.525323 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Oct 8 20:03:28.526339 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 8 20:03:28.546112 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 8 20:03:28.603155 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 8 20:03:28.626533 kernel: mousedev: PS/2 mouse device common for all mice Oct 8 20:03:28.651648 kernel: kvm_amd: TSC scaling supported Oct 8 20:03:28.651803 kernel: kvm_amd: Nested Virtualization enabled Oct 8 20:03:28.651833 kernel: kvm_amd: Nested Paging enabled Oct 8 20:03:28.651860 kernel: kvm_amd: LBR virtualization supported Oct 8 20:03:28.652926 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Oct 8 20:03:28.652969 kernel: kvm_amd: Virtual GIF supported Oct 8 20:03:28.672534 kernel: EDAC MC: Ver: 3.0.0 Oct 8 20:03:28.704064 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Oct 8 20:03:28.719549 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 8 20:03:28.733605 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Oct 8 20:03:28.741715 lvm[1429]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 8 20:03:28.770387 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Oct 8 20:03:28.772001 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 8 20:03:28.773261 systemd[1]: Reached target sysinit.target - System Initialization. Oct 8 20:03:28.774610 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 8 20:03:28.776051 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 8 20:03:28.777689 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 8 20:03:28.779059 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 8 20:03:28.780507 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 8 20:03:28.781926 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 8 20:03:28.781954 systemd[1]: Reached target paths.target - Path Units. Oct 8 20:03:28.783000 systemd[1]: Reached target timers.target - Timer Units. Oct 8 20:03:28.784846 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 8 20:03:28.787489 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 8 20:03:28.797496 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 8 20:03:28.799841 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Oct 8 20:03:28.801448 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 8 20:03:28.802691 systemd[1]: Reached target sockets.target - Socket Units. Oct 8 20:03:28.803715 systemd[1]: Reached target basic.target - Basic System. Oct 8 20:03:28.804784 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 8 20:03:28.804815 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 8 20:03:28.805836 systemd[1]: Starting containerd.service - containerd container runtime... Oct 8 20:03:28.808444 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 8 20:03:28.809696 lvm[1433]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 8 20:03:28.813062 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 8 20:03:28.816146 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 8 20:03:28.817608 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 8 20:03:28.818917 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 8 20:03:28.821598 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 8 20:03:28.828636 jq[1436]: false Oct 8 20:03:28.825290 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 8 20:03:28.831684 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 8 20:03:28.838047 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 8 20:03:28.839596 dbus-daemon[1435]: [system] SELinux support is enabled Oct 8 20:03:28.839877 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 8 20:03:28.840512 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 8 20:03:28.842701 systemd[1]: Starting update-engine.service - Update Engine... Oct 8 20:03:28.845120 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 8 20:03:28.846695 extend-filesystems[1437]: Found loop3 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found loop4 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found loop5 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found sr0 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda1 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda2 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda3 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found usr Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda4 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda6 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda7 Oct 8 20:03:28.855573 extend-filesystems[1437]: Found vda9 Oct 8 20:03:28.855573 extend-filesystems[1437]: Checking size of /dev/vda9 Oct 8 20:03:28.848939 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 8 20:03:28.853669 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Oct 8 20:03:28.874885 update_engine[1449]: I20241008 20:03:28.861340 1449 main.cc:92] Flatcar Update Engine starting Oct 8 20:03:28.874885 update_engine[1449]: I20241008 20:03:28.863306 1449 update_check_scheduler.cc:74] Next update check in 7m16s Oct 8 20:03:28.863915 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 8 20:03:28.875290 jq[1451]: true Oct 8 20:03:28.864761 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 8 20:03:28.865183 systemd[1]: motdgen.service: Deactivated successfully. Oct 8 20:03:28.865479 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 8 20:03:28.867057 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 8 20:03:28.867315 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 8 20:03:28.876194 extend-filesystems[1437]: Resized partition /dev/vda9 Oct 8 20:03:28.883076 extend-filesystems[1466]: resize2fs 1.47.1 (20-May-2024) Oct 8 20:03:28.885015 jq[1460]: true Oct 8 20:03:28.892198 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Oct 8 20:03:28.890014 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 8 20:03:28.890066 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 8 20:03:28.892282 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 8 20:03:28.892306 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 8 20:03:28.892413 (ntainerd)[1462]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 8 20:03:28.894080 systemd[1]: Started update-engine.service - Update Engine. Oct 8 20:03:28.902705 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 8 20:03:28.912766 tar[1457]: linux-amd64/helm Oct 8 20:03:28.913495 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1395) Oct 8 20:03:28.939533 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Oct 8 20:03:28.974535 locksmithd[1473]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 8 20:03:29.004880 systemd-logind[1448]: Watching system buttons on /dev/input/event1 (Power Button) Oct 8 20:03:29.005264 systemd-logind[1448]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 8 20:03:29.005929 extend-filesystems[1466]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Oct 8 20:03:29.005929 extend-filesystems[1466]: old_desc_blocks = 1, new_desc_blocks = 1 Oct 8 20:03:29.005929 extend-filesystems[1466]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Oct 8 20:03:29.016888 extend-filesystems[1437]: Resized filesystem in /dev/vda9 Oct 8 20:03:29.009525 systemd-logind[1448]: New seat seat0. Oct 8 20:03:29.011807 systemd[1]: Started systemd-logind.service - User Login Management. Oct 8 20:03:29.018026 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 8 20:03:29.018309 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 8 20:03:29.184606 containerd[1462]: time="2024-10-08T20:03:29.184533006Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Oct 8 20:03:29.190529 sshd_keygen[1456]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 8 20:03:29.209354 containerd[1462]: time="2024-10-08T20:03:29.209294482Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211087315Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.54-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211119805Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211138971Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211334248Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211353594Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211445246Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211461997Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211711415Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211730170Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211745148Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212505 containerd[1462]: time="2024-10-08T20:03:29.211756600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.211878278Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.212139408Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.212280953Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.212297294Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.212412670Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Oct 8 20:03:29.212810 containerd[1462]: time="2024-10-08T20:03:29.212489745Z" level=info msg="metadata content store policy set" policy=shared Oct 8 20:03:29.216514 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 8 20:03:29.225710 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 8 20:03:29.234117 systemd[1]: issuegen.service: Deactivated successfully. Oct 8 20:03:29.234378 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 8 20:03:29.241722 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 8 20:03:29.254594 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 8 20:03:29.265813 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 8 20:03:29.268312 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 8 20:03:29.269830 systemd[1]: Reached target getty.target - Login Prompts. Oct 8 20:03:29.340502 containerd[1462]: time="2024-10-08T20:03:29.340428515Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Oct 8 20:03:29.340624 containerd[1462]: time="2024-10-08T20:03:29.340523053Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Oct 8 20:03:29.340624 containerd[1462]: time="2024-10-08T20:03:29.340541537Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Oct 8 20:03:29.340624 containerd[1462]: time="2024-10-08T20:03:29.340556465Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Oct 8 20:03:29.340624 containerd[1462]: time="2024-10-08T20:03:29.340570993Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Oct 8 20:03:29.340781 containerd[1462]: time="2024-10-08T20:03:29.340758985Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Oct 8 20:03:29.340831 bash[1489]: Updated "/home/core/.ssh/authorized_keys" Oct 8 20:03:29.341108 containerd[1462]: time="2024-10-08T20:03:29.341032077Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341143837Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341162813Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341179604Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341196426Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341211644Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341233084Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341250778Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341268401Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341284511Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341299399Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341314687Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341338201Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341356406Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341651 containerd[1462]: time="2024-10-08T20:03:29.341373217Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341399737Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341415476Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341434753Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341449270Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341464118Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341501207Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341515865Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341527336Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341538547Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341550069Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341565888Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341586647Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341599031Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.341935 containerd[1462]: time="2024-10-08T20:03:29.341609931Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341654775Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341672639Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341686655Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341698046Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341707504Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341720018Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341735006Z" level=info msg="NRI interface is disabled by configuration." Oct 8 20:03:29.342179 containerd[1462]: time="2024-10-08T20:03:29.341744894Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Oct 8 20:03:29.342325 containerd[1462]: time="2024-10-08T20:03:29.342005433Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Oct 8 20:03:29.342325 containerd[1462]: time="2024-10-08T20:03:29.342058302Z" level=info msg="Connect containerd service" Oct 8 20:03:29.342325 containerd[1462]: time="2024-10-08T20:03:29.342102986Z" level=info msg="using legacy CRI server" Oct 8 20:03:29.342325 containerd[1462]: time="2024-10-08T20:03:29.342109778Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 8 20:03:29.342325 containerd[1462]: time="2024-10-08T20:03:29.342203043Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Oct 8 20:03:29.343287 containerd[1462]: time="2024-10-08T20:03:29.342836341Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 8 20:03:29.343287 containerd[1462]: time="2024-10-08T20:03:29.343185275Z" level=info msg="Start subscribing containerd event" Oct 8 20:03:29.343287 containerd[1462]: time="2024-10-08T20:03:29.343231192Z" level=info msg="Start recovering state" Oct 8 20:03:29.343364 containerd[1462]: time="2024-10-08T20:03:29.343297175Z" level=info msg="Start event monitor" Oct 8 20:03:29.343364 containerd[1462]: time="2024-10-08T20:03:29.343314798Z" level=info msg="Start snapshots syncer" Oct 8 20:03:29.343364 containerd[1462]: time="2024-10-08T20:03:29.343325749Z" level=info msg="Start cni network conf syncer for default" Oct 8 20:03:29.343364 containerd[1462]: time="2024-10-08T20:03:29.343336920Z" level=info msg="Start streaming server" Oct 8 20:03:29.343630 containerd[1462]: time="2024-10-08T20:03:29.343603029Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 8 20:03:29.343672 containerd[1462]: time="2024-10-08T20:03:29.343658493Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 8 20:03:29.343779 containerd[1462]: time="2024-10-08T20:03:29.343743212Z" level=info msg="containerd successfully booted in 0.161028s" Oct 8 20:03:29.343813 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 8 20:03:29.346658 systemd[1]: Started containerd.service - containerd container runtime. Oct 8 20:03:29.350048 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Oct 8 20:03:29.391005 tar[1457]: linux-amd64/LICENSE Oct 8 20:03:29.391126 tar[1457]: linux-amd64/README.md Oct 8 20:03:29.405820 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 8 20:03:29.751739 systemd-networkd[1389]: eth0: Gained IPv6LL Oct 8 20:03:29.755309 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 8 20:03:29.757334 systemd[1]: Reached target network-online.target - Network is Online. Oct 8 20:03:29.776711 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Oct 8 20:03:29.779555 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:03:29.781823 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 8 20:03:29.804228 systemd[1]: coreos-metadata.service: Deactivated successfully. Oct 8 20:03:29.804552 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Oct 8 20:03:29.806289 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 8 20:03:29.806813 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 8 20:03:30.457045 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:03:30.458978 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 8 20:03:30.462476 systemd[1]: Startup finished in 698ms (kernel) + 5.857s (initrd) + 4.834s (userspace) = 11.390s. Oct 8 20:03:30.473680 (kubelet)[1548]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 8 20:03:30.950409 kubelet[1548]: E1008 20:03:30.950338 1548 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 8 20:03:30.954546 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 8 20:03:30.954761 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 8 20:03:30.955098 systemd[1]: kubelet.service: Consumed 1.022s CPU time. Oct 8 20:03:38.672500 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 8 20:03:38.674040 systemd[1]: Started sshd@0-10.0.0.126:22-10.0.0.1:54310.service - OpenSSH per-connection server daemon (10.0.0.1:54310). Oct 8 20:03:38.723930 sshd[1562]: Accepted publickey for core from 10.0.0.1 port 54310 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:38.726440 sshd[1562]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:38.738209 systemd-logind[1448]: New session 1 of user core. Oct 8 20:03:38.740041 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 8 20:03:38.746912 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 8 20:03:38.762533 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 8 20:03:38.773920 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 8 20:03:38.777684 (systemd)[1566]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 8 20:03:38.930713 systemd[1566]: Queued start job for default target default.target. Oct 8 20:03:38.949422 systemd[1566]: Created slice app.slice - User Application Slice. Oct 8 20:03:38.949459 systemd[1566]: Reached target paths.target - Paths. Oct 8 20:03:38.949494 systemd[1566]: Reached target timers.target - Timers. Oct 8 20:03:38.951524 systemd[1566]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 8 20:03:38.965050 systemd[1566]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 8 20:03:38.965247 systemd[1566]: Reached target sockets.target - Sockets. Oct 8 20:03:38.965270 systemd[1566]: Reached target basic.target - Basic System. Oct 8 20:03:38.965321 systemd[1566]: Reached target default.target - Main User Target. Oct 8 20:03:38.965366 systemd[1566]: Startup finished in 178ms. Oct 8 20:03:38.965890 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 8 20:03:38.967985 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 8 20:03:39.034323 systemd[1]: Started sshd@1-10.0.0.126:22-10.0.0.1:54320.service - OpenSSH per-connection server daemon (10.0.0.1:54320). Oct 8 20:03:39.076999 sshd[1577]: Accepted publickey for core from 10.0.0.1 port 54320 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:39.078652 sshd[1577]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:39.083104 systemd-logind[1448]: New session 2 of user core. Oct 8 20:03:39.096791 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 8 20:03:39.151762 sshd[1577]: pam_unix(sshd:session): session closed for user core Oct 8 20:03:39.164999 systemd[1]: sshd@1-10.0.0.126:22-10.0.0.1:54320.service: Deactivated successfully. Oct 8 20:03:39.167492 systemd[1]: session-2.scope: Deactivated successfully. Oct 8 20:03:39.169422 systemd-logind[1448]: Session 2 logged out. Waiting for processes to exit. Oct 8 20:03:39.176899 systemd[1]: Started sshd@2-10.0.0.126:22-10.0.0.1:54324.service - OpenSSH per-connection server daemon (10.0.0.1:54324). Oct 8 20:03:39.177817 systemd-logind[1448]: Removed session 2. Oct 8 20:03:39.210928 sshd[1584]: Accepted publickey for core from 10.0.0.1 port 54324 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:39.212887 sshd[1584]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:39.217389 systemd-logind[1448]: New session 3 of user core. Oct 8 20:03:39.231645 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 8 20:03:39.282595 sshd[1584]: pam_unix(sshd:session): session closed for user core Oct 8 20:03:39.295902 systemd[1]: sshd@2-10.0.0.126:22-10.0.0.1:54324.service: Deactivated successfully. Oct 8 20:03:39.297906 systemd[1]: session-3.scope: Deactivated successfully. Oct 8 20:03:39.299661 systemd-logind[1448]: Session 3 logged out. Waiting for processes to exit. Oct 8 20:03:39.310835 systemd[1]: Started sshd@3-10.0.0.126:22-10.0.0.1:54330.service - OpenSSH per-connection server daemon (10.0.0.1:54330). Oct 8 20:03:39.312207 systemd-logind[1448]: Removed session 3. Oct 8 20:03:39.343505 sshd[1591]: Accepted publickey for core from 10.0.0.1 port 54330 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:39.345385 sshd[1591]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:39.349959 systemd-logind[1448]: New session 4 of user core. Oct 8 20:03:39.363594 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 8 20:03:39.420402 sshd[1591]: pam_unix(sshd:session): session closed for user core Oct 8 20:03:39.432566 systemd[1]: sshd@3-10.0.0.126:22-10.0.0.1:54330.service: Deactivated successfully. Oct 8 20:03:39.434456 systemd[1]: session-4.scope: Deactivated successfully. Oct 8 20:03:39.436108 systemd-logind[1448]: Session 4 logged out. Waiting for processes to exit. Oct 8 20:03:39.449042 systemd[1]: Started sshd@4-10.0.0.126:22-10.0.0.1:54336.service - OpenSSH per-connection server daemon (10.0.0.1:54336). Oct 8 20:03:39.450163 systemd-logind[1448]: Removed session 4. Oct 8 20:03:39.480326 sshd[1599]: Accepted publickey for core from 10.0.0.1 port 54336 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:39.481998 sshd[1599]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:39.486403 systemd-logind[1448]: New session 5 of user core. Oct 8 20:03:39.495610 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 8 20:03:39.937514 sudo[1602]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 8 20:03:39.937878 sudo[1602]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 8 20:03:39.958265 sudo[1602]: pam_unix(sudo:session): session closed for user root Oct 8 20:03:39.960500 sshd[1599]: pam_unix(sshd:session): session closed for user core Oct 8 20:03:39.973688 systemd[1]: sshd@4-10.0.0.126:22-10.0.0.1:54336.service: Deactivated successfully. Oct 8 20:03:39.975612 systemd[1]: session-5.scope: Deactivated successfully. Oct 8 20:03:39.978198 systemd-logind[1448]: Session 5 logged out. Waiting for processes to exit. Oct 8 20:03:39.983841 systemd[1]: Started sshd@5-10.0.0.126:22-10.0.0.1:54338.service - OpenSSH per-connection server daemon (10.0.0.1:54338). Oct 8 20:03:39.984923 systemd-logind[1448]: Removed session 5. Oct 8 20:03:40.019553 sshd[1607]: Accepted publickey for core from 10.0.0.1 port 54338 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:40.021260 sshd[1607]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:40.025488 systemd-logind[1448]: New session 6 of user core. Oct 8 20:03:40.042746 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 8 20:03:40.097996 sudo[1612]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 8 20:03:40.098348 sudo[1612]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 8 20:03:40.102653 sudo[1612]: pam_unix(sudo:session): session closed for user root Oct 8 20:03:40.109012 sudo[1611]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Oct 8 20:03:40.109375 sudo[1611]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 8 20:03:40.128816 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Oct 8 20:03:40.130439 auditctl[1615]: No rules Oct 8 20:03:40.131853 systemd[1]: audit-rules.service: Deactivated successfully. Oct 8 20:03:40.132153 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Oct 8 20:03:40.134146 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 8 20:03:40.167089 augenrules[1633]: No rules Oct 8 20:03:40.168940 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 8 20:03:40.170215 sudo[1611]: pam_unix(sudo:session): session closed for user root Oct 8 20:03:40.172191 sshd[1607]: pam_unix(sshd:session): session closed for user core Oct 8 20:03:40.192211 systemd[1]: sshd@5-10.0.0.126:22-10.0.0.1:54338.service: Deactivated successfully. Oct 8 20:03:40.194065 systemd[1]: session-6.scope: Deactivated successfully. Oct 8 20:03:40.195828 systemd-logind[1448]: Session 6 logged out. Waiting for processes to exit. Oct 8 20:03:40.208828 systemd[1]: Started sshd@6-10.0.0.126:22-10.0.0.1:54346.service - OpenSSH per-connection server daemon (10.0.0.1:54346). Oct 8 20:03:40.209815 systemd-logind[1448]: Removed session 6. Oct 8 20:03:40.239841 sshd[1641]: Accepted publickey for core from 10.0.0.1 port 54346 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:03:40.241574 sshd[1641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:03:40.245606 systemd-logind[1448]: New session 7 of user core. Oct 8 20:03:40.259618 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 8 20:03:40.313933 sudo[1644]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 8 20:03:40.314354 sudo[1644]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 8 20:03:40.628765 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 8 20:03:40.628896 (dockerd)[1662]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 8 20:03:40.936988 dockerd[1662]: time="2024-10-08T20:03:40.936830969Z" level=info msg="Starting up" Oct 8 20:03:41.073642 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 8 20:03:41.088716 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:03:41.419006 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:03:41.425619 (kubelet)[1694]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 8 20:03:41.527491 kubelet[1694]: E1008 20:03:41.527416 1694 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 8 20:03:41.533902 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 8 20:03:41.534152 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 8 20:03:42.751040 dockerd[1662]: time="2024-10-08T20:03:42.750983102Z" level=info msg="Loading containers: start." Oct 8 20:03:42.877500 kernel: Initializing XFRM netlink socket Oct 8 20:03:42.971195 systemd-networkd[1389]: docker0: Link UP Oct 8 20:03:42.994271 dockerd[1662]: time="2024-10-08T20:03:42.994218425Z" level=info msg="Loading containers: done." Oct 8 20:03:43.009320 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3681401597-merged.mount: Deactivated successfully. Oct 8 20:03:43.012672 dockerd[1662]: time="2024-10-08T20:03:43.012624203Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 8 20:03:43.012776 dockerd[1662]: time="2024-10-08T20:03:43.012752153Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Oct 8 20:03:43.012905 dockerd[1662]: time="2024-10-08T20:03:43.012880734Z" level=info msg="Daemon has completed initialization" Oct 8 20:03:43.055668 dockerd[1662]: time="2024-10-08T20:03:43.055572977Z" level=info msg="API listen on /run/docker.sock" Oct 8 20:03:43.055812 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 8 20:03:44.046976 containerd[1462]: time="2024-10-08T20:03:44.046931228Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.0\"" Oct 8 20:03:45.206558 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1380535368.mount: Deactivated successfully. Oct 8 20:03:47.334969 containerd[1462]: time="2024-10-08T20:03:47.334881248Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:47.335691 containerd[1462]: time="2024-10-08T20:03:47.335543379Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.0: active requests=0, bytes read=28066621" Oct 8 20:03:47.337145 containerd[1462]: time="2024-10-08T20:03:47.337087575Z" level=info msg="ImageCreate event name:\"sha256:604f5db92eaa823d11c141d8825f1460206f6bf29babca2a909a698dc22055d3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:47.340087 containerd[1462]: time="2024-10-08T20:03:47.340027689Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:470179274deb9dc3a81df55cfc24823ce153147d4ebf2ed649a4f271f51eaddf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:47.341359 containerd[1462]: time="2024-10-08T20:03:47.341306347Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.0\" with image id \"sha256:604f5db92eaa823d11c141d8825f1460206f6bf29babca2a909a698dc22055d3\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.0\", repo digest \"registry.k8s.io/kube-apiserver@sha256:470179274deb9dc3a81df55cfc24823ce153147d4ebf2ed649a4f271f51eaddf\", size \"28063421\" in 3.294330155s" Oct 8 20:03:47.341408 containerd[1462]: time="2024-10-08T20:03:47.341360809Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.0\" returns image reference \"sha256:604f5db92eaa823d11c141d8825f1460206f6bf29babca2a909a698dc22055d3\"" Oct 8 20:03:47.343293 containerd[1462]: time="2024-10-08T20:03:47.343259830Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.0\"" Oct 8 20:03:48.626826 containerd[1462]: time="2024-10-08T20:03:48.626758819Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:48.627658 containerd[1462]: time="2024-10-08T20:03:48.627614614Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.0: active requests=0, bytes read=24690922" Oct 8 20:03:48.628975 containerd[1462]: time="2024-10-08T20:03:48.628947093Z" level=info msg="ImageCreate event name:\"sha256:045733566833c40b15806c9b87d27f08e455e069833752e0e6ad7a76d37cb2b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:48.632035 containerd[1462]: time="2024-10-08T20:03:48.632003916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f6f3c33dda209e8434b83dacf5244c03b59b0018d93325ff21296a142b68497d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:48.633288 containerd[1462]: time="2024-10-08T20:03:48.633256705Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.0\" with image id \"sha256:045733566833c40b15806c9b87d27f08e455e069833752e0e6ad7a76d37cb2b1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.0\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f6f3c33dda209e8434b83dacf5244c03b59b0018d93325ff21296a142b68497d\", size \"26240868\" in 1.289963732s" Oct 8 20:03:48.633343 containerd[1462]: time="2024-10-08T20:03:48.633289096Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.0\" returns image reference \"sha256:045733566833c40b15806c9b87d27f08e455e069833752e0e6ad7a76d37cb2b1\"" Oct 8 20:03:48.633854 containerd[1462]: time="2024-10-08T20:03:48.633810824Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.0\"" Oct 8 20:03:50.043838 containerd[1462]: time="2024-10-08T20:03:50.043775692Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:50.045101 containerd[1462]: time="2024-10-08T20:03:50.045048098Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.0: active requests=0, bytes read=18646758" Oct 8 20:03:50.046223 containerd[1462]: time="2024-10-08T20:03:50.046188036Z" level=info msg="ImageCreate event name:\"sha256:1766f54c897f0e57040741e6741462f2e3a7d754705f446c9f729c7e1230fb94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:50.048946 containerd[1462]: time="2024-10-08T20:03:50.048886867Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:96ddae9c9b2e79342e0551e2d2ec422c0c02629a74d928924aaa069706619808\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:50.050072 containerd[1462]: time="2024-10-08T20:03:50.049997219Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.0\" with image id \"sha256:1766f54c897f0e57040741e6741462f2e3a7d754705f446c9f729c7e1230fb94\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.0\", repo digest \"registry.k8s.io/kube-scheduler@sha256:96ddae9c9b2e79342e0551e2d2ec422c0c02629a74d928924aaa069706619808\", size \"20196722\" in 1.416149176s" Oct 8 20:03:50.050166 containerd[1462]: time="2024-10-08T20:03:50.050072771Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.0\" returns image reference \"sha256:1766f54c897f0e57040741e6741462f2e3a7d754705f446c9f729c7e1230fb94\"" Oct 8 20:03:50.050570 containerd[1462]: time="2024-10-08T20:03:50.050542442Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.0\"" Oct 8 20:03:51.481614 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2235160997.mount: Deactivated successfully. Oct 8 20:03:51.573579 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Oct 8 20:03:51.582889 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:03:51.737780 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:03:51.742463 (kubelet)[1905]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 8 20:03:51.777543 kubelet[1905]: E1008 20:03:51.777487 1905 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 8 20:03:51.781449 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 8 20:03:51.781694 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 8 20:03:52.093934 containerd[1462]: time="2024-10-08T20:03:52.093850784Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:52.095754 containerd[1462]: time="2024-10-08T20:03:52.095710662Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.0: active requests=0, bytes read=30208881" Oct 8 20:03:52.097036 containerd[1462]: time="2024-10-08T20:03:52.096999679Z" level=info msg="ImageCreate event name:\"sha256:ad83b2ca7b09e6162f96f933eecded731cbebf049c78f941fd0ce560a86b6494\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:52.099370 containerd[1462]: time="2024-10-08T20:03:52.099303510Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c727efb1c6f15a68060bf7f207f5c7a765355b7e3340c513e582ec819c5cd2fe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:52.099953 containerd[1462]: time="2024-10-08T20:03:52.099904828Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.0\" with image id \"sha256:ad83b2ca7b09e6162f96f933eecded731cbebf049c78f941fd0ce560a86b6494\", repo tag \"registry.k8s.io/kube-proxy:v1.31.0\", repo digest \"registry.k8s.io/kube-proxy@sha256:c727efb1c6f15a68060bf7f207f5c7a765355b7e3340c513e582ec819c5cd2fe\", size \"30207900\" in 2.049262368s" Oct 8 20:03:52.099953 containerd[1462]: time="2024-10-08T20:03:52.099944332Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.0\" returns image reference \"sha256:ad83b2ca7b09e6162f96f933eecded731cbebf049c78f941fd0ce560a86b6494\"" Oct 8 20:03:52.100445 containerd[1462]: time="2024-10-08T20:03:52.100418270Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Oct 8 20:03:52.683702 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3864138839.mount: Deactivated successfully. Oct 8 20:03:53.793556 containerd[1462]: time="2024-10-08T20:03:53.793488251Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:53.794452 containerd[1462]: time="2024-10-08T20:03:53.794373541Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Oct 8 20:03:53.795585 containerd[1462]: time="2024-10-08T20:03:53.795548174Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:53.798766 containerd[1462]: time="2024-10-08T20:03:53.798729310Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:53.800134 containerd[1462]: time="2024-10-08T20:03:53.800086866Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.699638108s" Oct 8 20:03:53.800177 containerd[1462]: time="2024-10-08T20:03:53.800131500Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Oct 8 20:03:53.800808 containerd[1462]: time="2024-10-08T20:03:53.800770197Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 8 20:03:54.792716 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount384564531.mount: Deactivated successfully. Oct 8 20:03:54.800105 containerd[1462]: time="2024-10-08T20:03:54.800044213Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:54.800929 containerd[1462]: time="2024-10-08T20:03:54.800846688Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Oct 8 20:03:54.802123 containerd[1462]: time="2024-10-08T20:03:54.802086914Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:54.805338 containerd[1462]: time="2024-10-08T20:03:54.805265174Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:54.806103 containerd[1462]: time="2024-10-08T20:03:54.806063772Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.005251255s" Oct 8 20:03:54.806103 containerd[1462]: time="2024-10-08T20:03:54.806100951Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Oct 8 20:03:54.806678 containerd[1462]: time="2024-10-08T20:03:54.806645322Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Oct 8 20:03:55.660314 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1116782941.mount: Deactivated successfully. Oct 8 20:03:57.716174 containerd[1462]: time="2024-10-08T20:03:57.716103445Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:57.717069 containerd[1462]: time="2024-10-08T20:03:57.717031736Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56241740" Oct 8 20:03:57.718605 containerd[1462]: time="2024-10-08T20:03:57.718561374Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:57.722598 containerd[1462]: time="2024-10-08T20:03:57.722535677Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:03:57.723752 containerd[1462]: time="2024-10-08T20:03:57.723672399Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.916998313s" Oct 8 20:03:57.723752 containerd[1462]: time="2024-10-08T20:03:57.723735127Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Oct 8 20:03:59.640979 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:03:59.651807 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:03:59.679494 systemd[1]: Reloading requested from client PID 2045 ('systemctl') (unit session-7.scope)... Oct 8 20:03:59.679514 systemd[1]: Reloading... Oct 8 20:03:59.766126 zram_generator::config[2087]: No configuration found. Oct 8 20:04:00.536780 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 8 20:04:00.644649 systemd[1]: Reloading finished in 964 ms. Oct 8 20:04:00.720155 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 8 20:04:00.720259 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 8 20:04:00.720567 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:04:00.730950 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:04:00.879017 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:04:00.893957 (kubelet)[2132]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 8 20:04:00.964583 kubelet[2132]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 8 20:04:00.964583 kubelet[2132]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 8 20:04:00.964583 kubelet[2132]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 8 20:04:00.965659 kubelet[2132]: I1008 20:04:00.965598 2132 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 8 20:04:01.174857 kubelet[2132]: I1008 20:04:01.174748 2132 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Oct 8 20:04:01.174857 kubelet[2132]: I1008 20:04:01.174778 2132 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 8 20:04:01.175136 kubelet[2132]: I1008 20:04:01.175099 2132 server.go:929] "Client rotation is on, will bootstrap in background" Oct 8 20:04:01.200590 kubelet[2132]: E1008 20:04:01.200525 2132 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.126:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:01.200902 kubelet[2132]: I1008 20:04:01.200877 2132 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 8 20:04:01.208846 kubelet[2132]: E1008 20:04:01.208790 2132 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Oct 8 20:04:01.208846 kubelet[2132]: I1008 20:04:01.208829 2132 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Oct 8 20:04:01.222011 kubelet[2132]: I1008 20:04:01.221964 2132 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 8 20:04:01.223249 kubelet[2132]: I1008 20:04:01.223212 2132 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 8 20:04:01.223512 kubelet[2132]: I1008 20:04:01.223454 2132 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 8 20:04:01.223862 kubelet[2132]: I1008 20:04:01.223511 2132 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 8 20:04:01.223954 kubelet[2132]: I1008 20:04:01.223886 2132 topology_manager.go:138] "Creating topology manager with none policy" Oct 8 20:04:01.223975 kubelet[2132]: I1008 20:04:01.223902 2132 container_manager_linux.go:300] "Creating device plugin manager" Oct 8 20:04:01.225648 kubelet[2132]: I1008 20:04:01.224145 2132 state_mem.go:36] "Initialized new in-memory state store" Oct 8 20:04:01.228617 kubelet[2132]: I1008 20:04:01.228582 2132 kubelet.go:408] "Attempting to sync node with API server" Oct 8 20:04:01.228666 kubelet[2132]: I1008 20:04:01.228623 2132 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 8 20:04:01.228710 kubelet[2132]: I1008 20:04:01.228678 2132 kubelet.go:314] "Adding apiserver pod source" Oct 8 20:04:01.228710 kubelet[2132]: I1008 20:04:01.228696 2132 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 8 20:04:01.232915 kubelet[2132]: W1008 20:04:01.232749 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:01.232915 kubelet[2132]: E1008 20:04:01.232856 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:01.232915 kubelet[2132]: W1008 20:04:01.232770 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.126:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:01.232915 kubelet[2132]: E1008 20:04:01.232894 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.126:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:01.234363 kubelet[2132]: I1008 20:04:01.234337 2132 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 8 20:04:01.235872 kubelet[2132]: I1008 20:04:01.235855 2132 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 8 20:04:01.236653 kubelet[2132]: W1008 20:04:01.236627 2132 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 8 20:04:01.237603 kubelet[2132]: I1008 20:04:01.237291 2132 server.go:1269] "Started kubelet" Oct 8 20:04:01.237603 kubelet[2132]: I1008 20:04:01.237357 2132 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 8 20:04:01.238315 kubelet[2132]: I1008 20:04:01.238286 2132 server.go:460] "Adding debug handlers to kubelet server" Oct 8 20:04:01.239014 kubelet[2132]: I1008 20:04:01.238757 2132 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 8 20:04:01.239014 kubelet[2132]: I1008 20:04:01.238965 2132 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 8 20:04:01.239104 kubelet[2132]: I1008 20:04:01.239085 2132 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 8 20:04:01.239148 kubelet[2132]: I1008 20:04:01.239109 2132 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 8 20:04:01.242043 kubelet[2132]: I1008 20:04:01.241444 2132 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 8 20:04:01.242043 kubelet[2132]: I1008 20:04:01.241583 2132 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 8 20:04:01.242043 kubelet[2132]: I1008 20:04:01.241670 2132 reconciler.go:26] "Reconciler: start to sync state" Oct 8 20:04:01.242043 kubelet[2132]: W1008 20:04:01.241815 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.126:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:01.242043 kubelet[2132]: E1008 20:04:01.241857 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.126:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:01.242043 kubelet[2132]: I1008 20:04:01.241972 2132 factory.go:221] Registration of the systemd container factory successfully Oct 8 20:04:01.242043 kubelet[2132]: E1008 20:04:01.242011 2132 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 8 20:04:01.242360 kubelet[2132]: I1008 20:04:01.242088 2132 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 8 20:04:01.242360 kubelet[2132]: E1008 20:04:01.242108 2132 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.126:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.126:6443: connect: connection refused" interval="200ms" Oct 8 20:04:01.243332 kubelet[2132]: E1008 20:04:01.242859 2132 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 8 20:04:01.243332 kubelet[2132]: I1008 20:04:01.243290 2132 factory.go:221] Registration of the containerd container factory successfully Oct 8 20:04:01.278745 kubelet[2132]: E1008 20:04:01.275378 2132 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.126:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.126:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.17fc92de7a64f6e7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2024-10-08 20:04:01.237260007 +0000 UTC m=+0.331327104,LastTimestamp:2024-10-08 20:04:01.237260007 +0000 UTC m=+0.331327104,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Oct 8 20:04:01.292815 kubelet[2132]: I1008 20:04:01.292770 2132 cpu_manager.go:214] "Starting CPU manager" policy="none" Oct 8 20:04:01.292815 kubelet[2132]: I1008 20:04:01.292798 2132 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Oct 8 20:04:01.292815 kubelet[2132]: I1008 20:04:01.292818 2132 state_mem.go:36] "Initialized new in-memory state store" Oct 8 20:04:01.294433 kubelet[2132]: I1008 20:04:01.294369 2132 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 8 20:04:01.295873 kubelet[2132]: I1008 20:04:01.295834 2132 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 8 20:04:01.295935 kubelet[2132]: I1008 20:04:01.295887 2132 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 8 20:04:01.295935 kubelet[2132]: I1008 20:04:01.295913 2132 kubelet.go:2321] "Starting kubelet main sync loop" Oct 8 20:04:01.296111 kubelet[2132]: E1008 20:04:01.295960 2132 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 8 20:04:01.296519 kubelet[2132]: W1008 20:04:01.296482 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.126:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:01.296564 kubelet[2132]: E1008 20:04:01.296532 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.126:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:01.342206 kubelet[2132]: E1008 20:04:01.342140 2132 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 8 20:04:01.396498 kubelet[2132]: E1008 20:04:01.396426 2132 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 8 20:04:01.442933 kubelet[2132]: E1008 20:04:01.442777 2132 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 8 20:04:01.443333 kubelet[2132]: E1008 20:04:01.443277 2132 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.126:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.126:6443: connect: connection refused" interval="400ms" Oct 8 20:04:01.543721 kubelet[2132]: E1008 20:04:01.543641 2132 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 8 20:04:01.596863 kubelet[2132]: E1008 20:04:01.596770 2132 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 8 20:04:01.644118 kubelet[2132]: E1008 20:04:01.644071 2132 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Oct 8 20:04:01.666180 kubelet[2132]: I1008 20:04:01.666151 2132 policy_none.go:49] "None policy: Start" Oct 8 20:04:01.666827 kubelet[2132]: I1008 20:04:01.666811 2132 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 8 20:04:01.666890 kubelet[2132]: I1008 20:04:01.666832 2132 state_mem.go:35] "Initializing new in-memory state store" Oct 8 20:04:01.673137 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Oct 8 20:04:01.689103 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Oct 8 20:04:01.691926 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Oct 8 20:04:01.699380 kubelet[2132]: I1008 20:04:01.699301 2132 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 8 20:04:01.699578 kubelet[2132]: I1008 20:04:01.699543 2132 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 8 20:04:01.699615 kubelet[2132]: I1008 20:04:01.699562 2132 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 8 20:04:01.699827 kubelet[2132]: I1008 20:04:01.699795 2132 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 8 20:04:01.701189 kubelet[2132]: E1008 20:04:01.701165 2132 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Oct 8 20:04:01.801153 kubelet[2132]: I1008 20:04:01.801100 2132 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:01.801648 kubelet[2132]: E1008 20:04:01.801593 2132 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.126:6443/api/v1/nodes\": dial tcp 10.0.0.126:6443: connect: connection refused" node="localhost" Oct 8 20:04:01.844594 kubelet[2132]: E1008 20:04:01.844524 2132 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.126:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.126:6443: connect: connection refused" interval="800ms" Oct 8 20:04:02.002480 kubelet[2132]: I1008 20:04:02.002323 2132 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:02.002917 kubelet[2132]: E1008 20:04:02.002750 2132 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.126:6443/api/v1/nodes\": dial tcp 10.0.0.126:6443: connect: connection refused" node="localhost" Oct 8 20:04:02.005551 systemd[1]: Created slice kubepods-burstable-pod824cd4c200e7e9af876e3a359bfb6982.slice - libcontainer container kubepods-burstable-pod824cd4c200e7e9af876e3a359bfb6982.slice. Oct 8 20:04:02.032408 systemd[1]: Created slice kubepods-burstable-pod344660bab292c4b91cf719f133c08ba2.slice - libcontainer container kubepods-burstable-pod344660bab292c4b91cf719f133c08ba2.slice. Oct 8 20:04:02.036337 systemd[1]: Created slice kubepods-burstable-pod1510be5a54dc8eef4f27b06886c891dc.slice - libcontainer container kubepods-burstable-pod1510be5a54dc8eef4f27b06886c891dc.slice. Oct 8 20:04:02.045179 kubelet[2132]: I1008 20:04:02.045137 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:02.045179 kubelet[2132]: I1008 20:04:02.045172 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:02.045279 kubelet[2132]: I1008 20:04:02.045189 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1510be5a54dc8eef4f27b06886c891dc-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"1510be5a54dc8eef4f27b06886c891dc\") " pod="kube-system/kube-scheduler-localhost" Oct 8 20:04:02.045279 kubelet[2132]: I1008 20:04:02.045205 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:02.045279 kubelet[2132]: I1008 20:04:02.045222 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:02.045279 kubelet[2132]: I1008 20:04:02.045239 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:02.045279 kubelet[2132]: I1008 20:04:02.045261 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:02.045417 kubelet[2132]: I1008 20:04:02.045304 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:02.045417 kubelet[2132]: I1008 20:04:02.045340 2132 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:02.265145 kubelet[2132]: W1008 20:04:02.264982 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.126:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:02.265145 kubelet[2132]: E1008 20:04:02.265058 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.126:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:02.329923 kubelet[2132]: E1008 20:04:02.329857 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:02.330737 containerd[1462]: time="2024-10-08T20:04:02.330691051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:824cd4c200e7e9af876e3a359bfb6982,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:02.334936 kubelet[2132]: E1008 20:04:02.334898 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:02.335405 containerd[1462]: time="2024-10-08T20:04:02.335378791Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:344660bab292c4b91cf719f133c08ba2,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:02.338789 kubelet[2132]: E1008 20:04:02.338765 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:02.339314 containerd[1462]: time="2024-10-08T20:04:02.339271554Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:1510be5a54dc8eef4f27b06886c891dc,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:02.380197 kubelet[2132]: W1008 20:04:02.380111 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.126:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:02.380248 kubelet[2132]: E1008 20:04:02.380204 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.126:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:02.404558 kubelet[2132]: I1008 20:04:02.404525 2132 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:02.404894 kubelet[2132]: E1008 20:04:02.404847 2132 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.126:6443/api/v1/nodes\": dial tcp 10.0.0.126:6443: connect: connection refused" node="localhost" Oct 8 20:04:02.415126 kubelet[2132]: W1008 20:04:02.415072 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:02.415126 kubelet[2132]: E1008 20:04:02.415126 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:02.645035 kubelet[2132]: E1008 20:04:02.644979 2132 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.126:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.126:6443: connect: connection refused" interval="1.6s" Oct 8 20:04:02.704828 kubelet[2132]: W1008 20:04:02.704786 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.126:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:02.704873 kubelet[2132]: E1008 20:04:02.704835 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.126:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:02.908949 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2430147018.mount: Deactivated successfully. Oct 8 20:04:02.918022 containerd[1462]: time="2024-10-08T20:04:02.917964735Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 8 20:04:02.919116 containerd[1462]: time="2024-10-08T20:04:02.919084378Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 8 20:04:02.919677 containerd[1462]: time="2024-10-08T20:04:02.919580812Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Oct 8 20:04:02.920691 containerd[1462]: time="2024-10-08T20:04:02.920655928Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 8 20:04:02.921534 containerd[1462]: time="2024-10-08T20:04:02.921494119Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 8 20:04:02.922606 containerd[1462]: time="2024-10-08T20:04:02.922557533Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 8 20:04:02.923492 containerd[1462]: time="2024-10-08T20:04:02.923399431Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 8 20:04:02.925438 containerd[1462]: time="2024-10-08T20:04:02.925401730Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 8 20:04:02.929252 containerd[1462]: time="2024-10-08T20:04:02.929212404Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 598.428725ms" Oct 8 20:04:02.930968 containerd[1462]: time="2024-10-08T20:04:02.930909085Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 595.468455ms" Oct 8 20:04:02.934333 containerd[1462]: time="2024-10-08T20:04:02.934281597Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 594.9246ms" Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.189786048Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.189939201Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.189980079Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.190000118Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.190066185Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:03.190182 containerd[1462]: time="2024-10-08T20:04:03.190086414Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.190452 containerd[1462]: time="2024-10-08T20:04:03.190177909Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.190452 containerd[1462]: time="2024-10-08T20:04:03.190174834Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.226302 containerd[1462]: time="2024-10-08T20:04:03.198437410Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:03.226302 containerd[1462]: time="2024-10-08T20:04:03.198592177Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:03.226302 containerd[1462]: time="2024-10-08T20:04:03.198646401Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.226517 kubelet[2132]: I1008 20:04:03.207013 2132 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:03.226517 kubelet[2132]: E1008 20:04:03.207632 2132 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.126:6443/api/v1/nodes\": dial tcp 10.0.0.126:6443: connect: connection refused" node="localhost" Oct 8 20:04:03.235994 containerd[1462]: time="2024-10-08T20:04:03.235158962Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:03.264905 systemd[1]: Started cri-containerd-349411960e9bd4c6f813fb622ec8b4926cd370847646f3fc46995708cab649af.scope - libcontainer container 349411960e9bd4c6f813fb622ec8b4926cd370847646f3fc46995708cab649af. Oct 8 20:04:03.266101 kubelet[2132]: E1008 20:04:03.266013 2132 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.126:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:03.268523 systemd[1]: Started cri-containerd-e9d26b07e6c310713abca5ccc5acdaf3f239456b8ecd19dc7636d01775fb9821.scope - libcontainer container e9d26b07e6c310713abca5ccc5acdaf3f239456b8ecd19dc7636d01775fb9821. Oct 8 20:04:03.282975 systemd[1]: Started cri-containerd-4478fa637050ef53b22c7d5c03422cafb799ca8297d7990f070d4a342addf55a.scope - libcontainer container 4478fa637050ef53b22c7d5c03422cafb799ca8297d7990f070d4a342addf55a. Oct 8 20:04:03.335321 containerd[1462]: time="2024-10-08T20:04:03.335239588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:344660bab292c4b91cf719f133c08ba2,Namespace:kube-system,Attempt:0,} returns sandbox id \"e9d26b07e6c310713abca5ccc5acdaf3f239456b8ecd19dc7636d01775fb9821\"" Oct 8 20:04:03.337261 kubelet[2132]: E1008 20:04:03.336921 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:03.337519 containerd[1462]: time="2024-10-08T20:04:03.337440262Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:824cd4c200e7e9af876e3a359bfb6982,Namespace:kube-system,Attempt:0,} returns sandbox id \"349411960e9bd4c6f813fb622ec8b4926cd370847646f3fc46995708cab649af\"" Oct 8 20:04:03.338333 kubelet[2132]: E1008 20:04:03.338303 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:03.339931 containerd[1462]: time="2024-10-08T20:04:03.339896906Z" level=info msg="CreateContainer within sandbox \"e9d26b07e6c310713abca5ccc5acdaf3f239456b8ecd19dc7636d01775fb9821\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 8 20:04:03.340244 containerd[1462]: time="2024-10-08T20:04:03.340219826Z" level=info msg="CreateContainer within sandbox \"349411960e9bd4c6f813fb622ec8b4926cd370847646f3fc46995708cab649af\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 8 20:04:03.345336 containerd[1462]: time="2024-10-08T20:04:03.345297811Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:1510be5a54dc8eef4f27b06886c891dc,Namespace:kube-system,Attempt:0,} returns sandbox id \"4478fa637050ef53b22c7d5c03422cafb799ca8297d7990f070d4a342addf55a\"" Oct 8 20:04:03.346352 kubelet[2132]: E1008 20:04:03.346324 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:03.348124 containerd[1462]: time="2024-10-08T20:04:03.348063628Z" level=info msg="CreateContainer within sandbox \"4478fa637050ef53b22c7d5c03422cafb799ca8297d7990f070d4a342addf55a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 8 20:04:03.449625 kubelet[2132]: E1008 20:04:03.449374 2132 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.126:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.126:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.17fc92de7a64f6e7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2024-10-08 20:04:01.237260007 +0000 UTC m=+0.331327104,LastTimestamp:2024-10-08 20:04:01.237260007 +0000 UTC m=+0.331327104,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Oct 8 20:04:04.098613 containerd[1462]: time="2024-10-08T20:04:04.098554751Z" level=info msg="CreateContainer within sandbox \"e9d26b07e6c310713abca5ccc5acdaf3f239456b8ecd19dc7636d01775fb9821\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"82e6d11df34f6b408a464f9d175bba47107b1a23dcc7bb802383307f6cfc4027\"" Oct 8 20:04:04.099379 containerd[1462]: time="2024-10-08T20:04:04.099345205Z" level=info msg="StartContainer for \"82e6d11df34f6b408a464f9d175bba47107b1a23dcc7bb802383307f6cfc4027\"" Oct 8 20:04:04.101957 kubelet[2132]: W1008 20:04:04.101907 2132 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.126:6443: connect: connection refused Oct 8 20:04:04.101957 kubelet[2132]: E1008 20:04:04.101956 2132 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.126:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.126:6443: connect: connection refused" logger="UnhandledError" Oct 8 20:04:04.102718 containerd[1462]: time="2024-10-08T20:04:04.102654617Z" level=info msg="CreateContainer within sandbox \"349411960e9bd4c6f813fb622ec8b4926cd370847646f3fc46995708cab649af\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9fd9aff74ad9f17e44a35ea1364fa976e2db64f5adfd094a16239b279d0e6dab\"" Oct 8 20:04:04.103157 containerd[1462]: time="2024-10-08T20:04:04.103125269Z" level=info msg="StartContainer for \"9fd9aff74ad9f17e44a35ea1364fa976e2db64f5adfd094a16239b279d0e6dab\"" Oct 8 20:04:04.104856 containerd[1462]: time="2024-10-08T20:04:04.104814887Z" level=info msg="CreateContainer within sandbox \"4478fa637050ef53b22c7d5c03422cafb799ca8297d7990f070d4a342addf55a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3705c2b49cebf5b1859c8f851857af1d9d36922ec944470a739d14b491d76c6f\"" Oct 8 20:04:04.105491 containerd[1462]: time="2024-10-08T20:04:04.105178304Z" level=info msg="StartContainer for \"3705c2b49cebf5b1859c8f851857af1d9d36922ec944470a739d14b491d76c6f\"" Oct 8 20:04:04.133692 systemd[1]: Started cri-containerd-82e6d11df34f6b408a464f9d175bba47107b1a23dcc7bb802383307f6cfc4027.scope - libcontainer container 82e6d11df34f6b408a464f9d175bba47107b1a23dcc7bb802383307f6cfc4027. Oct 8 20:04:04.138363 systemd[1]: Started cri-containerd-3705c2b49cebf5b1859c8f851857af1d9d36922ec944470a739d14b491d76c6f.scope - libcontainer container 3705c2b49cebf5b1859c8f851857af1d9d36922ec944470a739d14b491d76c6f. Oct 8 20:04:04.139971 systemd[1]: Started cri-containerd-9fd9aff74ad9f17e44a35ea1364fa976e2db64f5adfd094a16239b279d0e6dab.scope - libcontainer container 9fd9aff74ad9f17e44a35ea1364fa976e2db64f5adfd094a16239b279d0e6dab. Oct 8 20:04:04.203648 containerd[1462]: time="2024-10-08T20:04:04.198002509Z" level=info msg="StartContainer for \"82e6d11df34f6b408a464f9d175bba47107b1a23dcc7bb802383307f6cfc4027\" returns successfully" Oct 8 20:04:04.204952 containerd[1462]: time="2024-10-08T20:04:04.204913012Z" level=info msg="StartContainer for \"3705c2b49cebf5b1859c8f851857af1d9d36922ec944470a739d14b491d76c6f\" returns successfully" Oct 8 20:04:04.226636 containerd[1462]: time="2024-10-08T20:04:04.226602188Z" level=info msg="StartContainer for \"9fd9aff74ad9f17e44a35ea1364fa976e2db64f5adfd094a16239b279d0e6dab\" returns successfully" Oct 8 20:04:04.246196 kubelet[2132]: E1008 20:04:04.245972 2132 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.126:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.126:6443: connect: connection refused" interval="3.2s" Oct 8 20:04:04.304099 kubelet[2132]: E1008 20:04:04.304029 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:04.306880 kubelet[2132]: E1008 20:04:04.306848 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:04.309032 kubelet[2132]: E1008 20:04:04.308961 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:04.810224 kubelet[2132]: I1008 20:04:04.810186 2132 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:05.312767 kubelet[2132]: E1008 20:04:05.312741 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:05.628530 kubelet[2132]: I1008 20:04:05.628482 2132 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Oct 8 20:04:05.628530 kubelet[2132]: E1008 20:04:05.628523 2132 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Oct 8 20:04:06.233876 kubelet[2132]: I1008 20:04:06.233837 2132 apiserver.go:52] "Watching apiserver" Oct 8 20:04:06.242216 kubelet[2132]: I1008 20:04:06.242174 2132 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 8 20:04:06.918545 kubelet[2132]: E1008 20:04:06.918514 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:07.314991 kubelet[2132]: E1008 20:04:07.314849 2132 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:07.823306 systemd[1]: Reloading requested from client PID 2412 ('systemctl') (unit session-7.scope)... Oct 8 20:04:07.823320 systemd[1]: Reloading... Oct 8 20:04:07.897500 zram_generator::config[2451]: No configuration found. Oct 8 20:04:08.011214 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 8 20:04:08.101081 systemd[1]: Reloading finished in 277 ms. Oct 8 20:04:08.148695 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:04:08.172954 systemd[1]: kubelet.service: Deactivated successfully. Oct 8 20:04:08.173280 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:04:08.184767 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 8 20:04:08.329755 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 8 20:04:08.335197 (kubelet)[2496]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 8 20:04:08.380033 kubelet[2496]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 8 20:04:08.380033 kubelet[2496]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 8 20:04:08.380033 kubelet[2496]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 8 20:04:08.380456 kubelet[2496]: I1008 20:04:08.380028 2496 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 8 20:04:08.387178 kubelet[2496]: I1008 20:04:08.387139 2496 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Oct 8 20:04:08.387178 kubelet[2496]: I1008 20:04:08.387169 2496 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 8 20:04:08.388025 kubelet[2496]: I1008 20:04:08.387993 2496 server.go:929] "Client rotation is on, will bootstrap in background" Oct 8 20:04:08.389226 kubelet[2496]: I1008 20:04:08.389198 2496 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 8 20:04:08.391035 kubelet[2496]: I1008 20:04:08.390992 2496 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 8 20:04:08.394403 kubelet[2496]: E1008 20:04:08.394369 2496 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Oct 8 20:04:08.394403 kubelet[2496]: I1008 20:04:08.394400 2496 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Oct 8 20:04:08.400579 kubelet[2496]: I1008 20:04:08.400534 2496 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 8 20:04:08.400715 kubelet[2496]: I1008 20:04:08.400682 2496 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 8 20:04:08.400901 kubelet[2496]: I1008 20:04:08.400850 2496 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 8 20:04:08.401089 kubelet[2496]: I1008 20:04:08.400886 2496 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 8 20:04:08.401089 kubelet[2496]: I1008 20:04:08.401079 2496 topology_manager.go:138] "Creating topology manager with none policy" Oct 8 20:04:08.401089 kubelet[2496]: I1008 20:04:08.401088 2496 container_manager_linux.go:300] "Creating device plugin manager" Oct 8 20:04:08.401358 kubelet[2496]: I1008 20:04:08.401138 2496 state_mem.go:36] "Initialized new in-memory state store" Oct 8 20:04:08.401358 kubelet[2496]: I1008 20:04:08.401258 2496 kubelet.go:408] "Attempting to sync node with API server" Oct 8 20:04:08.401358 kubelet[2496]: I1008 20:04:08.401272 2496 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 8 20:04:08.401358 kubelet[2496]: I1008 20:04:08.401297 2496 kubelet.go:314] "Adding apiserver pod source" Oct 8 20:04:08.401358 kubelet[2496]: I1008 20:04:08.401318 2496 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 8 20:04:08.402332 kubelet[2496]: I1008 20:04:08.402263 2496 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 8 20:04:08.403846 kubelet[2496]: I1008 20:04:08.403791 2496 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 8 20:04:08.406329 kubelet[2496]: I1008 20:04:08.406315 2496 server.go:1269] "Started kubelet" Oct 8 20:04:08.407155 kubelet[2496]: I1008 20:04:08.407129 2496 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 8 20:04:08.408452 kubelet[2496]: I1008 20:04:08.407906 2496 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 8 20:04:08.408452 kubelet[2496]: I1008 20:04:08.408358 2496 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 8 20:04:08.411052 kubelet[2496]: I1008 20:04:08.410507 2496 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 8 20:04:08.411243 kubelet[2496]: I1008 20:04:08.411140 2496 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 8 20:04:08.412183 kubelet[2496]: I1008 20:04:08.411893 2496 server.go:460] "Adding debug handlers to kubelet server" Oct 8 20:04:08.413653 kubelet[2496]: I1008 20:04:08.413628 2496 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 8 20:04:08.413778 kubelet[2496]: I1008 20:04:08.413749 2496 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 8 20:04:08.414231 kubelet[2496]: I1008 20:04:08.414194 2496 reconciler.go:26] "Reconciler: start to sync state" Oct 8 20:04:08.414621 kubelet[2496]: I1008 20:04:08.414371 2496 factory.go:221] Registration of the systemd container factory successfully Oct 8 20:04:08.414621 kubelet[2496]: I1008 20:04:08.414543 2496 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 8 20:04:08.416303 kubelet[2496]: E1008 20:04:08.416266 2496 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 8 20:04:08.417392 kubelet[2496]: I1008 20:04:08.417357 2496 factory.go:221] Registration of the containerd container factory successfully Oct 8 20:04:08.425349 kubelet[2496]: I1008 20:04:08.425304 2496 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 8 20:04:08.426934 kubelet[2496]: I1008 20:04:08.426894 2496 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 8 20:04:08.426982 kubelet[2496]: I1008 20:04:08.426944 2496 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 8 20:04:08.426982 kubelet[2496]: I1008 20:04:08.426971 2496 kubelet.go:2321] "Starting kubelet main sync loop" Oct 8 20:04:08.427035 kubelet[2496]: E1008 20:04:08.427018 2496 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 8 20:04:08.460180 kubelet[2496]: I1008 20:04:08.460126 2496 cpu_manager.go:214] "Starting CPU manager" policy="none" Oct 8 20:04:08.460180 kubelet[2496]: I1008 20:04:08.460146 2496 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Oct 8 20:04:08.460180 kubelet[2496]: I1008 20:04:08.460165 2496 state_mem.go:36] "Initialized new in-memory state store" Oct 8 20:04:08.460356 kubelet[2496]: I1008 20:04:08.460307 2496 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 8 20:04:08.460356 kubelet[2496]: I1008 20:04:08.460317 2496 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 8 20:04:08.460356 kubelet[2496]: I1008 20:04:08.460337 2496 policy_none.go:49] "None policy: Start" Oct 8 20:04:08.461042 kubelet[2496]: I1008 20:04:08.461005 2496 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 8 20:04:08.461042 kubelet[2496]: I1008 20:04:08.461042 2496 state_mem.go:35] "Initializing new in-memory state store" Oct 8 20:04:08.461289 kubelet[2496]: I1008 20:04:08.461266 2496 state_mem.go:75] "Updated machine memory state" Oct 8 20:04:08.465839 kubelet[2496]: I1008 20:04:08.465802 2496 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 8 20:04:08.466122 kubelet[2496]: I1008 20:04:08.465981 2496 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 8 20:04:08.466122 kubelet[2496]: I1008 20:04:08.465991 2496 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 8 20:04:08.466366 kubelet[2496]: I1008 20:04:08.466239 2496 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 8 20:04:08.534252 kubelet[2496]: E1008 20:04:08.534201 2496 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Oct 8 20:04:08.575054 kubelet[2496]: I1008 20:04:08.574999 2496 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Oct 8 20:04:08.580316 kubelet[2496]: I1008 20:04:08.580279 2496 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Oct 8 20:04:08.580682 kubelet[2496]: I1008 20:04:08.580646 2496 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Oct 8 20:04:08.715990 kubelet[2496]: I1008 20:04:08.715867 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:08.715990 kubelet[2496]: I1008 20:04:08.715899 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:08.715990 kubelet[2496]: I1008 20:04:08.715917 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:08.715990 kubelet[2496]: I1008 20:04:08.715933 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/824cd4c200e7e9af876e3a359bfb6982-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"824cd4c200e7e9af876e3a359bfb6982\") " pod="kube-system/kube-apiserver-localhost" Oct 8 20:04:08.715990 kubelet[2496]: I1008 20:04:08.715948 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:08.716224 kubelet[2496]: I1008 20:04:08.716017 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:08.716224 kubelet[2496]: I1008 20:04:08.716079 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:08.716224 kubelet[2496]: I1008 20:04:08.716104 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/344660bab292c4b91cf719f133c08ba2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"344660bab292c4b91cf719f133c08ba2\") " pod="kube-system/kube-controller-manager-localhost" Oct 8 20:04:08.716224 kubelet[2496]: I1008 20:04:08.716168 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/1510be5a54dc8eef4f27b06886c891dc-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"1510be5a54dc8eef4f27b06886c891dc\") " pod="kube-system/kube-scheduler-localhost" Oct 8 20:04:08.820367 sudo[2531]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Oct 8 20:04:08.820779 sudo[2531]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Oct 8 20:04:08.834554 kubelet[2496]: E1008 20:04:08.834521 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:08.834631 kubelet[2496]: E1008 20:04:08.834604 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:08.834631 kubelet[2496]: E1008 20:04:08.834619 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:09.402383 kubelet[2496]: I1008 20:04:09.402332 2496 apiserver.go:52] "Watching apiserver" Oct 8 20:04:09.413974 kubelet[2496]: I1008 20:04:09.413921 2496 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 8 20:04:09.439351 kubelet[2496]: E1008 20:04:09.439310 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:09.440025 kubelet[2496]: E1008 20:04:09.439996 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:09.440277 kubelet[2496]: E1008 20:04:09.440259 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:09.459107 sudo[2531]: pam_unix(sudo:session): session closed for user root Oct 8 20:04:09.557243 kubelet[2496]: I1008 20:04:09.557178 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.557157649 podStartE2EDuration="1.557157649s" podCreationTimestamp="2024-10-08 20:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:09.523317051 +0000 UTC m=+1.184541005" watchObservedRunningTime="2024-10-08 20:04:09.557157649 +0000 UTC m=+1.218381603" Oct 8 20:04:09.557400 kubelet[2496]: I1008 20:04:09.557308 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.557304428 podStartE2EDuration="3.557304428s" podCreationTimestamp="2024-10-08 20:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:09.556944172 +0000 UTC m=+1.218168126" watchObservedRunningTime="2024-10-08 20:04:09.557304428 +0000 UTC m=+1.218528382" Oct 8 20:04:09.566760 kubelet[2496]: I1008 20:04:09.566690 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.566674589 podStartE2EDuration="1.566674589s" podCreationTimestamp="2024-10-08 20:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:09.565120148 +0000 UTC m=+1.226344112" watchObservedRunningTime="2024-10-08 20:04:09.566674589 +0000 UTC m=+1.227898543" Oct 8 20:04:10.440815 kubelet[2496]: E1008 20:04:10.440780 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:10.440815 kubelet[2496]: E1008 20:04:10.440785 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:10.441215 kubelet[2496]: E1008 20:04:10.441066 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:10.978155 sudo[1644]: pam_unix(sudo:session): session closed for user root Oct 8 20:04:10.979827 sshd[1641]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:10.984106 systemd[1]: sshd@6-10.0.0.126:22-10.0.0.1:54346.service: Deactivated successfully. Oct 8 20:04:10.986142 systemd[1]: session-7.scope: Deactivated successfully. Oct 8 20:04:10.986330 systemd[1]: session-7.scope: Consumed 5.273s CPU time, 154.4M memory peak, 0B memory swap peak. Oct 8 20:04:10.986813 systemd-logind[1448]: Session 7 logged out. Waiting for processes to exit. Oct 8 20:04:10.987630 systemd-logind[1448]: Removed session 7. Oct 8 20:04:13.298059 kubelet[2496]: I1008 20:04:13.298024 2496 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 8 20:04:13.298542 containerd[1462]: time="2024-10-08T20:04:13.298394039Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 8 20:04:13.298794 kubelet[2496]: I1008 20:04:13.298608 2496 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 8 20:04:14.074786 update_engine[1449]: I20241008 20:04:14.074690 1449 update_attempter.cc:509] Updating boot flags... Oct 8 20:04:14.316492 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2578) Oct 8 20:04:14.380525 systemd[1]: Created slice kubepods-besteffort-podff3faa2a_52a1_4b7f_9242_e936d3843af7.slice - libcontainer container kubepods-besteffort-podff3faa2a_52a1_4b7f_9242_e936d3843af7.slice. Oct 8 20:04:14.412493 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2580) Oct 8 20:04:14.412831 systemd[1]: Created slice kubepods-burstable-pod798ba836_9610_40d9_b6b9_ba0cbd07a5ed.slice - libcontainer container kubepods-burstable-pod798ba836_9610_40d9_b6b9_ba0cbd07a5ed.slice. Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439623 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hostproc\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439663 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-etc-cni-netd\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439681 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-lib-modules\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439701 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ff3faa2a-52a1-4b7f-9242-e936d3843af7-xtables-lock\") pod \"kube-proxy-qdkg2\" (UID: \"ff3faa2a-52a1-4b7f-9242-e936d3843af7\") " pod="kube-system/kube-proxy-qdkg2" Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439717 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ff3faa2a-52a1-4b7f-9242-e936d3843af7-lib-modules\") pod \"kube-proxy-qdkg2\" (UID: \"ff3faa2a-52a1-4b7f-9242-e936d3843af7\") " pod="kube-system/kube-proxy-qdkg2" Oct 8 20:04:14.439790 kubelet[2496]: I1008 20:04:14.439743 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8j8\" (UniqueName: \"kubernetes.io/projected/ff3faa2a-52a1-4b7f-9242-e936d3843af7-kube-api-access-rz8j8\") pod \"kube-proxy-qdkg2\" (UID: \"ff3faa2a-52a1-4b7f-9242-e936d3843af7\") " pod="kube-system/kube-proxy-qdkg2" Oct 8 20:04:14.440423 kubelet[2496]: I1008 20:04:14.439762 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2wrj\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-kube-api-access-k2wrj\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440423 kubelet[2496]: I1008 20:04:14.439780 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-xtables-lock\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440423 kubelet[2496]: I1008 20:04:14.439796 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/ff3faa2a-52a1-4b7f-9242-e936d3843af7-kube-proxy\") pod \"kube-proxy-qdkg2\" (UID: \"ff3faa2a-52a1-4b7f-9242-e936d3843af7\") " pod="kube-system/kube-proxy-qdkg2" Oct 8 20:04:14.440423 kubelet[2496]: I1008 20:04:14.439811 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-config-path\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440423 kubelet[2496]: I1008 20:04:14.439829 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-kernel\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439854 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-bpf-maps\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439870 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-cgroup\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439884 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cni-path\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439901 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hubble-tls\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439917 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-run\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440604 kubelet[2496]: I1008 20:04:14.439934 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-clustermesh-secrets\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.440788 kubelet[2496]: I1008 20:04:14.439952 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-net\") pod \"cilium-nbqd6\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " pod="kube-system/cilium-nbqd6" Oct 8 20:04:14.465984 systemd[1]: Created slice kubepods-besteffort-pod37102e20_313b_42f0_bbca_a02d167c080b.slice - libcontainer container kubepods-besteffort-pod37102e20_313b_42f0_bbca_a02d167c080b.slice. Oct 8 20:04:14.541089 kubelet[2496]: I1008 20:04:14.541034 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/37102e20-313b-42f0-bbca-a02d167c080b-cilium-config-path\") pod \"cilium-operator-5d85765b45-vh7nj\" (UID: \"37102e20-313b-42f0-bbca-a02d167c080b\") " pod="kube-system/cilium-operator-5d85765b45-vh7nj" Oct 8 20:04:14.541543 kubelet[2496]: I1008 20:04:14.541290 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfkhr\" (UniqueName: \"kubernetes.io/projected/37102e20-313b-42f0-bbca-a02d167c080b-kube-api-access-hfkhr\") pod \"cilium-operator-5d85765b45-vh7nj\" (UID: \"37102e20-313b-42f0-bbca-a02d167c080b\") " pod="kube-system/cilium-operator-5d85765b45-vh7nj" Oct 8 20:04:14.699633 kubelet[2496]: E1008 20:04:14.699458 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:14.700231 containerd[1462]: time="2024-10-08T20:04:14.700156336Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qdkg2,Uid:ff3faa2a-52a1-4b7f-9242-e936d3843af7,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:14.719109 kubelet[2496]: E1008 20:04:14.719058 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:14.719791 containerd[1462]: time="2024-10-08T20:04:14.719545936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nbqd6,Uid:798ba836-9610-40d9-b6b9-ba0cbd07a5ed,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:14.728817 containerd[1462]: time="2024-10-08T20:04:14.727994822Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:14.728817 containerd[1462]: time="2024-10-08T20:04:14.728774571Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:14.728817 containerd[1462]: time="2024-10-08T20:04:14.728794058Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.729023 containerd[1462]: time="2024-10-08T20:04:14.728920307Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.744601 containerd[1462]: time="2024-10-08T20:04:14.744483760Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:14.744601 containerd[1462]: time="2024-10-08T20:04:14.744553151Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:14.744601 containerd[1462]: time="2024-10-08T20:04:14.744567618Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.744834 containerd[1462]: time="2024-10-08T20:04:14.744648913Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.751692 systemd[1]: Started cri-containerd-ee47c76156d3ac98de74c2a8092cf8893eb2e46a62c6a4f6e48ae7754d7f6e47.scope - libcontainer container ee47c76156d3ac98de74c2a8092cf8893eb2e46a62c6a4f6e48ae7754d7f6e47. Oct 8 20:04:14.760668 systemd[1]: Started cri-containerd-f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8.scope - libcontainer container f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8. Oct 8 20:04:14.770610 kubelet[2496]: E1008 20:04:14.770109 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:14.771570 containerd[1462]: time="2024-10-08T20:04:14.771531035Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-vh7nj,Uid:37102e20-313b-42f0-bbca-a02d167c080b,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:14.779540 containerd[1462]: time="2024-10-08T20:04:14.779482848Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qdkg2,Uid:ff3faa2a-52a1-4b7f-9242-e936d3843af7,Namespace:kube-system,Attempt:0,} returns sandbox id \"ee47c76156d3ac98de74c2a8092cf8893eb2e46a62c6a4f6e48ae7754d7f6e47\"" Oct 8 20:04:14.780384 kubelet[2496]: E1008 20:04:14.780350 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:14.783942 containerd[1462]: time="2024-10-08T20:04:14.783866924Z" level=info msg="CreateContainer within sandbox \"ee47c76156d3ac98de74c2a8092cf8893eb2e46a62c6a4f6e48ae7754d7f6e47\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 8 20:04:14.792512 containerd[1462]: time="2024-10-08T20:04:14.790759488Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nbqd6,Uid:798ba836-9610-40d9-b6b9-ba0cbd07a5ed,Namespace:kube-system,Attempt:0,} returns sandbox id \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\"" Oct 8 20:04:14.792650 kubelet[2496]: E1008 20:04:14.791794 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:14.794254 containerd[1462]: time="2024-10-08T20:04:14.794190437Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Oct 8 20:04:14.808315 containerd[1462]: time="2024-10-08T20:04:14.808258271Z" level=info msg="CreateContainer within sandbox \"ee47c76156d3ac98de74c2a8092cf8893eb2e46a62c6a4f6e48ae7754d7f6e47\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9316fc1ad178c26c0d2a10dff93c6aa775a17285c668f6e3f2c4d4e7cbc7ad19\"" Oct 8 20:04:14.808513 containerd[1462]: time="2024-10-08T20:04:14.808248413Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:14.808513 containerd[1462]: time="2024-10-08T20:04:14.808313806Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:14.808513 containerd[1462]: time="2024-10-08T20:04:14.808329196Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.808636 containerd[1462]: time="2024-10-08T20:04:14.808424948Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:14.809077 containerd[1462]: time="2024-10-08T20:04:14.809040225Z" level=info msg="StartContainer for \"9316fc1ad178c26c0d2a10dff93c6aa775a17285c668f6e3f2c4d4e7cbc7ad19\"" Oct 8 20:04:14.831652 systemd[1]: Started cri-containerd-4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3.scope - libcontainer container 4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3. Oct 8 20:04:14.834843 systemd[1]: Started cri-containerd-9316fc1ad178c26c0d2a10dff93c6aa775a17285c668f6e3f2c4d4e7cbc7ad19.scope - libcontainer container 9316fc1ad178c26c0d2a10dff93c6aa775a17285c668f6e3f2c4d4e7cbc7ad19. Oct 8 20:04:14.866857 containerd[1462]: time="2024-10-08T20:04:14.866794668Z" level=info msg="StartContainer for \"9316fc1ad178c26c0d2a10dff93c6aa775a17285c668f6e3f2c4d4e7cbc7ad19\" returns successfully" Oct 8 20:04:14.880913 containerd[1462]: time="2024-10-08T20:04:14.880833657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-vh7nj,Uid:37102e20-313b-42f0-bbca-a02d167c080b,Namespace:kube-system,Attempt:0,} returns sandbox id \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\"" Oct 8 20:04:14.881749 kubelet[2496]: E1008 20:04:14.881691 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:15.454608 kubelet[2496]: E1008 20:04:15.454569 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:15.465973 kubelet[2496]: I1008 20:04:15.465888 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-qdkg2" podStartSLOduration=1.465870281 podStartE2EDuration="1.465870281s" podCreationTimestamp="2024-10-08 20:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:15.4656588 +0000 UTC m=+7.126882764" watchObservedRunningTime="2024-10-08 20:04:15.465870281 +0000 UTC m=+7.127094235" Oct 8 20:04:17.001240 kubelet[2496]: E1008 20:04:17.001197 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:17.458934 kubelet[2496]: E1008 20:04:17.458889 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:18.460547 kubelet[2496]: E1008 20:04:18.460496 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:19.521953 kubelet[2496]: E1008 20:04:19.521902 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:20.276816 kubelet[2496]: E1008 20:04:20.276774 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:21.575669 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1566276999.mount: Deactivated successfully. Oct 8 20:04:24.557840 containerd[1462]: time="2024-10-08T20:04:24.557781231Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:04:24.558589 containerd[1462]: time="2024-10-08T20:04:24.558517119Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735327" Oct 8 20:04:24.559743 containerd[1462]: time="2024-10-08T20:04:24.559707566Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:04:24.561259 containerd[1462]: time="2024-10-08T20:04:24.561225089Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 9.766964109s" Oct 8 20:04:24.561302 containerd[1462]: time="2024-10-08T20:04:24.561257279Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Oct 8 20:04:24.566242 containerd[1462]: time="2024-10-08T20:04:24.566199413Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Oct 8 20:04:24.576821 containerd[1462]: time="2024-10-08T20:04:24.576779915Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Oct 8 20:04:24.590204 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount160540379.mount: Deactivated successfully. Oct 8 20:04:24.591564 containerd[1462]: time="2024-10-08T20:04:24.591536375Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\"" Oct 8 20:04:24.594039 containerd[1462]: time="2024-10-08T20:04:24.594019640Z" level=info msg="StartContainer for \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\"" Oct 8 20:04:24.635719 systemd[1]: Started cri-containerd-649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48.scope - libcontainer container 649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48. Oct 8 20:04:24.664122 containerd[1462]: time="2024-10-08T20:04:24.664077493Z" level=info msg="StartContainer for \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\" returns successfully" Oct 8 20:04:24.674936 systemd[1]: cri-containerd-649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48.scope: Deactivated successfully. Oct 8 20:04:24.870437 containerd[1462]: time="2024-10-08T20:04:24.867207446Z" level=info msg="shim disconnected" id=649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48 namespace=k8s.io Oct 8 20:04:24.870437 containerd[1462]: time="2024-10-08T20:04:24.870439534Z" level=warning msg="cleaning up after shim disconnected" id=649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48 namespace=k8s.io Oct 8 20:04:24.870437 containerd[1462]: time="2024-10-08T20:04:24.870485911Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:04:25.476270 kubelet[2496]: E1008 20:04:25.476221 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:25.479044 containerd[1462]: time="2024-10-08T20:04:25.479004830Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Oct 8 20:04:25.502073 containerd[1462]: time="2024-10-08T20:04:25.501997153Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\"" Oct 8 20:04:25.502727 containerd[1462]: time="2024-10-08T20:04:25.502682405Z" level=info msg="StartContainer for \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\"" Oct 8 20:04:25.533623 systemd[1]: Started cri-containerd-7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad.scope - libcontainer container 7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad. Oct 8 20:04:25.558836 containerd[1462]: time="2024-10-08T20:04:25.558793016Z" level=info msg="StartContainer for \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\" returns successfully" Oct 8 20:04:25.570909 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 8 20:04:25.571138 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 8 20:04:25.571212 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Oct 8 20:04:25.580013 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 8 20:04:25.580308 systemd[1]: cri-containerd-7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad.scope: Deactivated successfully. Oct 8 20:04:25.588990 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48-rootfs.mount: Deactivated successfully. Oct 8 20:04:25.594434 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad-rootfs.mount: Deactivated successfully. Oct 8 20:04:25.600347 containerd[1462]: time="2024-10-08T20:04:25.600269688Z" level=info msg="shim disconnected" id=7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad namespace=k8s.io Oct 8 20:04:25.600448 containerd[1462]: time="2024-10-08T20:04:25.600344359Z" level=warning msg="cleaning up after shim disconnected" id=7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad namespace=k8s.io Oct 8 20:04:25.600448 containerd[1462]: time="2024-10-08T20:04:25.600357504Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:04:25.608848 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 8 20:04:26.481011 kubelet[2496]: E1008 20:04:26.480717 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:26.487206 containerd[1462]: time="2024-10-08T20:04:26.487053513Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Oct 8 20:04:26.514781 containerd[1462]: time="2024-10-08T20:04:26.514716475Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\"" Oct 8 20:04:26.515519 containerd[1462]: time="2024-10-08T20:04:26.515463924Z" level=info msg="StartContainer for \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\"" Oct 8 20:04:26.564820 systemd[1]: Started cri-containerd-7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24.scope - libcontainer container 7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24. Oct 8 20:04:26.644020 containerd[1462]: time="2024-10-08T20:04:26.643823398Z" level=info msg="StartContainer for \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\" returns successfully" Oct 8 20:04:26.644195 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2256754486.mount: Deactivated successfully. Oct 8 20:04:26.645104 systemd[1]: cri-containerd-7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24.scope: Deactivated successfully. Oct 8 20:04:26.669119 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24-rootfs.mount: Deactivated successfully. Oct 8 20:04:26.681210 containerd[1462]: time="2024-10-08T20:04:26.681141692Z" level=info msg="shim disconnected" id=7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24 namespace=k8s.io Oct 8 20:04:26.681210 containerd[1462]: time="2024-10-08T20:04:26.681194492Z" level=warning msg="cleaning up after shim disconnected" id=7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24 namespace=k8s.io Oct 8 20:04:26.681210 containerd[1462]: time="2024-10-08T20:04:26.681202657Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:04:26.959193 containerd[1462]: time="2024-10-08T20:04:26.959140917Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:04:26.959831 containerd[1462]: time="2024-10-08T20:04:26.959763610Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18907149" Oct 8 20:04:26.960812 containerd[1462]: time="2024-10-08T20:04:26.960774867Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 8 20:04:26.962128 containerd[1462]: time="2024-10-08T20:04:26.962090567Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.395857061s" Oct 8 20:04:26.962128 containerd[1462]: time="2024-10-08T20:04:26.962120645Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Oct 8 20:04:26.963998 containerd[1462]: time="2024-10-08T20:04:26.963962326Z" level=info msg="CreateContainer within sandbox \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Oct 8 20:04:26.977990 containerd[1462]: time="2024-10-08T20:04:26.977943274Z" level=info msg="CreateContainer within sandbox \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\"" Oct 8 20:04:26.978443 containerd[1462]: time="2024-10-08T20:04:26.978415906Z" level=info msg="StartContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\"" Oct 8 20:04:27.010650 systemd[1]: Started cri-containerd-443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f.scope - libcontainer container 443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f. Oct 8 20:04:27.046250 containerd[1462]: time="2024-10-08T20:04:27.046193136Z" level=info msg="StartContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" returns successfully" Oct 8 20:04:27.485664 kubelet[2496]: E1008 20:04:27.485622 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:27.489191 containerd[1462]: time="2024-10-08T20:04:27.488872112Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Oct 8 20:04:27.495401 kubelet[2496]: E1008 20:04:27.495351 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:27.510949 containerd[1462]: time="2024-10-08T20:04:27.510897222Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\"" Oct 8 20:04:27.512487 containerd[1462]: time="2024-10-08T20:04:27.511384541Z" level=info msg="StartContainer for \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\"" Oct 8 20:04:27.560674 systemd[1]: Started cri-containerd-a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7.scope - libcontainer container a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7. Oct 8 20:04:27.612555 systemd[1]: cri-containerd-a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7.scope: Deactivated successfully. Oct 8 20:04:27.621640 containerd[1462]: time="2024-10-08T20:04:27.621554473Z" level=info msg="StartContainer for \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\" returns successfully" Oct 8 20:04:27.657542 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7-rootfs.mount: Deactivated successfully. Oct 8 20:04:28.009099 containerd[1462]: time="2024-10-08T20:04:28.009038712Z" level=info msg="shim disconnected" id=a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7 namespace=k8s.io Oct 8 20:04:28.009783 containerd[1462]: time="2024-10-08T20:04:28.009736096Z" level=warning msg="cleaning up after shim disconnected" id=a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7 namespace=k8s.io Oct 8 20:04:28.009783 containerd[1462]: time="2024-10-08T20:04:28.009760302Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:04:28.498959 kubelet[2496]: E1008 20:04:28.498129 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:28.498959 kubelet[2496]: E1008 20:04:28.498686 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:28.503092 containerd[1462]: time="2024-10-08T20:04:28.502833647Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Oct 8 20:04:28.653980 kubelet[2496]: I1008 20:04:28.653915 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-vh7nj" podStartSLOduration=2.573376998 podStartE2EDuration="14.653897212s" podCreationTimestamp="2024-10-08 20:04:14 +0000 UTC" firstStartedPulling="2024-10-08 20:04:14.882247489 +0000 UTC m=+6.543471444" lastFinishedPulling="2024-10-08 20:04:26.962767704 +0000 UTC m=+18.623991658" observedRunningTime="2024-10-08 20:04:27.527184769 +0000 UTC m=+19.188408723" watchObservedRunningTime="2024-10-08 20:04:28.653897212 +0000 UTC m=+20.315121156" Oct 8 20:04:29.187774 containerd[1462]: time="2024-10-08T20:04:29.187710202Z" level=info msg="CreateContainer within sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\"" Oct 8 20:04:29.188366 containerd[1462]: time="2024-10-08T20:04:29.188300965Z" level=info msg="StartContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\"" Oct 8 20:04:29.219601 systemd[1]: Started cri-containerd-3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93.scope - libcontainer container 3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93. Oct 8 20:04:29.409662 containerd[1462]: time="2024-10-08T20:04:29.409604637Z" level=info msg="StartContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" returns successfully" Oct 8 20:04:29.590651 kubelet[2496]: I1008 20:04:29.590611 2496 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Oct 8 20:04:29.659574 systemd[1]: Created slice kubepods-burstable-pod677bd8c4_98db_4ad3_96f2_b3a82703f8b9.slice - libcontainer container kubepods-burstable-pod677bd8c4_98db_4ad3_96f2_b3a82703f8b9.slice. Oct 8 20:04:29.666272 systemd[1]: Created slice kubepods-burstable-pod87be2043_a5cb_462b_a162_5c88cef9ac39.slice - libcontainer container kubepods-burstable-pod87be2043_a5cb_462b_a162_5c88cef9ac39.slice. Oct 8 20:04:29.745343 kubelet[2496]: I1008 20:04:29.745309 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87be2043-a5cb-462b-a162-5c88cef9ac39-config-volume\") pod \"coredns-6f6b679f8f-7dgnr\" (UID: \"87be2043-a5cb-462b-a162-5c88cef9ac39\") " pod="kube-system/coredns-6f6b679f8f-7dgnr" Oct 8 20:04:29.745598 kubelet[2496]: I1008 20:04:29.745358 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89l6b\" (UniqueName: \"kubernetes.io/projected/677bd8c4-98db-4ad3-96f2-b3a82703f8b9-kube-api-access-89l6b\") pod \"coredns-6f6b679f8f-94x7t\" (UID: \"677bd8c4-98db-4ad3-96f2-b3a82703f8b9\") " pod="kube-system/coredns-6f6b679f8f-94x7t" Oct 8 20:04:29.745598 kubelet[2496]: I1008 20:04:29.745381 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/677bd8c4-98db-4ad3-96f2-b3a82703f8b9-config-volume\") pod \"coredns-6f6b679f8f-94x7t\" (UID: \"677bd8c4-98db-4ad3-96f2-b3a82703f8b9\") " pod="kube-system/coredns-6f6b679f8f-94x7t" Oct 8 20:04:29.745598 kubelet[2496]: I1008 20:04:29.745396 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5p4b\" (UniqueName: \"kubernetes.io/projected/87be2043-a5cb-462b-a162-5c88cef9ac39-kube-api-access-s5p4b\") pod \"coredns-6f6b679f8f-7dgnr\" (UID: \"87be2043-a5cb-462b-a162-5c88cef9ac39\") " pod="kube-system/coredns-6f6b679f8f-7dgnr" Oct 8 20:04:29.977081 kubelet[2496]: E1008 20:04:29.976940 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:29.977081 kubelet[2496]: E1008 20:04:29.976945 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:29.986772 containerd[1462]: time="2024-10-08T20:04:29.986714649Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-94x7t,Uid:677bd8c4-98db-4ad3-96f2-b3a82703f8b9,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:29.988020 containerd[1462]: time="2024-10-08T20:04:29.987976215Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7dgnr,Uid:87be2043-a5cb-462b-a162-5c88cef9ac39,Namespace:kube-system,Attempt:0,}" Oct 8 20:04:30.507723 kubelet[2496]: E1008 20:04:30.507671 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:30.537987 kubelet[2496]: I1008 20:04:30.537910 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-nbqd6" podStartSLOduration=6.764807062 podStartE2EDuration="16.537892413s" podCreationTimestamp="2024-10-08 20:04:14 +0000 UTC" firstStartedPulling="2024-10-08 20:04:14.792850745 +0000 UTC m=+6.454074699" lastFinishedPulling="2024-10-08 20:04:24.565936096 +0000 UTC m=+16.227160050" observedRunningTime="2024-10-08 20:04:30.537526775 +0000 UTC m=+22.198750759" watchObservedRunningTime="2024-10-08 20:04:30.537892413 +0000 UTC m=+22.199116368" Oct 8 20:04:31.507496 kubelet[2496]: E1008 20:04:31.507452 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:31.646611 systemd-networkd[1389]: cilium_host: Link UP Oct 8 20:04:31.646793 systemd-networkd[1389]: cilium_net: Link UP Oct 8 20:04:31.646965 systemd-networkd[1389]: cilium_net: Gained carrier Oct 8 20:04:31.647142 systemd-networkd[1389]: cilium_host: Gained carrier Oct 8 20:04:31.651609 systemd-networkd[1389]: cilium_net: Gained IPv6LL Oct 8 20:04:31.754210 systemd-networkd[1389]: cilium_vxlan: Link UP Oct 8 20:04:31.754223 systemd-networkd[1389]: cilium_vxlan: Gained carrier Oct 8 20:04:31.958614 kernel: NET: Registered PF_ALG protocol family Oct 8 20:04:32.471631 systemd-networkd[1389]: cilium_host: Gained IPv6LL Oct 8 20:04:32.509049 kubelet[2496]: E1008 20:04:32.509024 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:32.601375 systemd-networkd[1389]: lxc_health: Link UP Oct 8 20:04:32.610923 systemd-networkd[1389]: lxc_health: Gained carrier Oct 8 20:04:32.919676 systemd-networkd[1389]: cilium_vxlan: Gained IPv6LL Oct 8 20:04:33.204109 systemd-networkd[1389]: lxc61b90b48203e: Link UP Oct 8 20:04:33.213343 systemd-networkd[1389]: lxc4fe1f673b121: Link UP Oct 8 20:04:33.224488 kernel: eth0: renamed from tmpba5e9 Oct 8 20:04:33.231549 kernel: eth0: renamed from tmp40ade Oct 8 20:04:33.238646 systemd-networkd[1389]: lxc4fe1f673b121: Gained carrier Oct 8 20:04:33.238842 systemd-networkd[1389]: lxc61b90b48203e: Gained carrier Oct 8 20:04:33.510611 kubelet[2496]: E1008 20:04:33.510491 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:34.071652 systemd-networkd[1389]: lxc_health: Gained IPv6LL Oct 8 20:04:34.511783 kubelet[2496]: E1008 20:04:34.511743 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:34.647682 systemd-networkd[1389]: lxc61b90b48203e: Gained IPv6LL Oct 8 20:04:34.904616 systemd-networkd[1389]: lxc4fe1f673b121: Gained IPv6LL Oct 8 20:04:36.755029 containerd[1462]: time="2024-10-08T20:04:36.754811915Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:36.755029 containerd[1462]: time="2024-10-08T20:04:36.754899259Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:36.755029 containerd[1462]: time="2024-10-08T20:04:36.754913345Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:36.755803 containerd[1462]: time="2024-10-08T20:04:36.755024425Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:36.778607 systemd[1]: Started cri-containerd-ba5e9bd4d87383b3728284c5ba3338190b81ce14e942d429119a38e99eacb266.scope - libcontainer container ba5e9bd4d87383b3728284c5ba3338190b81ce14e942d429119a38e99eacb266. Oct 8 20:04:36.792189 systemd-resolved[1354]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 8 20:04:36.819068 containerd[1462]: time="2024-10-08T20:04:36.819010173Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-94x7t,Uid:677bd8c4-98db-4ad3-96f2-b3a82703f8b9,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba5e9bd4d87383b3728284c5ba3338190b81ce14e942d429119a38e99eacb266\"" Oct 8 20:04:36.819803 kubelet[2496]: E1008 20:04:36.819778 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:36.820436 containerd[1462]: time="2024-10-08T20:04:36.819647672Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:04:36.820436 containerd[1462]: time="2024-10-08T20:04:36.819711943Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:04:36.820436 containerd[1462]: time="2024-10-08T20:04:36.819725108Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:36.820436 containerd[1462]: time="2024-10-08T20:04:36.819825837Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:04:36.824505 containerd[1462]: time="2024-10-08T20:04:36.822188901Z" level=info msg="CreateContainer within sandbox \"ba5e9bd4d87383b3728284c5ba3338190b81ce14e942d429119a38e99eacb266\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 8 20:04:36.844619 systemd[1]: Started cri-containerd-40ade049211d8c2951c6f7615bcb5aed2d6c43dd377e3c7e422821e153d82d6c.scope - libcontainer container 40ade049211d8c2951c6f7615bcb5aed2d6c43dd377e3c7e422821e153d82d6c. Oct 8 20:04:36.857026 systemd-resolved[1354]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Oct 8 20:04:36.882514 containerd[1462]: time="2024-10-08T20:04:36.882398238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7dgnr,Uid:87be2043-a5cb-462b-a162-5c88cef9ac39,Namespace:kube-system,Attempt:0,} returns sandbox id \"40ade049211d8c2951c6f7615bcb5aed2d6c43dd377e3c7e422821e153d82d6c\"" Oct 8 20:04:36.883048 kubelet[2496]: E1008 20:04:36.883019 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:36.888787 containerd[1462]: time="2024-10-08T20:04:36.888759321Z" level=info msg="CreateContainer within sandbox \"40ade049211d8c2951c6f7615bcb5aed2d6c43dd377e3c7e422821e153d82d6c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 8 20:04:37.162298 systemd[1]: Started sshd@7-10.0.0.126:22-10.0.0.1:51982.service - OpenSSH per-connection server daemon (10.0.0.1:51982). Oct 8 20:04:37.545322 sshd[3818]: Accepted publickey for core from 10.0.0.1 port 51982 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:37.547021 sshd[3818]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:37.550892 systemd-logind[1448]: New session 8 of user core. Oct 8 20:04:37.567612 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 8 20:04:37.759700 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1717203144.mount: Deactivated successfully. Oct 8 20:04:37.806163 containerd[1462]: time="2024-10-08T20:04:37.806027705Z" level=info msg="CreateContainer within sandbox \"ba5e9bd4d87383b3728284c5ba3338190b81ce14e942d429119a38e99eacb266\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859\"" Oct 8 20:04:37.806876 containerd[1462]: time="2024-10-08T20:04:37.806832208Z" level=info msg="StartContainer for \"74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859\"" Oct 8 20:04:37.831931 systemd[1]: run-containerd-runc-k8s.io-74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859-runc.kp875C.mount: Deactivated successfully. Oct 8 20:04:37.848705 systemd[1]: Started cri-containerd-74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859.scope - libcontainer container 74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859. Oct 8 20:04:38.054463 sshd[3818]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:38.058718 systemd[1]: sshd@7-10.0.0.126:22-10.0.0.1:51982.service: Deactivated successfully. Oct 8 20:04:38.060953 systemd[1]: session-8.scope: Deactivated successfully. Oct 8 20:04:38.061633 systemd-logind[1448]: Session 8 logged out. Waiting for processes to exit. Oct 8 20:04:38.062604 systemd-logind[1448]: Removed session 8. Oct 8 20:04:38.283550 containerd[1462]: time="2024-10-08T20:04:38.283455745Z" level=info msg="CreateContainer within sandbox \"40ade049211d8c2951c6f7615bcb5aed2d6c43dd377e3c7e422821e153d82d6c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e63fe4e5e4c6d4375e316ad5ed4915c4772b27943d671637c7bce9f78f7a201f\"" Oct 8 20:04:38.283952 containerd[1462]: time="2024-10-08T20:04:38.283522811Z" level=info msg="StartContainer for \"74e0c7d20756ff29c04d9e16af53cc433d71da7b2d4a74f55a993b07f3e7b859\" returns successfully" Oct 8 20:04:38.284545 containerd[1462]: time="2024-10-08T20:04:38.284315752Z" level=info msg="StartContainer for \"e63fe4e5e4c6d4375e316ad5ed4915c4772b27943d671637c7bce9f78f7a201f\"" Oct 8 20:04:38.320662 systemd[1]: Started cri-containerd-e63fe4e5e4c6d4375e316ad5ed4915c4772b27943d671637c7bce9f78f7a201f.scope - libcontainer container e63fe4e5e4c6d4375e316ad5ed4915c4772b27943d671637c7bce9f78f7a201f. Oct 8 20:04:38.425230 containerd[1462]: time="2024-10-08T20:04:38.425158155Z" level=info msg="StartContainer for \"e63fe4e5e4c6d4375e316ad5ed4915c4772b27943d671637c7bce9f78f7a201f\" returns successfully" Oct 8 20:04:38.541162 kubelet[2496]: E1008 20:04:38.541048 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:38.542886 kubelet[2496]: E1008 20:04:38.542854 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:38.695352 kubelet[2496]: I1008 20:04:38.695275 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-94x7t" podStartSLOduration=24.695252392 podStartE2EDuration="24.695252392s" podCreationTimestamp="2024-10-08 20:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:38.692927943 +0000 UTC m=+30.354151887" watchObservedRunningTime="2024-10-08 20:04:38.695252392 +0000 UTC m=+30.356476346" Oct 8 20:04:38.695804 kubelet[2496]: I1008 20:04:38.695766 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-7dgnr" podStartSLOduration=24.695758735 podStartE2EDuration="24.695758735s" podCreationTimestamp="2024-10-08 20:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:04:38.670252124 +0000 UTC m=+30.331476078" watchObservedRunningTime="2024-10-08 20:04:38.695758735 +0000 UTC m=+30.356982689" Oct 8 20:04:39.545073 kubelet[2496]: E1008 20:04:39.544861 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:39.545073 kubelet[2496]: E1008 20:04:39.544934 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:40.546139 kubelet[2496]: E1008 20:04:40.546107 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:40.546139 kubelet[2496]: E1008 20:04:40.546136 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:04:43.073850 systemd[1]: Started sshd@8-10.0.0.126:22-10.0.0.1:39296.service - OpenSSH per-connection server daemon (10.0.0.1:39296). Oct 8 20:04:43.114780 sshd[3929]: Accepted publickey for core from 10.0.0.1 port 39296 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:43.116563 sshd[3929]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:43.120677 systemd-logind[1448]: New session 9 of user core. Oct 8 20:04:43.132657 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 8 20:04:43.284710 sshd[3929]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:43.288961 systemd[1]: sshd@8-10.0.0.126:22-10.0.0.1:39296.service: Deactivated successfully. Oct 8 20:04:43.291029 systemd[1]: session-9.scope: Deactivated successfully. Oct 8 20:04:43.291699 systemd-logind[1448]: Session 9 logged out. Waiting for processes to exit. Oct 8 20:04:43.292696 systemd-logind[1448]: Removed session 9. Oct 8 20:04:48.296511 systemd[1]: Started sshd@9-10.0.0.126:22-10.0.0.1:39302.service - OpenSSH per-connection server daemon (10.0.0.1:39302). Oct 8 20:04:48.332145 sshd[3947]: Accepted publickey for core from 10.0.0.1 port 39302 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:48.333692 sshd[3947]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:48.337765 systemd-logind[1448]: New session 10 of user core. Oct 8 20:04:48.347586 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 8 20:04:48.474819 sshd[3947]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:48.479345 systemd[1]: sshd@9-10.0.0.126:22-10.0.0.1:39302.service: Deactivated successfully. Oct 8 20:04:48.481362 systemd[1]: session-10.scope: Deactivated successfully. Oct 8 20:04:48.482063 systemd-logind[1448]: Session 10 logged out. Waiting for processes to exit. Oct 8 20:04:48.482911 systemd-logind[1448]: Removed session 10. Oct 8 20:04:53.490244 systemd[1]: Started sshd@10-10.0.0.126:22-10.0.0.1:38710.service - OpenSSH per-connection server daemon (10.0.0.1:38710). Oct 8 20:04:53.524518 sshd[3962]: Accepted publickey for core from 10.0.0.1 port 38710 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:53.525940 sshd[3962]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:53.529691 systemd-logind[1448]: New session 11 of user core. Oct 8 20:04:53.540597 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 8 20:04:53.643057 sshd[3962]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:53.661200 systemd[1]: sshd@10-10.0.0.126:22-10.0.0.1:38710.service: Deactivated successfully. Oct 8 20:04:53.663199 systemd[1]: session-11.scope: Deactivated successfully. Oct 8 20:04:53.665120 systemd-logind[1448]: Session 11 logged out. Waiting for processes to exit. Oct 8 20:04:53.666627 systemd[1]: Started sshd@11-10.0.0.126:22-10.0.0.1:38716.service - OpenSSH per-connection server daemon (10.0.0.1:38716). Oct 8 20:04:53.667584 systemd-logind[1448]: Removed session 11. Oct 8 20:04:53.714421 sshd[3978]: Accepted publickey for core from 10.0.0.1 port 38716 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:53.715972 sshd[3978]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:53.720024 systemd-logind[1448]: New session 12 of user core. Oct 8 20:04:53.729613 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 8 20:04:53.874811 sshd[3978]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:53.885627 systemd[1]: sshd@11-10.0.0.126:22-10.0.0.1:38716.service: Deactivated successfully. Oct 8 20:04:53.888187 systemd[1]: session-12.scope: Deactivated successfully. Oct 8 20:04:53.891047 systemd-logind[1448]: Session 12 logged out. Waiting for processes to exit. Oct 8 20:04:53.900490 systemd[1]: Started sshd@12-10.0.0.126:22-10.0.0.1:38728.service - OpenSSH per-connection server daemon (10.0.0.1:38728). Oct 8 20:04:53.901598 systemd-logind[1448]: Removed session 12. Oct 8 20:04:53.934966 sshd[3991]: Accepted publickey for core from 10.0.0.1 port 38728 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:53.936717 sshd[3991]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:53.941022 systemd-logind[1448]: New session 13 of user core. Oct 8 20:04:53.951601 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 8 20:04:54.068341 sshd[3991]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:54.072410 systemd[1]: sshd@12-10.0.0.126:22-10.0.0.1:38728.service: Deactivated successfully. Oct 8 20:04:54.074441 systemd[1]: session-13.scope: Deactivated successfully. Oct 8 20:04:54.075150 systemd-logind[1448]: Session 13 logged out. Waiting for processes to exit. Oct 8 20:04:54.076225 systemd-logind[1448]: Removed session 13. Oct 8 20:04:59.081946 systemd[1]: Started sshd@13-10.0.0.126:22-10.0.0.1:38758.service - OpenSSH per-connection server daemon (10.0.0.1:38758). Oct 8 20:04:59.115595 sshd[4007]: Accepted publickey for core from 10.0.0.1 port 38758 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:04:59.116974 sshd[4007]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:04:59.120626 systemd-logind[1448]: New session 14 of user core. Oct 8 20:04:59.130585 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 8 20:04:59.234345 sshd[4007]: pam_unix(sshd:session): session closed for user core Oct 8 20:04:59.238797 systemd[1]: sshd@13-10.0.0.126:22-10.0.0.1:38758.service: Deactivated successfully. Oct 8 20:04:59.240858 systemd[1]: session-14.scope: Deactivated successfully. Oct 8 20:04:59.241559 systemd-logind[1448]: Session 14 logged out. Waiting for processes to exit. Oct 8 20:04:59.242408 systemd-logind[1448]: Removed session 14. Oct 8 20:05:04.247451 systemd[1]: Started sshd@14-10.0.0.126:22-10.0.0.1:58670.service - OpenSSH per-connection server daemon (10.0.0.1:58670). Oct 8 20:05:04.287845 sshd[4022]: Accepted publickey for core from 10.0.0.1 port 58670 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:04.289870 sshd[4022]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:04.294148 systemd-logind[1448]: New session 15 of user core. Oct 8 20:05:04.313652 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 8 20:05:04.426141 sshd[4022]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:04.430463 systemd[1]: sshd@14-10.0.0.126:22-10.0.0.1:58670.service: Deactivated successfully. Oct 8 20:05:04.432453 systemd[1]: session-15.scope: Deactivated successfully. Oct 8 20:05:04.433247 systemd-logind[1448]: Session 15 logged out. Waiting for processes to exit. Oct 8 20:05:04.434263 systemd-logind[1448]: Removed session 15. Oct 8 20:05:09.437544 systemd[1]: Started sshd@15-10.0.0.126:22-10.0.0.1:58682.service - OpenSSH per-connection server daemon (10.0.0.1:58682). Oct 8 20:05:09.473879 sshd[4039]: Accepted publickey for core from 10.0.0.1 port 58682 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:09.475522 sshd[4039]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:09.479502 systemd-logind[1448]: New session 16 of user core. Oct 8 20:05:09.487621 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 8 20:05:09.596148 sshd[4039]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:09.612384 systemd[1]: sshd@15-10.0.0.126:22-10.0.0.1:58682.service: Deactivated successfully. Oct 8 20:05:09.614388 systemd[1]: session-16.scope: Deactivated successfully. Oct 8 20:05:09.615904 systemd-logind[1448]: Session 16 logged out. Waiting for processes to exit. Oct 8 20:05:09.620780 systemd[1]: Started sshd@16-10.0.0.126:22-10.0.0.1:58686.service - OpenSSH per-connection server daemon (10.0.0.1:58686). Oct 8 20:05:09.621748 systemd-logind[1448]: Removed session 16. Oct 8 20:05:09.651893 sshd[4053]: Accepted publickey for core from 10.0.0.1 port 58686 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:09.653363 sshd[4053]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:09.657231 systemd-logind[1448]: New session 17 of user core. Oct 8 20:05:09.667578 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 8 20:05:10.064906 sshd[4053]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:10.075102 systemd[1]: sshd@16-10.0.0.126:22-10.0.0.1:58686.service: Deactivated successfully. Oct 8 20:05:10.077623 systemd[1]: session-17.scope: Deactivated successfully. Oct 8 20:05:10.079673 systemd-logind[1448]: Session 17 logged out. Waiting for processes to exit. Oct 8 20:05:10.084809 systemd[1]: Started sshd@17-10.0.0.126:22-10.0.0.1:58696.service - OpenSSH per-connection server daemon (10.0.0.1:58696). Oct 8 20:05:10.085908 systemd-logind[1448]: Removed session 17. Oct 8 20:05:10.121202 sshd[4066]: Accepted publickey for core from 10.0.0.1 port 58696 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:10.122734 sshd[4066]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:10.126928 systemd-logind[1448]: New session 18 of user core. Oct 8 20:05:10.138606 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 8 20:05:12.027689 sshd[4066]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:12.035311 systemd[1]: sshd@17-10.0.0.126:22-10.0.0.1:58696.service: Deactivated successfully. Oct 8 20:05:12.037370 systemd[1]: session-18.scope: Deactivated successfully. Oct 8 20:05:12.039050 systemd-logind[1448]: Session 18 logged out. Waiting for processes to exit. Oct 8 20:05:12.044731 systemd[1]: Started sshd@18-10.0.0.126:22-10.0.0.1:56440.service - OpenSSH per-connection server daemon (10.0.0.1:56440). Oct 8 20:05:12.045601 systemd-logind[1448]: Removed session 18. Oct 8 20:05:12.076466 sshd[4085]: Accepted publickey for core from 10.0.0.1 port 56440 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:12.078235 sshd[4085]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:12.082177 systemd-logind[1448]: New session 19 of user core. Oct 8 20:05:12.090631 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 8 20:05:12.446804 sshd[4085]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:12.454256 systemd[1]: sshd@18-10.0.0.126:22-10.0.0.1:56440.service: Deactivated successfully. Oct 8 20:05:12.456253 systemd[1]: session-19.scope: Deactivated successfully. Oct 8 20:05:12.457814 systemd-logind[1448]: Session 19 logged out. Waiting for processes to exit. Oct 8 20:05:12.465825 systemd[1]: Started sshd@19-10.0.0.126:22-10.0.0.1:56452.service - OpenSSH per-connection server daemon (10.0.0.1:56452). Oct 8 20:05:12.466818 systemd-logind[1448]: Removed session 19. Oct 8 20:05:12.495922 sshd[4098]: Accepted publickey for core from 10.0.0.1 port 56452 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:12.497443 sshd[4098]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:12.501937 systemd-logind[1448]: New session 20 of user core. Oct 8 20:05:12.506633 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 8 20:05:12.633043 sshd[4098]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:12.636865 systemd[1]: sshd@19-10.0.0.126:22-10.0.0.1:56452.service: Deactivated successfully. Oct 8 20:05:12.638976 systemd[1]: session-20.scope: Deactivated successfully. Oct 8 20:05:12.639611 systemd-logind[1448]: Session 20 logged out. Waiting for processes to exit. Oct 8 20:05:12.640415 systemd-logind[1448]: Removed session 20. Oct 8 20:05:17.644411 systemd[1]: Started sshd@20-10.0.0.126:22-10.0.0.1:56500.service - OpenSSH per-connection server daemon (10.0.0.1:56500). Oct 8 20:05:17.679885 sshd[4116]: Accepted publickey for core from 10.0.0.1 port 56500 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:17.681952 sshd[4116]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:17.686610 systemd-logind[1448]: New session 21 of user core. Oct 8 20:05:17.696655 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 8 20:05:17.824323 sshd[4116]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:17.827013 systemd[1]: sshd@20-10.0.0.126:22-10.0.0.1:56500.service: Deactivated successfully. Oct 8 20:05:17.830126 systemd[1]: session-21.scope: Deactivated successfully. Oct 8 20:05:17.830872 systemd-logind[1448]: Session 21 logged out. Waiting for processes to exit. Oct 8 20:05:17.831838 systemd-logind[1448]: Removed session 21. Oct 8 20:05:22.428458 kubelet[2496]: E1008 20:05:22.428421 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:22.837022 systemd[1]: Started sshd@21-10.0.0.126:22-10.0.0.1:58060.service - OpenSSH per-connection server daemon (10.0.0.1:58060). Oct 8 20:05:22.876386 sshd[4134]: Accepted publickey for core from 10.0.0.1 port 58060 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:22.878121 sshd[4134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:22.882015 systemd-logind[1448]: New session 22 of user core. Oct 8 20:05:22.892594 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 8 20:05:22.997584 sshd[4134]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:23.001798 systemd[1]: sshd@21-10.0.0.126:22-10.0.0.1:58060.service: Deactivated successfully. Oct 8 20:05:23.003882 systemd[1]: session-22.scope: Deactivated successfully. Oct 8 20:05:23.004606 systemd-logind[1448]: Session 22 logged out. Waiting for processes to exit. Oct 8 20:05:23.005602 systemd-logind[1448]: Removed session 22. Oct 8 20:05:28.008487 systemd[1]: Started sshd@22-10.0.0.126:22-10.0.0.1:58072.service - OpenSSH per-connection server daemon (10.0.0.1:58072). Oct 8 20:05:28.047290 sshd[4148]: Accepted publickey for core from 10.0.0.1 port 58072 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:28.049147 sshd[4148]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:28.054049 systemd-logind[1448]: New session 23 of user core. Oct 8 20:05:28.060737 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 8 20:05:28.175329 sshd[4148]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:28.179989 systemd[1]: sshd@22-10.0.0.126:22-10.0.0.1:58072.service: Deactivated successfully. Oct 8 20:05:28.182148 systemd[1]: session-23.scope: Deactivated successfully. Oct 8 20:05:28.182933 systemd-logind[1448]: Session 23 logged out. Waiting for processes to exit. Oct 8 20:05:28.183911 systemd-logind[1448]: Removed session 23. Oct 8 20:05:29.428724 kubelet[2496]: E1008 20:05:29.428677 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:33.186071 systemd[1]: Started sshd@23-10.0.0.126:22-10.0.0.1:54786.service - OpenSSH per-connection server daemon (10.0.0.1:54786). Oct 8 20:05:33.221624 sshd[4163]: Accepted publickey for core from 10.0.0.1 port 54786 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:33.223323 sshd[4163]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:33.227378 systemd-logind[1448]: New session 24 of user core. Oct 8 20:05:33.237619 systemd[1]: Started session-24.scope - Session 24 of User core. Oct 8 20:05:33.355653 sshd[4163]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:33.375655 systemd[1]: sshd@23-10.0.0.126:22-10.0.0.1:54786.service: Deactivated successfully. Oct 8 20:05:33.377409 systemd[1]: session-24.scope: Deactivated successfully. Oct 8 20:05:33.379165 systemd-logind[1448]: Session 24 logged out. Waiting for processes to exit. Oct 8 20:05:33.386028 systemd[1]: Started sshd@24-10.0.0.126:22-10.0.0.1:54790.service - OpenSSH per-connection server daemon (10.0.0.1:54790). Oct 8 20:05:33.387246 systemd-logind[1448]: Removed session 24. Oct 8 20:05:33.416238 sshd[4177]: Accepted publickey for core from 10.0.0.1 port 54790 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:33.417756 sshd[4177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:33.422272 systemd-logind[1448]: New session 25 of user core. Oct 8 20:05:33.434714 systemd[1]: Started session-25.scope - Session 25 of User core. Oct 8 20:05:34.427965 kubelet[2496]: E1008 20:05:34.427921 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:34.929980 containerd[1462]: time="2024-10-08T20:05:34.929916353Z" level=info msg="StopContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" with timeout 30 (s)" Oct 8 20:05:34.930643 containerd[1462]: time="2024-10-08T20:05:34.930321975Z" level=info msg="Stop container \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" with signal terminated" Oct 8 20:05:34.947157 systemd[1]: cri-containerd-443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f.scope: Deactivated successfully. Oct 8 20:05:34.973159 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f-rootfs.mount: Deactivated successfully. Oct 8 20:05:34.989580 containerd[1462]: time="2024-10-08T20:05:34.989534596Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 8 20:05:34.997797 containerd[1462]: time="2024-10-08T20:05:34.997748271Z" level=info msg="StopContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" with timeout 2 (s)" Oct 8 20:05:34.998025 containerd[1462]: time="2024-10-08T20:05:34.997992986Z" level=info msg="Stop container \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" with signal terminated" Oct 8 20:05:35.006312 systemd-networkd[1389]: lxc_health: Link DOWN Oct 8 20:05:35.006321 systemd-networkd[1389]: lxc_health: Lost carrier Oct 8 20:05:35.094010 systemd[1]: cri-containerd-3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93.scope: Deactivated successfully. Oct 8 20:05:35.094427 systemd[1]: cri-containerd-3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93.scope: Consumed 7.017s CPU time. Oct 8 20:05:35.112438 containerd[1462]: time="2024-10-08T20:05:35.112171886Z" level=info msg="shim disconnected" id=443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f namespace=k8s.io Oct 8 20:05:35.112438 containerd[1462]: time="2024-10-08T20:05:35.112436549Z" level=warning msg="cleaning up after shim disconnected" id=443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f namespace=k8s.io Oct 8 20:05:35.112637 containerd[1462]: time="2024-10-08T20:05:35.112463411Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:35.118745 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93-rootfs.mount: Deactivated successfully. Oct 8 20:05:35.240213 containerd[1462]: time="2024-10-08T20:05:35.239841003Z" level=info msg="StopContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" returns successfully" Oct 8 20:05:35.240829 containerd[1462]: time="2024-10-08T20:05:35.240768967Z" level=info msg="shim disconnected" id=3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93 namespace=k8s.io Oct 8 20:05:35.240829 containerd[1462]: time="2024-10-08T20:05:35.240828059Z" level=warning msg="cleaning up after shim disconnected" id=3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93 namespace=k8s.io Oct 8 20:05:35.240944 containerd[1462]: time="2024-10-08T20:05:35.240845923Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:35.244872 containerd[1462]: time="2024-10-08T20:05:35.244812482Z" level=info msg="StopPodSandbox for \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\"" Oct 8 20:05:35.244998 containerd[1462]: time="2024-10-08T20:05:35.244888166Z" level=info msg="Container to stop \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.247953 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3-shm.mount: Deactivated successfully. Oct 8 20:05:35.256065 systemd[1]: cri-containerd-4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3.scope: Deactivated successfully. Oct 8 20:05:35.280825 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3-rootfs.mount: Deactivated successfully. Oct 8 20:05:35.375813 containerd[1462]: time="2024-10-08T20:05:35.375737196Z" level=info msg="StopContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" returns successfully" Oct 8 20:05:35.376387 containerd[1462]: time="2024-10-08T20:05:35.376345582Z" level=info msg="StopPodSandbox for \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\"" Oct 8 20:05:35.376446 containerd[1462]: time="2024-10-08T20:05:35.376390888Z" level=info msg="Container to stop \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.376446 containerd[1462]: time="2024-10-08T20:05:35.376404955Z" level=info msg="Container to stop \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.376446 containerd[1462]: time="2024-10-08T20:05:35.376413902Z" level=info msg="Container to stop \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.376446 containerd[1462]: time="2024-10-08T20:05:35.376422258Z" level=info msg="Container to stop \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.376446 containerd[1462]: time="2024-10-08T20:05:35.376432848Z" level=info msg="Container to stop \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 8 20:05:35.382713 systemd[1]: cri-containerd-f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8.scope: Deactivated successfully. Oct 8 20:05:35.543902 containerd[1462]: time="2024-10-08T20:05:35.543722324Z" level=info msg="shim disconnected" id=4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3 namespace=k8s.io Oct 8 20:05:35.543902 containerd[1462]: time="2024-10-08T20:05:35.543778531Z" level=warning msg="cleaning up after shim disconnected" id=4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3 namespace=k8s.io Oct 8 20:05:35.543902 containerd[1462]: time="2024-10-08T20:05:35.543787618Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:35.566484 containerd[1462]: time="2024-10-08T20:05:35.566414681Z" level=info msg="TearDown network for sandbox \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\" successfully" Oct 8 20:05:35.566484 containerd[1462]: time="2024-10-08T20:05:35.566451882Z" level=info msg="StopPodSandbox for \"4d20cf56b504f8e18bc4ee0c0327a3c128b5216d45f19fe38904ba88d6f336f3\" returns successfully" Oct 8 20:05:35.598235 containerd[1462]: time="2024-10-08T20:05:35.598156426Z" level=info msg="shim disconnected" id=f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8 namespace=k8s.io Oct 8 20:05:35.598455 containerd[1462]: time="2024-10-08T20:05:35.598427963Z" level=warning msg="cleaning up after shim disconnected" id=f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8 namespace=k8s.io Oct 8 20:05:35.598455 containerd[1462]: time="2024-10-08T20:05:35.598448120Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:35.630710 containerd[1462]: time="2024-10-08T20:05:35.630646214Z" level=info msg="TearDown network for sandbox \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" successfully" Oct 8 20:05:35.630710 containerd[1462]: time="2024-10-08T20:05:35.630687973Z" level=info msg="StopPodSandbox for \"f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8\" returns successfully" Oct 8 20:05:35.645415 kubelet[2496]: I1008 20:05:35.645379 2496 scope.go:117] "RemoveContainer" containerID="3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93" Oct 8 20:05:35.646531 containerd[1462]: time="2024-10-08T20:05:35.646455158Z" level=info msg="RemoveContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\"" Oct 8 20:05:35.746183 kubelet[2496]: I1008 20:05:35.746132 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-clustermesh-secrets\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746183 kubelet[2496]: I1008 20:05:35.746171 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-net\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746208 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-etc-cni-netd\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746235 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-config-path\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746253 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-bpf-maps\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746275 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/37102e20-313b-42f0-bbca-a02d167c080b-cilium-config-path\") pod \"37102e20-313b-42f0-bbca-a02d167c080b\" (UID: \"37102e20-313b-42f0-bbca-a02d167c080b\") " Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746272 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.746402 kubelet[2496]: I1008 20:05:35.746298 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-xtables-lock\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746321 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-cgroup\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746345 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-lib-modules\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746372 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2wrj\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-kube-api-access-k2wrj\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746400 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-kernel\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746422 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-run\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746608 kubelet[2496]: I1008 20:05:35.746442 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfkhr\" (UniqueName: \"kubernetes.io/projected/37102e20-313b-42f0-bbca-a02d167c080b-kube-api-access-hfkhr\") pod \"37102e20-313b-42f0-bbca-a02d167c080b\" (UID: \"37102e20-313b-42f0-bbca-a02d167c080b\") " Oct 8 20:05:35.746798 kubelet[2496]: I1008 20:05:35.746462 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hostproc\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746798 kubelet[2496]: I1008 20:05:35.746496 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cni-path\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746798 kubelet[2496]: I1008 20:05:35.746516 2496 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hubble-tls\") pod \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\" (UID: \"798ba836-9610-40d9-b6b9-ba0cbd07a5ed\") " Oct 8 20:05:35.746798 kubelet[2496]: I1008 20:05:35.746552 2496 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.746926 kubelet[2496]: I1008 20:05:35.746882 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.749419 kubelet[2496]: I1008 20:05:35.749384 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750559 kubelet[2496]: I1008 20:05:35.749548 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750559 kubelet[2496]: I1008 20:05:35.749575 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hostproc" (OuterVolumeSpecName: "hostproc") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750559 kubelet[2496]: I1008 20:05:35.749576 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750559 kubelet[2496]: I1008 20:05:35.749590 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750559 kubelet[2496]: I1008 20:05:35.749603 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cni-path" (OuterVolumeSpecName: "cni-path") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750754 kubelet[2496]: I1008 20:05:35.749666 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-kube-api-access-k2wrj" (OuterVolumeSpecName: "kube-api-access-k2wrj") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "kube-api-access-k2wrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 8 20:05:35.750754 kubelet[2496]: I1008 20:05:35.749690 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750754 kubelet[2496]: I1008 20:05:35.749705 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 8 20:05:35.750754 kubelet[2496]: I1008 20:05:35.750194 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 8 20:05:35.750754 kubelet[2496]: I1008 20:05:35.750459 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 8 20:05:35.750908 kubelet[2496]: I1008 20:05:35.750708 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "798ba836-9610-40d9-b6b9-ba0cbd07a5ed" (UID: "798ba836-9610-40d9-b6b9-ba0cbd07a5ed"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 8 20:05:35.752400 kubelet[2496]: I1008 20:05:35.752375 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37102e20-313b-42f0-bbca-a02d167c080b-kube-api-access-hfkhr" (OuterVolumeSpecName: "kube-api-access-hfkhr") pod "37102e20-313b-42f0-bbca-a02d167c080b" (UID: "37102e20-313b-42f0-bbca-a02d167c080b"). InnerVolumeSpecName "kube-api-access-hfkhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 8 20:05:35.753173 kubelet[2496]: I1008 20:05:35.753155 2496 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37102e20-313b-42f0-bbca-a02d167c080b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "37102e20-313b-42f0-bbca-a02d167c080b" (UID: "37102e20-313b-42f0-bbca-a02d167c080b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 8 20:05:35.794873 containerd[1462]: time="2024-10-08T20:05:35.794765445Z" level=info msg="RemoveContainer for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" returns successfully" Oct 8 20:05:35.795149 kubelet[2496]: I1008 20:05:35.795113 2496 scope.go:117] "RemoveContainer" containerID="a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7" Oct 8 20:05:35.796182 containerd[1462]: time="2024-10-08T20:05:35.796153533Z" level=info msg="RemoveContainer for \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\"" Oct 8 20:05:35.847621 kubelet[2496]: I1008 20:05:35.847590 2496 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/37102e20-313b-42f0-bbca-a02d167c080b-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847621 kubelet[2496]: I1008 20:05:35.847622 2496 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-xtables-lock\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847621 kubelet[2496]: I1008 20:05:35.847634 2496 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847646 2496 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-hfkhr\" (UniqueName: \"kubernetes.io/projected/37102e20-313b-42f0-bbca-a02d167c080b-kube-api-access-hfkhr\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847662 2496 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-lib-modules\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847676 2496 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-k2wrj\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-kube-api-access-k2wrj\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847687 2496 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847697 2496 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-run\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847707 2496 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hostproc\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847716 2496 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cni-path\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847791 kubelet[2496]: I1008 20:05:35.847724 2496 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-hubble-tls\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847962 kubelet[2496]: I1008 20:05:35.847734 2496 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847962 kubelet[2496]: I1008 20:05:35.847747 2496 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847962 kubelet[2496]: I1008 20:05:35.847760 2496 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.847962 kubelet[2496]: I1008 20:05:35.847770 2496 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/798ba836-9610-40d9-b6b9-ba0cbd07a5ed-bpf-maps\") on node \"localhost\" DevicePath \"\"" Oct 8 20:05:35.870569 containerd[1462]: time="2024-10-08T20:05:35.870523464Z" level=info msg="RemoveContainer for \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\" returns successfully" Oct 8 20:05:35.870833 kubelet[2496]: I1008 20:05:35.870805 2496 scope.go:117] "RemoveContainer" containerID="7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24" Oct 8 20:05:35.871855 containerd[1462]: time="2024-10-08T20:05:35.871826831Z" level=info msg="RemoveContainer for \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\"" Oct 8 20:05:35.937341 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8-rootfs.mount: Deactivated successfully. Oct 8 20:05:35.937448 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f734242b7250ad9ecf158760e8f9860f3431008dc7b18802a4df33cf0b72f5e8-shm.mount: Deactivated successfully. Oct 8 20:05:35.937546 systemd[1]: var-lib-kubelet-pods-37102e20\x2d313b\x2d42f0\x2dbbca\x2da02d167c080b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dhfkhr.mount: Deactivated successfully. Oct 8 20:05:35.937622 systemd[1]: var-lib-kubelet-pods-798ba836\x2d9610\x2d40d9\x2db6b9\x2dba0cbd07a5ed-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dk2wrj.mount: Deactivated successfully. Oct 8 20:05:35.937709 systemd[1]: var-lib-kubelet-pods-798ba836\x2d9610\x2d40d9\x2db6b9\x2dba0cbd07a5ed-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Oct 8 20:05:35.937806 systemd[1]: var-lib-kubelet-pods-798ba836\x2d9610\x2d40d9\x2db6b9\x2dba0cbd07a5ed-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Oct 8 20:05:35.954011 systemd[1]: Removed slice kubepods-burstable-pod798ba836_9610_40d9_b6b9_ba0cbd07a5ed.slice - libcontainer container kubepods-burstable-pod798ba836_9610_40d9_b6b9_ba0cbd07a5ed.slice. Oct 8 20:05:35.954128 systemd[1]: kubepods-burstable-pod798ba836_9610_40d9_b6b9_ba0cbd07a5ed.slice: Consumed 7.127s CPU time. Oct 8 20:05:35.955306 systemd[1]: Removed slice kubepods-besteffort-pod37102e20_313b_42f0_bbca_a02d167c080b.slice - libcontainer container kubepods-besteffort-pod37102e20_313b_42f0_bbca_a02d167c080b.slice. Oct 8 20:05:35.958422 containerd[1462]: time="2024-10-08T20:05:35.958373133Z" level=info msg="RemoveContainer for \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\" returns successfully" Oct 8 20:05:35.959002 kubelet[2496]: I1008 20:05:35.958964 2496 scope.go:117] "RemoveContainer" containerID="7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad" Oct 8 20:05:35.960072 containerd[1462]: time="2024-10-08T20:05:35.960045782Z" level=info msg="RemoveContainer for \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\"" Oct 8 20:05:36.071298 containerd[1462]: time="2024-10-08T20:05:36.071177748Z" level=info msg="RemoveContainer for \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\" returns successfully" Oct 8 20:05:36.071512 kubelet[2496]: I1008 20:05:36.071485 2496 scope.go:117] "RemoveContainer" containerID="649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48" Oct 8 20:05:36.072557 containerd[1462]: time="2024-10-08T20:05:36.072533815Z" level=info msg="RemoveContainer for \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\"" Oct 8 20:05:36.103448 containerd[1462]: time="2024-10-08T20:05:36.103381773Z" level=info msg="RemoveContainer for \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\" returns successfully" Oct 8 20:05:36.103701 kubelet[2496]: I1008 20:05:36.103675 2496 scope.go:117] "RemoveContainer" containerID="3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93" Oct 8 20:05:36.108188 containerd[1462]: time="2024-10-08T20:05:36.108138962Z" level=error msg="ContainerStatus for \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\": not found" Oct 8 20:05:36.116883 kubelet[2496]: E1008 20:05:36.116848 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\": not found" containerID="3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93" Oct 8 20:05:36.117023 kubelet[2496]: I1008 20:05:36.116913 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93"} err="failed to get container status \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\": rpc error: code = NotFound desc = an error occurred when try to find container \"3783070d2100dce44f2f6713b04bd70aa9d431780265468aaf9bd6096255de93\": not found" Oct 8 20:05:36.117023 kubelet[2496]: I1008 20:05:36.117016 2496 scope.go:117] "RemoveContainer" containerID="a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7" Oct 8 20:05:36.117235 containerd[1462]: time="2024-10-08T20:05:36.117195068Z" level=error msg="ContainerStatus for \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\": not found" Oct 8 20:05:36.117362 kubelet[2496]: E1008 20:05:36.117340 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\": not found" containerID="a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7" Oct 8 20:05:36.117418 kubelet[2496]: I1008 20:05:36.117359 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7"} err="failed to get container status \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\": rpc error: code = NotFound desc = an error occurred when try to find container \"a5ced450ad54aead153044fd1425d07096657e9486995001fc27604ef328c3b7\": not found" Oct 8 20:05:36.117418 kubelet[2496]: I1008 20:05:36.117373 2496 scope.go:117] "RemoveContainer" containerID="7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24" Oct 8 20:05:36.117584 containerd[1462]: time="2024-10-08T20:05:36.117552086Z" level=error msg="ContainerStatus for \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\": not found" Oct 8 20:05:36.117694 kubelet[2496]: E1008 20:05:36.117669 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\": not found" containerID="7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24" Oct 8 20:05:36.117732 kubelet[2496]: I1008 20:05:36.117693 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24"} err="failed to get container status \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\": rpc error: code = NotFound desc = an error occurred when try to find container \"7f864bfbd25fdcc99cc5c55d3cb9eac49c5e9d4a9040f05e78b8f68e82630a24\": not found" Oct 8 20:05:36.117732 kubelet[2496]: I1008 20:05:36.117708 2496 scope.go:117] "RemoveContainer" containerID="7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad" Oct 8 20:05:36.117883 containerd[1462]: time="2024-10-08T20:05:36.117855963Z" level=error msg="ContainerStatus for \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\": not found" Oct 8 20:05:36.117971 kubelet[2496]: E1008 20:05:36.117946 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\": not found" containerID="7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad" Oct 8 20:05:36.118012 kubelet[2496]: I1008 20:05:36.117981 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad"} err="failed to get container status \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\": rpc error: code = NotFound desc = an error occurred when try to find container \"7e83b6619cd4464473c1ecc9e58cc623e0d7bf22f661411a0d52dd5b19e2bcad\": not found" Oct 8 20:05:36.118012 kubelet[2496]: I1008 20:05:36.117997 2496 scope.go:117] "RemoveContainer" containerID="649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48" Oct 8 20:05:36.118140 containerd[1462]: time="2024-10-08T20:05:36.118115335Z" level=error msg="ContainerStatus for \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\": not found" Oct 8 20:05:36.118261 kubelet[2496]: E1008 20:05:36.118231 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\": not found" containerID="649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48" Oct 8 20:05:36.118301 kubelet[2496]: I1008 20:05:36.118260 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48"} err="failed to get container status \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\": rpc error: code = NotFound desc = an error occurred when try to find container \"649d159f1d2fb450562b395bfaf20d7a0d0ff37a2a9ab381d3954204e0945a48\": not found" Oct 8 20:05:36.118301 kubelet[2496]: I1008 20:05:36.118278 2496 scope.go:117] "RemoveContainer" containerID="443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f" Oct 8 20:05:36.119149 containerd[1462]: time="2024-10-08T20:05:36.119127348Z" level=info msg="RemoveContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\"" Oct 8 20:05:36.174326 containerd[1462]: time="2024-10-08T20:05:36.174274554Z" level=info msg="RemoveContainer for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" returns successfully" Oct 8 20:05:36.174579 kubelet[2496]: I1008 20:05:36.174547 2496 scope.go:117] "RemoveContainer" containerID="443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f" Oct 8 20:05:36.174841 containerd[1462]: time="2024-10-08T20:05:36.174800863Z" level=error msg="ContainerStatus for \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\": not found" Oct 8 20:05:36.174949 kubelet[2496]: E1008 20:05:36.174925 2496 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\": not found" containerID="443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f" Oct 8 20:05:36.174992 kubelet[2496]: I1008 20:05:36.174954 2496 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f"} err="failed to get container status \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\": rpc error: code = NotFound desc = an error occurred when try to find container \"443d9de8b5563f7284997143f0607ba15fb036f22c3e45e66af0231eaa26312f\": not found" Oct 8 20:05:36.430923 kubelet[2496]: I1008 20:05:36.430875 2496 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37102e20-313b-42f0-bbca-a02d167c080b" path="/var/lib/kubelet/pods/37102e20-313b-42f0-bbca-a02d167c080b/volumes" Oct 8 20:05:36.431590 kubelet[2496]: I1008 20:05:36.431569 2496 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" path="/var/lib/kubelet/pods/798ba836-9610-40d9-b6b9-ba0cbd07a5ed/volumes" Oct 8 20:05:36.883020 sshd[4177]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:36.894995 systemd[1]: sshd@24-10.0.0.126:22-10.0.0.1:54790.service: Deactivated successfully. Oct 8 20:05:36.896986 systemd[1]: session-25.scope: Deactivated successfully. Oct 8 20:05:36.898736 systemd-logind[1448]: Session 25 logged out. Waiting for processes to exit. Oct 8 20:05:36.900242 systemd[1]: Started sshd@25-10.0.0.126:22-10.0.0.1:54796.service - OpenSSH per-connection server daemon (10.0.0.1:54796). Oct 8 20:05:36.901071 systemd-logind[1448]: Removed session 25. Oct 8 20:05:36.943206 sshd[4339]: Accepted publickey for core from 10.0.0.1 port 54796 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:36.945100 sshd[4339]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:36.949460 systemd-logind[1448]: New session 26 of user core. Oct 8 20:05:36.958600 systemd[1]: Started session-26.scope - Session 26 of User core. Oct 8 20:05:37.428260 kubelet[2496]: E1008 20:05:37.428187 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:38.334611 sshd[4339]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:38.343830 systemd[1]: sshd@25-10.0.0.126:22-10.0.0.1:54796.service: Deactivated successfully. Oct 8 20:05:38.345981 systemd[1]: session-26.scope: Deactivated successfully. Oct 8 20:05:38.347726 systemd-logind[1448]: Session 26 logged out. Waiting for processes to exit. Oct 8 20:05:38.361825 systemd[1]: Started sshd@26-10.0.0.126:22-10.0.0.1:54798.service - OpenSSH per-connection server daemon (10.0.0.1:54798). Oct 8 20:05:38.363522 systemd-logind[1448]: Removed session 26. Oct 8 20:05:38.396230 sshd[4352]: Accepted publickey for core from 10.0.0.1 port 54798 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:38.397904 sshd[4352]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:38.401890 systemd-logind[1448]: New session 27 of user core. Oct 8 20:05:38.412628 systemd[1]: Started session-27.scope - Session 27 of User core. Oct 8 20:05:38.463640 sshd[4352]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:38.474367 systemd[1]: sshd@26-10.0.0.126:22-10.0.0.1:54798.service: Deactivated successfully. Oct 8 20:05:38.476315 systemd[1]: session-27.scope: Deactivated successfully. Oct 8 20:05:38.478779 systemd-logind[1448]: Session 27 logged out. Waiting for processes to exit. Oct 8 20:05:38.488937 kubelet[2496]: E1008 20:05:38.488900 2496 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Oct 8 20:05:38.493764 systemd[1]: Started sshd@27-10.0.0.126:22-10.0.0.1:54802.service - OpenSSH per-connection server daemon (10.0.0.1:54802). Oct 8 20:05:38.495791 systemd-logind[1448]: Removed session 27. Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506168 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="37102e20-313b-42f0-bbca-a02d167c080b" containerName="cilium-operator" Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506216 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="clean-cilium-state" Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506228 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="cilium-agent" Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506237 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="mount-cgroup" Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506245 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="apply-sysctl-overwrites" Oct 8 20:05:38.506239 kubelet[2496]: E1008 20:05:38.506254 2496 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="mount-bpf-fs" Oct 8 20:05:38.506603 kubelet[2496]: I1008 20:05:38.506283 2496 memory_manager.go:354] "RemoveStaleState removing state" podUID="798ba836-9610-40d9-b6b9-ba0cbd07a5ed" containerName="cilium-agent" Oct 8 20:05:38.506603 kubelet[2496]: I1008 20:05:38.506293 2496 memory_manager.go:354] "RemoveStaleState removing state" podUID="37102e20-313b-42f0-bbca-a02d167c080b" containerName="cilium-operator" Oct 8 20:05:38.514679 systemd[1]: Created slice kubepods-burstable-podefc190af_8177_48b8_9f9d_1ae657fee9c7.slice - libcontainer container kubepods-burstable-podefc190af_8177_48b8_9f9d_1ae657fee9c7.slice. Oct 8 20:05:38.528067 sshd[4360]: Accepted publickey for core from 10.0.0.1 port 54802 ssh2: RSA SHA256:/xN8BdcoCidXIeJRfI4jO6TdLokQFeWhvR5OfwObqUI Oct 8 20:05:38.529966 sshd[4360]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 8 20:05:38.535852 systemd-logind[1448]: New session 28 of user core. Oct 8 20:05:38.545612 systemd[1]: Started session-28.scope - Session 28 of User core. Oct 8 20:05:38.663836 kubelet[2496]: I1008 20:05:38.663721 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/efc190af-8177-48b8-9f9d-1ae657fee9c7-cilium-config-path\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.663836 kubelet[2496]: I1008 20:05:38.663761 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-cilium-run\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.663836 kubelet[2496]: I1008 20:05:38.663785 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-host-proc-sys-kernel\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.663836 kubelet[2496]: I1008 20:05:38.663802 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/efc190af-8177-48b8-9f9d-1ae657fee9c7-hubble-tls\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.663836 kubelet[2496]: I1008 20:05:38.663817 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/efc190af-8177-48b8-9f9d-1ae657fee9c7-clustermesh-secrets\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663858 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-hostproc\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663881 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-cilium-cgroup\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663899 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-cni-path\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663916 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-etc-cni-netd\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663950 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/efc190af-8177-48b8-9f9d-1ae657fee9c7-cilium-ipsec-secrets\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664047 kubelet[2496]: I1008 20:05:38.663969 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8s7\" (UniqueName: \"kubernetes.io/projected/efc190af-8177-48b8-9f9d-1ae657fee9c7-kube-api-access-6n8s7\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664180 kubelet[2496]: I1008 20:05:38.663989 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-xtables-lock\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664180 kubelet[2496]: I1008 20:05:38.664003 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-host-proc-sys-net\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664180 kubelet[2496]: I1008 20:05:38.664017 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-bpf-maps\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:38.664180 kubelet[2496]: I1008 20:05:38.664031 2496 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efc190af-8177-48b8-9f9d-1ae657fee9c7-lib-modules\") pod \"cilium-zjcc2\" (UID: \"efc190af-8177-48b8-9f9d-1ae657fee9c7\") " pod="kube-system/cilium-zjcc2" Oct 8 20:05:39.418378 kubelet[2496]: E1008 20:05:39.418321 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:39.419092 containerd[1462]: time="2024-10-08T20:05:39.418873016Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zjcc2,Uid:efc190af-8177-48b8-9f9d-1ae657fee9c7,Namespace:kube-system,Attempt:0,}" Oct 8 20:05:40.123798 containerd[1462]: time="2024-10-08T20:05:40.123188086Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 8 20:05:40.123954 containerd[1462]: time="2024-10-08T20:05:40.123772045Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 8 20:05:40.123954 containerd[1462]: time="2024-10-08T20:05:40.123786271Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:05:40.123954 containerd[1462]: time="2024-10-08T20:05:40.123867365Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 8 20:05:40.147630 systemd[1]: Started cri-containerd-e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f.scope - libcontainer container e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f. Oct 8 20:05:40.168233 containerd[1462]: time="2024-10-08T20:05:40.168177975Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-zjcc2,Uid:efc190af-8177-48b8-9f9d-1ae657fee9c7,Namespace:kube-system,Attempt:0,} returns sandbox id \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\"" Oct 8 20:05:40.168847 kubelet[2496]: E1008 20:05:40.168823 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:40.170863 containerd[1462]: time="2024-10-08T20:05:40.170829156Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Oct 8 20:05:40.417349 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount78671455.mount: Deactivated successfully. Oct 8 20:05:40.932347 kubelet[2496]: I1008 20:05:40.932302 2496 setters.go:600] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-10-08T20:05:40Z","lastTransitionTime":"2024-10-08T20:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Oct 8 20:05:41.130911 containerd[1462]: time="2024-10-08T20:05:41.130767229Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a\"" Oct 8 20:05:41.132817 containerd[1462]: time="2024-10-08T20:05:41.132723661Z" level=info msg="StartContainer for \"e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a\"" Oct 8 20:05:41.166816 systemd[1]: Started cri-containerd-e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a.scope - libcontainer container e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a. Oct 8 20:05:41.414523 systemd[1]: cri-containerd-e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a.scope: Deactivated successfully. Oct 8 20:05:41.476980 containerd[1462]: time="2024-10-08T20:05:41.476920907Z" level=info msg="StartContainer for \"e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a\" returns successfully" Oct 8 20:05:41.500363 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a-rootfs.mount: Deactivated successfully. Oct 8 20:05:41.668274 kubelet[2496]: E1008 20:05:41.668142 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:42.162438 containerd[1462]: time="2024-10-08T20:05:42.162367220Z" level=info msg="shim disconnected" id=e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a namespace=k8s.io Oct 8 20:05:42.162438 containerd[1462]: time="2024-10-08T20:05:42.162438305Z" level=warning msg="cleaning up after shim disconnected" id=e41aab3380b93e4c3c0365b02fb1e1dbf86229142b2ddd7a3139884aaf06aa6a namespace=k8s.io Oct 8 20:05:42.162438 containerd[1462]: time="2024-10-08T20:05:42.162446871Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:42.671996 kubelet[2496]: E1008 20:05:42.671949 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:42.674403 containerd[1462]: time="2024-10-08T20:05:42.674359970Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Oct 8 20:05:42.693696 containerd[1462]: time="2024-10-08T20:05:42.693633467Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377\"" Oct 8 20:05:42.694777 containerd[1462]: time="2024-10-08T20:05:42.694422514Z" level=info msg="StartContainer for \"80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377\"" Oct 8 20:05:42.730641 systemd[1]: Started cri-containerd-80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377.scope - libcontainer container 80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377. Oct 8 20:05:42.760630 containerd[1462]: time="2024-10-08T20:05:42.760568605Z" level=info msg="StartContainer for \"80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377\" returns successfully" Oct 8 20:05:42.766782 systemd[1]: cri-containerd-80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377.scope: Deactivated successfully. Oct 8 20:05:42.789091 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377-rootfs.mount: Deactivated successfully. Oct 8 20:05:43.002114 containerd[1462]: time="2024-10-08T20:05:43.001944279Z" level=info msg="shim disconnected" id=80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377 namespace=k8s.io Oct 8 20:05:43.002114 containerd[1462]: time="2024-10-08T20:05:43.002023047Z" level=warning msg="cleaning up after shim disconnected" id=80f1a074d32026711a764cd52e6a30f9c9eea139a6a4384f17a6bf69cc0aa377 namespace=k8s.io Oct 8 20:05:43.002114 containerd[1462]: time="2024-10-08T20:05:43.002037896Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:43.490694 kubelet[2496]: E1008 20:05:43.490635 2496 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Oct 8 20:05:43.674845 kubelet[2496]: E1008 20:05:43.674810 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:43.676858 containerd[1462]: time="2024-10-08T20:05:43.676812639Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Oct 8 20:05:44.137971 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3864033542.mount: Deactivated successfully. Oct 8 20:05:44.567730 containerd[1462]: time="2024-10-08T20:05:44.567582533Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5\"" Oct 8 20:05:44.568532 containerd[1462]: time="2024-10-08T20:05:44.568486727Z" level=info msg="StartContainer for \"87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5\"" Oct 8 20:05:44.598621 systemd[1]: Started cri-containerd-87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5.scope - libcontainer container 87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5. Oct 8 20:05:44.728912 systemd[1]: cri-containerd-87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5.scope: Deactivated successfully. Oct 8 20:05:44.773268 containerd[1462]: time="2024-10-08T20:05:44.773192577Z" level=info msg="StartContainer for \"87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5\" returns successfully" Oct 8 20:05:44.792775 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5-rootfs.mount: Deactivated successfully. Oct 8 20:05:45.193714 containerd[1462]: time="2024-10-08T20:05:45.193647916Z" level=info msg="shim disconnected" id=87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5 namespace=k8s.io Oct 8 20:05:45.193714 containerd[1462]: time="2024-10-08T20:05:45.193708891Z" level=warning msg="cleaning up after shim disconnected" id=87f44a10fa63f37d95eff48da47345f2e6ddd5b308a5de90e3995a0943aaaaa5 namespace=k8s.io Oct 8 20:05:45.193714 containerd[1462]: time="2024-10-08T20:05:45.193722076Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:45.780636 kubelet[2496]: E1008 20:05:45.780603 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:45.782218 containerd[1462]: time="2024-10-08T20:05:45.782183033Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Oct 8 20:05:46.361776 containerd[1462]: time="2024-10-08T20:05:46.361664998Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4\"" Oct 8 20:05:46.362513 containerd[1462]: time="2024-10-08T20:05:46.362358281Z" level=info msg="StartContainer for \"61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4\"" Oct 8 20:05:46.393685 systemd[1]: Started cri-containerd-61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4.scope - libcontainer container 61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4. Oct 8 20:05:46.418546 systemd[1]: cri-containerd-61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4.scope: Deactivated successfully. Oct 8 20:05:46.428385 kubelet[2496]: E1008 20:05:46.428354 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:46.472408 containerd[1462]: time="2024-10-08T20:05:46.472361917Z" level=info msg="StartContainer for \"61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4\" returns successfully" Oct 8 20:05:46.489983 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4-rootfs.mount: Deactivated successfully. Oct 8 20:05:46.712055 containerd[1462]: time="2024-10-08T20:05:46.711878711Z" level=info msg="shim disconnected" id=61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4 namespace=k8s.io Oct 8 20:05:46.712055 containerd[1462]: time="2024-10-08T20:05:46.711946419Z" level=warning msg="cleaning up after shim disconnected" id=61f001e13bbcf206970bb2de5aec9443f00abf3d0262b0ebe2d853ee105accf4 namespace=k8s.io Oct 8 20:05:46.712055 containerd[1462]: time="2024-10-08T20:05:46.711954936Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 8 20:05:46.829842 kubelet[2496]: E1008 20:05:46.829812 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:46.831679 containerd[1462]: time="2024-10-08T20:05:46.831628808Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Oct 8 20:05:47.153945 containerd[1462]: time="2024-10-08T20:05:47.153871124Z" level=info msg="CreateContainer within sandbox \"e99dfc54c4f752597a816a3116f4262225543b1153cb21fa5aeed4e56f3d233f\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"e878564b8b7a7bfb02a7685031e2c47abfeb0384fc724119026d37a91a6d2baf\"" Oct 8 20:05:47.154594 containerd[1462]: time="2024-10-08T20:05:47.154556152Z" level=info msg="StartContainer for \"e878564b8b7a7bfb02a7685031e2c47abfeb0384fc724119026d37a91a6d2baf\"" Oct 8 20:05:47.190685 systemd[1]: Started cri-containerd-e878564b8b7a7bfb02a7685031e2c47abfeb0384fc724119026d37a91a6d2baf.scope - libcontainer container e878564b8b7a7bfb02a7685031e2c47abfeb0384fc724119026d37a91a6d2baf. Oct 8 20:05:47.266422 containerd[1462]: time="2024-10-08T20:05:47.266369271Z" level=info msg="StartContainer for \"e878564b8b7a7bfb02a7685031e2c47abfeb0384fc724119026d37a91a6d2baf\" returns successfully" Oct 8 20:05:47.697511 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Oct 8 20:05:47.835420 kubelet[2496]: E1008 20:05:47.835354 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:47.967238 kubelet[2496]: I1008 20:05:47.967032 2496 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-zjcc2" podStartSLOduration=9.967005273 podStartE2EDuration="9.967005273s" podCreationTimestamp="2024-10-08 20:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-08 20:05:47.966742454 +0000 UTC m=+99.627966408" watchObservedRunningTime="2024-10-08 20:05:47.967005273 +0000 UTC m=+99.628229227" Oct 8 20:05:48.428085 kubelet[2496]: E1008 20:05:48.428000 2496 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-6f6b679f8f-7dgnr" podUID="87be2043-a5cb-462b-a162-5c88cef9ac39" Oct 8 20:05:48.837669 kubelet[2496]: E1008 20:05:48.837623 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:49.839382 kubelet[2496]: E1008 20:05:49.839335 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:50.428235 kubelet[2496]: E1008 20:05:50.428200 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:51.036749 systemd-networkd[1389]: lxc_health: Link UP Oct 8 20:05:51.048646 systemd-networkd[1389]: lxc_health: Gained carrier Oct 8 20:05:51.421101 kubelet[2496]: E1008 20:05:51.421063 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:51.843561 kubelet[2496]: E1008 20:05:51.843518 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:52.845705 kubelet[2496]: E1008 20:05:52.845675 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:52.855719 systemd-networkd[1389]: lxc_health: Gained IPv6LL Oct 8 20:05:54.996912 kubelet[2496]: E1008 20:05:54.996157 2496 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:50738->127.0.0.1:43741: write tcp 127.0.0.1:50738->127.0.0.1:43741: write: broken pipe Oct 8 20:05:58.428196 kubelet[2496]: E1008 20:05:58.428140 2496 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Oct 8 20:05:59.181823 sshd[4360]: pam_unix(sshd:session): session closed for user core Oct 8 20:05:59.186684 systemd[1]: sshd@27-10.0.0.126:22-10.0.0.1:54802.service: Deactivated successfully. Oct 8 20:05:59.189150 systemd[1]: session-28.scope: Deactivated successfully. Oct 8 20:05:59.190177 systemd-logind[1448]: Session 28 logged out. Waiting for processes to exit. Oct 8 20:05:59.191948 systemd-logind[1448]: Removed session 28.