Nov 12 20:54:55.967816 kernel: Linux version 6.6.60-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Nov 12 16:20:46 -00 2024 Nov 12 20:54:55.967838 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c3abb3a2c1edae861df27d3f75f2daa0ffde49038bd42517f0a3aa15da59cfc7 Nov 12 20:54:55.967849 kernel: BIOS-provided physical RAM map: Nov 12 20:54:55.967856 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Nov 12 20:54:55.967862 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Nov 12 20:54:55.967868 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Nov 12 20:54:55.967875 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Nov 12 20:54:55.967883 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Nov 12 20:54:55.967891 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Nov 12 20:54:55.967903 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Nov 12 20:54:55.967926 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Nov 12 20:54:55.967935 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Nov 12 20:54:55.967944 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Nov 12 20:54:55.967952 kernel: NX (Execute Disable) protection: active Nov 12 20:54:55.967963 kernel: APIC: Static calls initialized Nov 12 20:54:55.967975 kernel: SMBIOS 2.8 present. Nov 12 20:54:55.967982 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Nov 12 20:54:55.967989 kernel: Hypervisor detected: KVM Nov 12 20:54:55.967996 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Nov 12 20:54:55.968002 kernel: kvm-clock: using sched offset of 2259093822 cycles Nov 12 20:54:55.968010 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Nov 12 20:54:55.968017 kernel: tsc: Detected 2794.744 MHz processor Nov 12 20:54:55.968024 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Nov 12 20:54:55.968031 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Nov 12 20:54:55.968040 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Nov 12 20:54:55.968047 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Nov 12 20:54:55.968054 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Nov 12 20:54:55.968061 kernel: Using GB pages for direct mapping Nov 12 20:54:55.968068 kernel: ACPI: Early table checksum verification disabled Nov 12 20:54:55.968075 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Nov 12 20:54:55.968082 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968089 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968096 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968105 kernel: ACPI: FACS 0x000000009CFE0000 000040 Nov 12 20:54:55.968112 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968119 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968126 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968133 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Nov 12 20:54:55.968139 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db] Nov 12 20:54:55.968147 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7] Nov 12 20:54:55.968157 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Nov 12 20:54:55.968167 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b] Nov 12 20:54:55.968174 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3] Nov 12 20:54:55.968182 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df] Nov 12 20:54:55.968189 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407] Nov 12 20:54:55.968196 kernel: No NUMA configuration found Nov 12 20:54:55.968203 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Nov 12 20:54:55.968212 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Nov 12 20:54:55.968220 kernel: Zone ranges: Nov 12 20:54:55.968227 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Nov 12 20:54:55.968234 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Nov 12 20:54:55.968241 kernel: Normal empty Nov 12 20:54:55.968248 kernel: Movable zone start for each node Nov 12 20:54:55.968255 kernel: Early memory node ranges Nov 12 20:54:55.968262 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Nov 12 20:54:55.968269 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Nov 12 20:54:55.968276 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Nov 12 20:54:55.968286 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Nov 12 20:54:55.968293 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Nov 12 20:54:55.968300 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Nov 12 20:54:55.968307 kernel: ACPI: PM-Timer IO Port: 0x608 Nov 12 20:54:55.968314 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Nov 12 20:54:55.968322 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Nov 12 20:54:55.968329 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Nov 12 20:54:55.968336 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Nov 12 20:54:55.968343 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Nov 12 20:54:55.968352 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Nov 12 20:54:55.968359 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Nov 12 20:54:55.968366 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Nov 12 20:54:55.968373 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Nov 12 20:54:55.968380 kernel: TSC deadline timer available Nov 12 20:54:55.968388 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Nov 12 20:54:55.968395 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Nov 12 20:54:55.968410 kernel: kvm-guest: KVM setup pv remote TLB flush Nov 12 20:54:55.968418 kernel: kvm-guest: setup PV sched yield Nov 12 20:54:55.968427 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Nov 12 20:54:55.968434 kernel: Booting paravirtualized kernel on KVM Nov 12 20:54:55.968442 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Nov 12 20:54:55.968449 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Nov 12 20:54:55.968456 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Nov 12 20:54:55.968463 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Nov 12 20:54:55.968470 kernel: pcpu-alloc: [0] 0 1 2 3 Nov 12 20:54:55.968477 kernel: kvm-guest: PV spinlocks enabled Nov 12 20:54:55.968484 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Nov 12 20:54:55.968495 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c3abb3a2c1edae861df27d3f75f2daa0ffde49038bd42517f0a3aa15da59cfc7 Nov 12 20:54:55.968503 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Nov 12 20:54:55.968510 kernel: random: crng init done Nov 12 20:54:55.968517 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Nov 12 20:54:55.968524 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Nov 12 20:54:55.968531 kernel: Fallback order for Node 0: 0 Nov 12 20:54:55.968539 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Nov 12 20:54:55.968546 kernel: Policy zone: DMA32 Nov 12 20:54:55.968555 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Nov 12 20:54:55.968563 kernel: Memory: 2434592K/2571752K available (12288K kernel code, 2305K rwdata, 22724K rodata, 42828K init, 2360K bss, 136900K reserved, 0K cma-reserved) Nov 12 20:54:55.968570 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Nov 12 20:54:55.968577 kernel: ftrace: allocating 37799 entries in 148 pages Nov 12 20:54:55.968584 kernel: ftrace: allocated 148 pages with 3 groups Nov 12 20:54:55.968591 kernel: Dynamic Preempt: voluntary Nov 12 20:54:55.968598 kernel: rcu: Preemptible hierarchical RCU implementation. Nov 12 20:54:55.968611 kernel: rcu: RCU event tracing is enabled. Nov 12 20:54:55.968618 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Nov 12 20:54:55.968628 kernel: Trampoline variant of Tasks RCU enabled. Nov 12 20:54:55.968636 kernel: Rude variant of Tasks RCU enabled. Nov 12 20:54:55.968643 kernel: Tracing variant of Tasks RCU enabled. Nov 12 20:54:55.968650 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Nov 12 20:54:55.968657 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Nov 12 20:54:55.968664 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Nov 12 20:54:55.968672 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Nov 12 20:54:55.968679 kernel: Console: colour VGA+ 80x25 Nov 12 20:54:55.968689 kernel: printk: console [ttyS0] enabled Nov 12 20:54:55.968699 kernel: ACPI: Core revision 20230628 Nov 12 20:54:55.968712 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Nov 12 20:54:55.968722 kernel: APIC: Switch to symmetric I/O mode setup Nov 12 20:54:55.968732 kernel: x2apic enabled Nov 12 20:54:55.968742 kernel: APIC: Switched APIC routing to: physical x2apic Nov 12 20:54:55.968749 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Nov 12 20:54:55.968756 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Nov 12 20:54:55.968763 kernel: kvm-guest: setup PV IPIs Nov 12 20:54:55.968781 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Nov 12 20:54:55.968788 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Nov 12 20:54:55.968796 kernel: Calibrating delay loop (skipped) preset value.. 5589.48 BogoMIPS (lpj=2794744) Nov 12 20:54:55.968803 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Nov 12 20:54:55.968813 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Nov 12 20:54:55.968820 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Nov 12 20:54:55.968828 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Nov 12 20:54:55.968835 kernel: Spectre V2 : Mitigation: Retpolines Nov 12 20:54:55.968843 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Nov 12 20:54:55.968853 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Nov 12 20:54:55.968860 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Nov 12 20:54:55.968867 kernel: RETBleed: Mitigation: untrained return thunk Nov 12 20:54:55.968875 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Nov 12 20:54:55.968883 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Nov 12 20:54:55.968890 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Nov 12 20:54:55.968898 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Nov 12 20:54:55.968918 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Nov 12 20:54:55.968941 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Nov 12 20:54:55.968948 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Nov 12 20:54:55.968956 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Nov 12 20:54:55.968963 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Nov 12 20:54:55.968971 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Nov 12 20:54:55.968978 kernel: Freeing SMP alternatives memory: 32K Nov 12 20:54:55.968986 kernel: pid_max: default: 32768 minimum: 301 Nov 12 20:54:55.968993 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Nov 12 20:54:55.969000 kernel: landlock: Up and running. Nov 12 20:54:55.969010 kernel: SELinux: Initializing. Nov 12 20:54:55.969018 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 12 20:54:55.969025 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Nov 12 20:54:55.969033 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Nov 12 20:54:55.969041 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 12 20:54:55.969048 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 12 20:54:55.969056 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Nov 12 20:54:55.969063 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Nov 12 20:54:55.969071 kernel: ... version: 0 Nov 12 20:54:55.969080 kernel: ... bit width: 48 Nov 12 20:54:55.969088 kernel: ... generic registers: 6 Nov 12 20:54:55.969095 kernel: ... value mask: 0000ffffffffffff Nov 12 20:54:55.969103 kernel: ... max period: 00007fffffffffff Nov 12 20:54:55.969110 kernel: ... fixed-purpose events: 0 Nov 12 20:54:55.969118 kernel: ... event mask: 000000000000003f Nov 12 20:54:55.969125 kernel: signal: max sigframe size: 1776 Nov 12 20:54:55.969132 kernel: rcu: Hierarchical SRCU implementation. Nov 12 20:54:55.969140 kernel: rcu: Max phase no-delay instances is 400. Nov 12 20:54:55.969150 kernel: smp: Bringing up secondary CPUs ... Nov 12 20:54:55.969157 kernel: smpboot: x86: Booting SMP configuration: Nov 12 20:54:55.969165 kernel: .... node #0, CPUs: #1 #2 #3 Nov 12 20:54:55.969172 kernel: smp: Brought up 1 node, 4 CPUs Nov 12 20:54:55.969180 kernel: smpboot: Max logical packages: 1 Nov 12 20:54:55.969187 kernel: smpboot: Total of 4 processors activated (22357.95 BogoMIPS) Nov 12 20:54:55.969195 kernel: devtmpfs: initialized Nov 12 20:54:55.969202 kernel: x86/mm: Memory block size: 128MB Nov 12 20:54:55.969210 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Nov 12 20:54:55.969220 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Nov 12 20:54:55.969227 kernel: pinctrl core: initialized pinctrl subsystem Nov 12 20:54:55.969235 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Nov 12 20:54:55.969242 kernel: audit: initializing netlink subsys (disabled) Nov 12 20:54:55.969250 kernel: audit: type=2000 audit(1731444895.054:1): state=initialized audit_enabled=0 res=1 Nov 12 20:54:55.969257 kernel: thermal_sys: Registered thermal governor 'step_wise' Nov 12 20:54:55.969265 kernel: thermal_sys: Registered thermal governor 'user_space' Nov 12 20:54:55.969272 kernel: cpuidle: using governor menu Nov 12 20:54:55.969280 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Nov 12 20:54:55.969290 kernel: dca service started, version 1.12.1 Nov 12 20:54:55.969297 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Nov 12 20:54:55.969305 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Nov 12 20:54:55.969312 kernel: PCI: Using configuration type 1 for base access Nov 12 20:54:55.969322 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Nov 12 20:54:55.969330 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Nov 12 20:54:55.969338 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Nov 12 20:54:55.969346 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Nov 12 20:54:55.969356 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Nov 12 20:54:55.969367 kernel: ACPI: Added _OSI(Module Device) Nov 12 20:54:55.969376 kernel: ACPI: Added _OSI(Processor Device) Nov 12 20:54:55.969383 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Nov 12 20:54:55.969391 kernel: ACPI: Added _OSI(Processor Aggregator Device) Nov 12 20:54:55.969406 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Nov 12 20:54:55.969414 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Nov 12 20:54:55.969421 kernel: ACPI: Interpreter enabled Nov 12 20:54:55.969429 kernel: ACPI: PM: (supports S0 S3 S5) Nov 12 20:54:55.969437 kernel: ACPI: Using IOAPIC for interrupt routing Nov 12 20:54:55.969447 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Nov 12 20:54:55.969454 kernel: PCI: Using E820 reservations for host bridge windows Nov 12 20:54:55.969462 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Nov 12 20:54:55.969469 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Nov 12 20:54:55.969653 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Nov 12 20:54:55.969783 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Nov 12 20:54:55.969961 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Nov 12 20:54:55.969975 kernel: PCI host bridge to bus 0000:00 Nov 12 20:54:55.970148 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Nov 12 20:54:55.970262 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Nov 12 20:54:55.970375 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Nov 12 20:54:55.970494 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Nov 12 20:54:55.970603 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Nov 12 20:54:55.970712 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Nov 12 20:54:55.970827 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Nov 12 20:54:55.970984 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Nov 12 20:54:55.971115 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Nov 12 20:54:55.971236 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Nov 12 20:54:55.971446 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Nov 12 20:54:55.971616 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Nov 12 20:54:55.971787 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Nov 12 20:54:55.971943 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Nov 12 20:54:55.972068 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Nov 12 20:54:55.972189 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Nov 12 20:54:55.972318 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Nov 12 20:54:55.972491 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Nov 12 20:54:55.972615 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Nov 12 20:54:55.972736 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Nov 12 20:54:55.972862 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Nov 12 20:54:55.973063 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Nov 12 20:54:55.973195 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Nov 12 20:54:55.973319 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Nov 12 20:54:55.973451 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Nov 12 20:54:55.973572 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Nov 12 20:54:55.973699 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Nov 12 20:54:55.973824 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Nov 12 20:54:55.973998 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Nov 12 20:54:55.974123 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Nov 12 20:54:55.974242 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Nov 12 20:54:55.974373 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Nov 12 20:54:55.974504 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Nov 12 20:54:55.974519 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Nov 12 20:54:55.974527 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Nov 12 20:54:55.974535 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Nov 12 20:54:55.974542 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Nov 12 20:54:55.974550 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Nov 12 20:54:55.974557 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Nov 12 20:54:55.974565 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Nov 12 20:54:55.974572 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Nov 12 20:54:55.974580 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Nov 12 20:54:55.974590 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Nov 12 20:54:55.974598 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Nov 12 20:54:55.974605 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Nov 12 20:54:55.974613 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Nov 12 20:54:55.974620 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Nov 12 20:54:55.974628 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Nov 12 20:54:55.974635 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Nov 12 20:54:55.974643 kernel: iommu: Default domain type: Translated Nov 12 20:54:55.974650 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Nov 12 20:54:55.974660 kernel: PCI: Using ACPI for IRQ routing Nov 12 20:54:55.974668 kernel: PCI: pci_cache_line_size set to 64 bytes Nov 12 20:54:55.974675 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Nov 12 20:54:55.974683 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Nov 12 20:54:55.974802 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Nov 12 20:54:55.974934 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Nov 12 20:54:55.975056 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Nov 12 20:54:55.975066 kernel: vgaarb: loaded Nov 12 20:54:55.975078 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Nov 12 20:54:55.975085 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Nov 12 20:54:55.975093 kernel: clocksource: Switched to clocksource kvm-clock Nov 12 20:54:55.975101 kernel: VFS: Disk quotas dquot_6.6.0 Nov 12 20:54:55.975109 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Nov 12 20:54:55.975116 kernel: pnp: PnP ACPI init Nov 12 20:54:55.975249 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Nov 12 20:54:55.975260 kernel: pnp: PnP ACPI: found 6 devices Nov 12 20:54:55.975269 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Nov 12 20:54:55.975279 kernel: NET: Registered PF_INET protocol family Nov 12 20:54:55.975287 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Nov 12 20:54:55.975295 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Nov 12 20:54:55.975303 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Nov 12 20:54:55.975311 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Nov 12 20:54:55.975319 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Nov 12 20:54:55.975327 kernel: TCP: Hash tables configured (established 32768 bind 32768) Nov 12 20:54:55.975335 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 12 20:54:55.975346 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Nov 12 20:54:55.975354 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Nov 12 20:54:55.975362 kernel: NET: Registered PF_XDP protocol family Nov 12 20:54:55.975485 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Nov 12 20:54:55.975596 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Nov 12 20:54:55.975705 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Nov 12 20:54:55.975826 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Nov 12 20:54:55.976043 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Nov 12 20:54:55.976188 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Nov 12 20:54:55.976209 kernel: PCI: CLS 0 bytes, default 64 Nov 12 20:54:55.976220 kernel: Initialise system trusted keyrings Nov 12 20:54:55.976230 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Nov 12 20:54:55.976241 kernel: Key type asymmetric registered Nov 12 20:54:55.976252 kernel: Asymmetric key parser 'x509' registered Nov 12 20:54:55.976263 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Nov 12 20:54:55.976274 kernel: io scheduler mq-deadline registered Nov 12 20:54:55.976286 kernel: io scheduler kyber registered Nov 12 20:54:55.976298 kernel: io scheduler bfq registered Nov 12 20:54:55.976317 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Nov 12 20:54:55.976330 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Nov 12 20:54:55.976343 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Nov 12 20:54:55.976355 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Nov 12 20:54:55.976368 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Nov 12 20:54:55.976382 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Nov 12 20:54:55.976408 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Nov 12 20:54:55.976422 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Nov 12 20:54:55.976435 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Nov 12 20:54:55.976615 kernel: rtc_cmos 00:04: RTC can wake from S4 Nov 12 20:54:55.976767 kernel: rtc_cmos 00:04: registered as rtc0 Nov 12 20:54:55.976785 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Nov 12 20:54:55.977029 kernel: rtc_cmos 00:04: setting system clock to 2024-11-12T20:54:55 UTC (1731444895) Nov 12 20:54:55.977183 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Nov 12 20:54:55.977199 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Nov 12 20:54:55.977212 kernel: NET: Registered PF_INET6 protocol family Nov 12 20:54:55.977225 kernel: Segment Routing with IPv6 Nov 12 20:54:55.977245 kernel: In-situ OAM (IOAM) with IPv6 Nov 12 20:54:55.977255 kernel: NET: Registered PF_PACKET protocol family Nov 12 20:54:55.977266 kernel: Key type dns_resolver registered Nov 12 20:54:55.977279 kernel: IPI shorthand broadcast: enabled Nov 12 20:54:55.977292 kernel: sched_clock: Marking stable (604004816, 110013072)->(933773352, -219755464) Nov 12 20:54:55.977306 kernel: registered taskstats version 1 Nov 12 20:54:55.977318 kernel: Loading compiled-in X.509 certificates Nov 12 20:54:55.977332 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.60-flatcar: 0473a73d840db5324524af106a53c13fc6fc218a' Nov 12 20:54:55.977344 kernel: Key type .fscrypt registered Nov 12 20:54:55.977360 kernel: Key type fscrypt-provisioning registered Nov 12 20:54:55.977370 kernel: ima: No TPM chip found, activating TPM-bypass! Nov 12 20:54:55.977381 kernel: ima: Allocated hash algorithm: sha1 Nov 12 20:54:55.977391 kernel: ima: No architecture policies found Nov 12 20:54:55.977410 kernel: clk: Disabling unused clocks Nov 12 20:54:55.977420 kernel: Freeing unused kernel image (initmem) memory: 42828K Nov 12 20:54:55.977430 kernel: Write protecting the kernel read-only data: 36864k Nov 12 20:54:55.977440 kernel: Freeing unused kernel image (rodata/data gap) memory: 1852K Nov 12 20:54:55.977454 kernel: Run /init as init process Nov 12 20:54:55.977465 kernel: with arguments: Nov 12 20:54:55.977475 kernel: /init Nov 12 20:54:55.977485 kernel: with environment: Nov 12 20:54:55.977495 kernel: HOME=/ Nov 12 20:54:55.977505 kernel: TERM=linux Nov 12 20:54:55.977515 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Nov 12 20:54:55.977528 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Nov 12 20:54:55.977545 systemd[1]: Detected virtualization kvm. Nov 12 20:54:55.977556 systemd[1]: Detected architecture x86-64. Nov 12 20:54:55.977567 systemd[1]: Running in initrd. Nov 12 20:54:55.977578 systemd[1]: No hostname configured, using default hostname. Nov 12 20:54:55.977588 systemd[1]: Hostname set to . Nov 12 20:54:55.977599 systemd[1]: Initializing machine ID from VM UUID. Nov 12 20:54:55.977610 systemd[1]: Queued start job for default target initrd.target. Nov 12 20:54:55.977622 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 12 20:54:55.977636 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 12 20:54:55.977649 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Nov 12 20:54:55.977676 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 12 20:54:55.977690 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Nov 12 20:54:55.977701 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Nov 12 20:54:55.977718 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Nov 12 20:54:55.977730 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Nov 12 20:54:55.977742 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 12 20:54:55.977754 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 12 20:54:55.977765 systemd[1]: Reached target paths.target - Path Units. Nov 12 20:54:55.977776 systemd[1]: Reached target slices.target - Slice Units. Nov 12 20:54:55.977788 systemd[1]: Reached target swap.target - Swaps. Nov 12 20:54:55.977799 systemd[1]: Reached target timers.target - Timer Units. Nov 12 20:54:55.977814 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Nov 12 20:54:55.977825 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 12 20:54:55.977836 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Nov 12 20:54:55.977848 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Nov 12 20:54:55.977858 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 12 20:54:55.977870 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 12 20:54:55.977882 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 12 20:54:55.977893 systemd[1]: Reached target sockets.target - Socket Units. Nov 12 20:54:55.977920 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Nov 12 20:54:55.977949 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 12 20:54:55.977961 systemd[1]: Finished network-cleanup.service - Network Cleanup. Nov 12 20:54:55.977971 systemd[1]: Starting systemd-fsck-usr.service... Nov 12 20:54:55.977982 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 12 20:54:55.977994 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 12 20:54:55.978005 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 12 20:54:55.978017 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Nov 12 20:54:55.978028 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 12 20:54:55.978039 systemd[1]: Finished systemd-fsck-usr.service. Nov 12 20:54:55.978055 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Nov 12 20:54:55.978092 systemd-journald[192]: Collecting audit messages is disabled. Nov 12 20:54:55.978120 systemd-journald[192]: Journal started Nov 12 20:54:55.978146 systemd-journald[192]: Runtime Journal (/run/log/journal/b29ddd8cc6a24005a4ffb090905e37d8) is 6.0M, max 48.4M, 42.3M free. Nov 12 20:54:55.971773 systemd-modules-load[193]: Inserted module 'overlay' Nov 12 20:54:56.012930 systemd[1]: Started systemd-journald.service - Journal Service. Nov 12 20:54:56.012962 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Nov 12 20:54:56.012974 kernel: Bridge firewalling registered Nov 12 20:54:55.999504 systemd-modules-load[193]: Inserted module 'br_netfilter' Nov 12 20:54:56.013137 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 12 20:54:56.013790 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Nov 12 20:54:56.029227 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 12 20:54:56.031279 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 12 20:54:56.034313 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 12 20:54:56.035770 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 12 20:54:56.043051 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 12 20:54:56.045973 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 12 20:54:56.048343 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 12 20:54:56.051172 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 12 20:54:56.063112 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 12 20:54:56.065956 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 12 20:54:56.068341 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Nov 12 20:54:56.084147 dracut-cmdline[230]: dracut-dracut-053 Nov 12 20:54:56.086927 dracut-cmdline[230]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=c3abb3a2c1edae861df27d3f75f2daa0ffde49038bd42517f0a3aa15da59cfc7 Nov 12 20:54:56.103839 systemd-resolved[221]: Positive Trust Anchors: Nov 12 20:54:56.103878 systemd-resolved[221]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 12 20:54:56.103934 systemd-resolved[221]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 12 20:54:56.107443 systemd-resolved[221]: Defaulting to hostname 'linux'. Nov 12 20:54:56.108877 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 12 20:54:56.114742 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 12 20:54:56.172975 kernel: SCSI subsystem initialized Nov 12 20:54:56.183000 kernel: Loading iSCSI transport class v2.0-870. Nov 12 20:54:56.194949 kernel: iscsi: registered transport (tcp) Nov 12 20:54:56.216953 kernel: iscsi: registered transport (qla4xxx) Nov 12 20:54:56.217025 kernel: QLogic iSCSI HBA Driver Nov 12 20:54:56.275046 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Nov 12 20:54:56.283207 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Nov 12 20:54:56.311822 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Nov 12 20:54:56.311902 kernel: device-mapper: uevent: version 1.0.3 Nov 12 20:54:56.311924 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Nov 12 20:54:56.359956 kernel: raid6: avx2x4 gen() 22161 MB/s Nov 12 20:54:56.376949 kernel: raid6: avx2x2 gen() 25549 MB/s Nov 12 20:54:56.394124 kernel: raid6: avx2x1 gen() 23740 MB/s Nov 12 20:54:56.394201 kernel: raid6: using algorithm avx2x2 gen() 25549 MB/s Nov 12 20:54:56.412107 kernel: raid6: .... xor() 17807 MB/s, rmw enabled Nov 12 20:54:56.412161 kernel: raid6: using avx2x2 recovery algorithm Nov 12 20:54:56.433959 kernel: xor: automatically using best checksumming function avx Nov 12 20:54:56.639959 kernel: Btrfs loaded, zoned=no, fsverity=no Nov 12 20:54:56.654496 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Nov 12 20:54:56.666118 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 12 20:54:56.724831 systemd-udevd[412]: Using default interface naming scheme 'v255'. Nov 12 20:54:56.729694 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 12 20:54:56.790095 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Nov 12 20:54:56.805425 dracut-pre-trigger[423]: rd.md=0: removing MD RAID activation Nov 12 20:54:56.841000 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Nov 12 20:54:56.873224 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 12 20:54:56.936299 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 12 20:54:56.948314 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Nov 12 20:54:56.962384 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Nov 12 20:54:56.963550 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Nov 12 20:54:56.964881 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 12 20:54:56.965411 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 12 20:54:56.981986 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Nov 12 20:54:57.008723 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Nov 12 20:54:57.009308 kernel: cryptd: max_cpu_qlen set to 1000 Nov 12 20:54:57.009322 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Nov 12 20:54:57.009333 kernel: GPT:9289727 != 19775487 Nov 12 20:54:57.009351 kernel: GPT:Alternate GPT header not at the end of the disk. Nov 12 20:54:57.009361 kernel: GPT:9289727 != 19775487 Nov 12 20:54:57.009378 kernel: GPT: Use GNU Parted to correct GPT errors. Nov 12 20:54:57.009389 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 12 20:54:57.009399 kernel: libata version 3.00 loaded. Nov 12 20:54:56.988183 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Nov 12 20:54:57.000515 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Nov 12 20:54:57.043438 kernel: ahci 0000:00:1f.2: version 3.0 Nov 12 20:54:57.058511 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Nov 12 20:54:57.058534 kernel: AVX2 version of gcm_enc/dec engaged. Nov 12 20:54:57.058548 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Nov 12 20:54:57.058738 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Nov 12 20:54:57.058973 kernel: AES CTR mode by8 optimization enabled Nov 12 20:54:57.058990 kernel: scsi host0: ahci Nov 12 20:54:57.059182 kernel: scsi host1: ahci Nov 12 20:54:57.059364 kernel: scsi host2: ahci Nov 12 20:54:57.059551 kernel: scsi host3: ahci Nov 12 20:54:57.059728 kernel: scsi host4: ahci Nov 12 20:54:57.059959 kernel: scsi host5: ahci Nov 12 20:54:57.060139 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Nov 12 20:54:57.060158 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Nov 12 20:54:57.060172 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Nov 12 20:54:57.060186 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Nov 12 20:54:57.060199 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Nov 12 20:54:57.060213 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Nov 12 20:54:57.047473 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 12 20:54:57.047585 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 12 20:54:57.049592 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 12 20:54:57.052416 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 12 20:54:57.130805 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (474) Nov 12 20:54:57.052620 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 12 20:54:57.058756 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Nov 12 20:54:57.136253 kernel: BTRFS: device fsid 9dfeafbb-8ab7-4be2-acae-f51db463fc77 devid 1 transid 37 /dev/vda3 scanned by (udev-worker) (464) Nov 12 20:54:57.131647 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 12 20:54:57.158961 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Nov 12 20:54:57.170262 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Nov 12 20:54:57.222148 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 12 20:54:57.222868 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 12 20:54:57.229714 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Nov 12 20:54:57.229990 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Nov 12 20:54:57.250052 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Nov 12 20:54:57.251320 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Nov 12 20:54:57.286225 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 12 20:54:57.372955 kernel: ata2: SATA link down (SStatus 0 SControl 300) Nov 12 20:54:57.373025 kernel: ata5: SATA link down (SStatus 0 SControl 300) Nov 12 20:54:57.374958 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Nov 12 20:54:57.375032 kernel: ata4: SATA link down (SStatus 0 SControl 300) Nov 12 20:54:57.375930 kernel: ata1: SATA link down (SStatus 0 SControl 300) Nov 12 20:54:57.376937 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Nov 12 20:54:57.378045 kernel: ata3.00: applying bridge limits Nov 12 20:54:57.378075 kernel: ata3.00: configured for UDMA/100 Nov 12 20:54:57.378940 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Nov 12 20:54:57.411957 disk-uuid[555]: Primary Header is updated. Nov 12 20:54:57.411957 disk-uuid[555]: Secondary Entries is updated. Nov 12 20:54:57.411957 disk-uuid[555]: Secondary Header is updated. Nov 12 20:54:57.416284 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 12 20:54:57.419938 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 12 20:54:57.432519 kernel: ata6: SATA link down (SStatus 0 SControl 300) Nov 12 20:54:57.461939 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Nov 12 20:54:57.476620 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Nov 12 20:54:57.476644 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Nov 12 20:54:58.478978 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Nov 12 20:54:58.479397 disk-uuid[564]: The operation has completed successfully. Nov 12 20:54:58.504829 systemd[1]: disk-uuid.service: Deactivated successfully. Nov 12 20:54:58.504970 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Nov 12 20:54:58.540167 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Nov 12 20:54:58.543600 sh[591]: Success Nov 12 20:54:58.557107 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Nov 12 20:54:58.592873 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Nov 12 20:54:58.607454 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Nov 12 20:54:58.612092 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Nov 12 20:54:58.641188 kernel: BTRFS info (device dm-0): first mount of filesystem 9dfeafbb-8ab7-4be2-acae-f51db463fc77 Nov 12 20:54:58.641266 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Nov 12 20:54:58.641278 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Nov 12 20:54:58.642242 kernel: BTRFS info (device dm-0): disabling log replay at mount time Nov 12 20:54:58.642990 kernel: BTRFS info (device dm-0): using free space tree Nov 12 20:54:58.648311 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Nov 12 20:54:58.650049 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Nov 12 20:54:58.657185 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Nov 12 20:54:58.658938 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Nov 12 20:54:58.665938 kernel: BTRFS info (device vda6): first mount of filesystem bdc43ff2-e8de-475f-88ba-e8c26a6bbaa6 Nov 12 20:54:58.665960 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 12 20:54:58.667416 kernel: BTRFS info (device vda6): using free space tree Nov 12 20:54:58.669974 kernel: BTRFS info (device vda6): auto enabling async discard Nov 12 20:54:58.679319 systemd[1]: mnt-oem.mount: Deactivated successfully. Nov 12 20:54:58.681251 kernel: BTRFS info (device vda6): last unmount of filesystem bdc43ff2-e8de-475f-88ba-e8c26a6bbaa6 Nov 12 20:54:58.772284 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 12 20:54:58.792037 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 12 20:54:58.816119 systemd-networkd[769]: lo: Link UP Nov 12 20:54:58.816127 systemd-networkd[769]: lo: Gained carrier Nov 12 20:54:58.817719 systemd-networkd[769]: Enumeration completed Nov 12 20:54:58.818133 systemd-networkd[769]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 12 20:54:58.818136 systemd-networkd[769]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 12 20:54:58.821208 systemd-networkd[769]: eth0: Link UP Nov 12 20:54:58.821211 systemd-networkd[769]: eth0: Gained carrier Nov 12 20:54:58.821218 systemd-networkd[769]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 12 20:54:58.821982 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 12 20:54:58.823037 systemd[1]: Reached target network.target - Network. Nov 12 20:54:58.847982 systemd-networkd[769]: eth0: DHCPv4 address 10.0.0.145/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 12 20:54:59.117786 systemd[1]: Finished ignition-setup.service - Ignition (setup). Nov 12 20:54:59.135107 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Nov 12 20:54:59.203239 ignition[774]: Ignition 2.19.0 Nov 12 20:54:59.203251 ignition[774]: Stage: fetch-offline Nov 12 20:54:59.203287 ignition[774]: no configs at "/usr/lib/ignition/base.d" Nov 12 20:54:59.203297 ignition[774]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:54:59.203390 ignition[774]: parsed url from cmdline: "" Nov 12 20:54:59.203394 ignition[774]: no config URL provided Nov 12 20:54:59.203399 ignition[774]: reading system config file "/usr/lib/ignition/user.ign" Nov 12 20:54:59.203408 ignition[774]: no config at "/usr/lib/ignition/user.ign" Nov 12 20:54:59.203433 ignition[774]: op(1): [started] loading QEMU firmware config module Nov 12 20:54:59.203438 ignition[774]: op(1): executing: "modprobe" "qemu_fw_cfg" Nov 12 20:54:59.236383 ignition[774]: op(1): [finished] loading QEMU firmware config module Nov 12 20:54:59.277342 ignition[774]: parsing config with SHA512: e12f2d5901983efb4b0dfb71b5d066d944ed17a1e288ad4118c8fefab5002d76c2c47c4449fb30d835d88c78bcf37259fd65ee00fa8b975ad93207fa3618fe4b Nov 12 20:54:59.281333 unknown[774]: fetched base config from "system" Nov 12 20:54:59.281351 unknown[774]: fetched user config from "qemu" Nov 12 20:54:59.281815 ignition[774]: fetch-offline: fetch-offline passed Nov 12 20:54:59.284232 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Nov 12 20:54:59.281891 ignition[774]: Ignition finished successfully Nov 12 20:54:59.285816 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Nov 12 20:54:59.293169 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Nov 12 20:54:59.305470 ignition[785]: Ignition 2.19.0 Nov 12 20:54:59.305485 ignition[785]: Stage: kargs Nov 12 20:54:59.305689 ignition[785]: no configs at "/usr/lib/ignition/base.d" Nov 12 20:54:59.305703 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:54:59.306625 ignition[785]: kargs: kargs passed Nov 12 20:54:59.309536 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Nov 12 20:54:59.306671 ignition[785]: Ignition finished successfully Nov 12 20:54:59.322096 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Nov 12 20:54:59.338379 ignition[793]: Ignition 2.19.0 Nov 12 20:54:59.348106 systemd[1]: Finished ignition-disks.service - Ignition (disks). Nov 12 20:54:59.338386 ignition[793]: Stage: disks Nov 12 20:54:59.350377 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Nov 12 20:54:59.338550 ignition[793]: no configs at "/usr/lib/ignition/base.d" Nov 12 20:54:59.352504 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Nov 12 20:54:59.338560 ignition[793]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:54:59.354272 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 12 20:54:59.339439 ignition[793]: disks: disks passed Nov 12 20:54:59.356354 systemd[1]: Reached target sysinit.target - System Initialization. Nov 12 20:54:59.339477 ignition[793]: Ignition finished successfully Nov 12 20:54:59.358249 systemd[1]: Reached target basic.target - Basic System. Nov 12 20:54:59.381135 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Nov 12 20:54:59.447335 systemd-fsck[804]: ROOT: clean, 14/553520 files, 52654/553472 blocks Nov 12 20:55:00.053199 systemd-networkd[769]: eth0: Gained IPv6LL Nov 12 20:55:00.062078 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Nov 12 20:55:00.071077 systemd[1]: Mounting sysroot.mount - /sysroot... Nov 12 20:55:00.187940 kernel: EXT4-fs (vda9): mounted filesystem cc5635ac-cac6-420e-b789-89e3a937cfb2 r/w with ordered data mode. Quota mode: none. Nov 12 20:55:00.188262 systemd[1]: Mounted sysroot.mount - /sysroot. Nov 12 20:55:00.189698 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Nov 12 20:55:00.201984 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 12 20:55:00.203810 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Nov 12 20:55:00.205278 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Nov 12 20:55:00.230392 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (812) Nov 12 20:55:00.230413 kernel: BTRFS info (device vda6): first mount of filesystem bdc43ff2-e8de-475f-88ba-e8c26a6bbaa6 Nov 12 20:55:00.230424 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 12 20:55:00.230435 kernel: BTRFS info (device vda6): using free space tree Nov 12 20:55:00.230445 kernel: BTRFS info (device vda6): auto enabling async discard Nov 12 20:55:00.205336 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Nov 12 20:55:00.205360 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Nov 12 20:55:00.225167 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Nov 12 20:55:00.231418 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 12 20:55:00.235273 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Nov 12 20:55:00.273229 initrd-setup-root[837]: cut: /sysroot/etc/passwd: No such file or directory Nov 12 20:55:00.280444 initrd-setup-root[844]: cut: /sysroot/etc/group: No such file or directory Nov 12 20:55:00.284151 initrd-setup-root[851]: cut: /sysroot/etc/shadow: No such file or directory Nov 12 20:55:00.287940 initrd-setup-root[858]: cut: /sysroot/etc/gshadow: No such file or directory Nov 12 20:55:00.373732 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Nov 12 20:55:00.391997 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Nov 12 20:55:00.393604 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Nov 12 20:55:00.400454 systemd[1]: sysroot-oem.mount: Deactivated successfully. Nov 12 20:55:00.401922 kernel: BTRFS info (device vda6): last unmount of filesystem bdc43ff2-e8de-475f-88ba-e8c26a6bbaa6 Nov 12 20:55:00.441709 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Nov 12 20:55:00.589115 ignition[931]: INFO : Ignition 2.19.0 Nov 12 20:55:00.589115 ignition[931]: INFO : Stage: mount Nov 12 20:55:00.590789 ignition[931]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 12 20:55:00.590789 ignition[931]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:55:00.590789 ignition[931]: INFO : mount: mount passed Nov 12 20:55:00.590789 ignition[931]: INFO : Ignition finished successfully Nov 12 20:55:00.596416 systemd[1]: Finished ignition-mount.service - Ignition (mount). Nov 12 20:55:00.608027 systemd[1]: Starting ignition-files.service - Ignition (files)... Nov 12 20:55:01.201068 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Nov 12 20:55:01.257939 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (941) Nov 12 20:55:01.257976 kernel: BTRFS info (device vda6): first mount of filesystem bdc43ff2-e8de-475f-88ba-e8c26a6bbaa6 Nov 12 20:55:01.259720 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Nov 12 20:55:01.259740 kernel: BTRFS info (device vda6): using free space tree Nov 12 20:55:01.262940 kernel: BTRFS info (device vda6): auto enabling async discard Nov 12 20:55:01.264409 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Nov 12 20:55:01.291363 ignition[959]: INFO : Ignition 2.19.0 Nov 12 20:55:01.291363 ignition[959]: INFO : Stage: files Nov 12 20:55:01.319076 ignition[959]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 12 20:55:01.319076 ignition[959]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:55:01.319076 ignition[959]: DEBUG : files: compiled without relabeling support, skipping Nov 12 20:55:01.323412 ignition[959]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Nov 12 20:55:01.323412 ignition[959]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Nov 12 20:55:01.328400 ignition[959]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Nov 12 20:55:01.330066 ignition[959]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Nov 12 20:55:01.332068 unknown[959]: wrote ssh authorized keys file for user: core Nov 12 20:55:01.333248 ignition[959]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Nov 12 20:55:01.335860 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Nov 12 20:55:01.338304 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Nov 12 20:55:01.376928 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Nov 12 20:55:01.471890 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Nov 12 20:55:01.471890 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Nov 12 20:55:01.475935 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Nov 12 20:55:01.922509 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Nov 12 20:55:02.022753 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Nov 12 20:55:02.024705 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Nov 12 20:55:02.326948 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Nov 12 20:55:03.323885 ignition[959]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Nov 12 20:55:03.323885 ignition[959]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Nov 12 20:55:03.328242 ignition[959]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Nov 12 20:55:03.369186 ignition[959]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Nov 12 20:55:03.375407 ignition[959]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Nov 12 20:55:03.377510 ignition[959]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Nov 12 20:55:03.377510 ignition[959]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Nov 12 20:55:03.380445 ignition[959]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Nov 12 20:55:03.382014 ignition[959]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Nov 12 20:55:03.383867 ignition[959]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Nov 12 20:55:03.385726 ignition[959]: INFO : files: files passed Nov 12 20:55:03.386479 ignition[959]: INFO : Ignition finished successfully Nov 12 20:55:03.389952 systemd[1]: Finished ignition-files.service - Ignition (files). Nov 12 20:55:03.399051 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Nov 12 20:55:03.401683 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Nov 12 20:55:03.404733 systemd[1]: ignition-quench.service: Deactivated successfully. Nov 12 20:55:03.404852 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Nov 12 20:55:03.416191 initrd-setup-root-after-ignition[986]: grep: /sysroot/oem/oem-release: No such file or directory Nov 12 20:55:03.420369 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 12 20:55:03.420369 initrd-setup-root-after-ignition[988]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Nov 12 20:55:03.423651 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Nov 12 20:55:03.424780 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 12 20:55:03.426655 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Nov 12 20:55:03.435032 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Nov 12 20:55:03.459687 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Nov 12 20:55:03.459818 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Nov 12 20:55:03.460613 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Nov 12 20:55:03.464117 systemd[1]: Reached target initrd.target - Initrd Default Target. Nov 12 20:55:03.464492 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Nov 12 20:55:03.465258 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Nov 12 20:55:03.483631 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 12 20:55:03.496034 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Nov 12 20:55:03.504449 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Nov 12 20:55:03.505842 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 12 20:55:03.508190 systemd[1]: Stopped target timers.target - Timer Units. Nov 12 20:55:03.510168 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Nov 12 20:55:03.510294 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Nov 12 20:55:03.512422 systemd[1]: Stopped target initrd.target - Initrd Default Target. Nov 12 20:55:03.514118 systemd[1]: Stopped target basic.target - Basic System. Nov 12 20:55:03.516117 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Nov 12 20:55:03.518397 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Nov 12 20:55:03.520258 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Nov 12 20:55:03.522341 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Nov 12 20:55:03.524418 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Nov 12 20:55:03.526617 systemd[1]: Stopped target sysinit.target - System Initialization. Nov 12 20:55:03.528552 systemd[1]: Stopped target local-fs.target - Local File Systems. Nov 12 20:55:03.530663 systemd[1]: Stopped target swap.target - Swaps. Nov 12 20:55:03.532387 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Nov 12 20:55:03.532518 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Nov 12 20:55:03.534586 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Nov 12 20:55:03.536140 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 12 20:55:03.538165 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Nov 12 20:55:03.538265 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 12 20:55:03.540320 systemd[1]: dracut-initqueue.service: Deactivated successfully. Nov 12 20:55:03.540427 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Nov 12 20:55:03.542614 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Nov 12 20:55:03.542722 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Nov 12 20:55:03.544716 systemd[1]: Stopped target paths.target - Path Units. Nov 12 20:55:03.546536 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Nov 12 20:55:03.550954 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 12 20:55:03.553491 systemd[1]: Stopped target slices.target - Slice Units. Nov 12 20:55:03.555217 systemd[1]: Stopped target sockets.target - Socket Units. Nov 12 20:55:03.557195 systemd[1]: iscsid.socket: Deactivated successfully. Nov 12 20:55:03.557297 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Nov 12 20:55:03.559633 systemd[1]: iscsiuio.socket: Deactivated successfully. Nov 12 20:55:03.559721 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Nov 12 20:55:03.561469 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Nov 12 20:55:03.561585 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Nov 12 20:55:03.563519 systemd[1]: ignition-files.service: Deactivated successfully. Nov 12 20:55:03.563654 systemd[1]: Stopped ignition-files.service - Ignition (files). Nov 12 20:55:03.575114 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Nov 12 20:55:03.576161 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Nov 12 20:55:03.576280 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Nov 12 20:55:03.580085 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Nov 12 20:55:03.582268 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Nov 12 20:55:03.582426 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Nov 12 20:55:03.583837 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Nov 12 20:55:03.583948 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Nov 12 20:55:03.590636 systemd[1]: initrd-cleanup.service: Deactivated successfully. Nov 12 20:55:03.590806 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Nov 12 20:55:03.608175 systemd[1]: sysroot-boot.mount: Deactivated successfully. Nov 12 20:55:03.610078 ignition[1012]: INFO : Ignition 2.19.0 Nov 12 20:55:03.610078 ignition[1012]: INFO : Stage: umount Nov 12 20:55:03.612188 ignition[1012]: INFO : no configs at "/usr/lib/ignition/base.d" Nov 12 20:55:03.612188 ignition[1012]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Nov 12 20:55:03.612188 ignition[1012]: INFO : umount: umount passed Nov 12 20:55:03.612188 ignition[1012]: INFO : Ignition finished successfully Nov 12 20:55:03.612287 systemd[1]: sysroot-boot.service: Deactivated successfully. Nov 12 20:55:03.612408 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Nov 12 20:55:03.614528 systemd[1]: ignition-mount.service: Deactivated successfully. Nov 12 20:55:03.614627 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Nov 12 20:55:03.616990 systemd[1]: Stopped target network.target - Network. Nov 12 20:55:03.618098 systemd[1]: ignition-disks.service: Deactivated successfully. Nov 12 20:55:03.618157 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Nov 12 20:55:03.620015 systemd[1]: ignition-kargs.service: Deactivated successfully. Nov 12 20:55:03.620063 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Nov 12 20:55:03.621958 systemd[1]: ignition-setup.service: Deactivated successfully. Nov 12 20:55:03.622005 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Nov 12 20:55:03.623813 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Nov 12 20:55:03.623863 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Nov 12 20:55:03.625706 systemd[1]: initrd-setup-root.service: Deactivated successfully. Nov 12 20:55:03.625750 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Nov 12 20:55:03.627784 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Nov 12 20:55:03.629609 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Nov 12 20:55:03.635957 systemd-networkd[769]: eth0: DHCPv6 lease lost Nov 12 20:55:03.638380 systemd[1]: systemd-networkd.service: Deactivated successfully. Nov 12 20:55:03.638508 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Nov 12 20:55:03.640536 systemd[1]: systemd-resolved.service: Deactivated successfully. Nov 12 20:55:03.640670 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Nov 12 20:55:03.643727 systemd[1]: systemd-networkd.socket: Deactivated successfully. Nov 12 20:55:03.643774 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Nov 12 20:55:03.649999 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Nov 12 20:55:03.651370 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Nov 12 20:55:03.651423 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Nov 12 20:55:03.653792 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 12 20:55:03.653851 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 12 20:55:03.656031 systemd[1]: systemd-modules-load.service: Deactivated successfully. Nov 12 20:55:03.656083 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Nov 12 20:55:03.658202 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Nov 12 20:55:03.658257 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 12 20:55:03.660486 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 12 20:55:03.671512 systemd[1]: network-cleanup.service: Deactivated successfully. Nov 12 20:55:03.671636 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Nov 12 20:55:03.676777 systemd[1]: systemd-udevd.service: Deactivated successfully. Nov 12 20:55:03.676971 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 12 20:55:03.679149 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Nov 12 20:55:03.679195 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Nov 12 20:55:03.681233 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Nov 12 20:55:03.681270 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Nov 12 20:55:03.683153 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Nov 12 20:55:03.683201 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Nov 12 20:55:03.685404 systemd[1]: dracut-cmdline.service: Deactivated successfully. Nov 12 20:55:03.685452 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Nov 12 20:55:03.687335 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Nov 12 20:55:03.687380 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Nov 12 20:55:03.700034 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Nov 12 20:55:03.701143 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Nov 12 20:55:03.701195 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 12 20:55:03.703441 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Nov 12 20:55:03.703499 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Nov 12 20:55:03.709778 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Nov 12 20:55:03.709890 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Nov 12 20:55:03.712145 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Nov 12 20:55:03.714697 systemd[1]: Starting initrd-switch-root.service - Switch Root... Nov 12 20:55:03.725474 systemd[1]: Switching root. Nov 12 20:55:03.758906 systemd-journald[192]: Journal stopped Nov 12 20:55:05.201562 systemd-journald[192]: Received SIGTERM from PID 1 (systemd). Nov 12 20:55:05.201646 kernel: SELinux: policy capability network_peer_controls=1 Nov 12 20:55:05.201675 kernel: SELinux: policy capability open_perms=1 Nov 12 20:55:05.201691 kernel: SELinux: policy capability extended_socket_class=1 Nov 12 20:55:05.201706 kernel: SELinux: policy capability always_check_network=0 Nov 12 20:55:05.201721 kernel: SELinux: policy capability cgroup_seclabel=1 Nov 12 20:55:05.201736 kernel: SELinux: policy capability nnp_nosuid_transition=1 Nov 12 20:55:05.201751 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Nov 12 20:55:05.201767 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Nov 12 20:55:05.201805 kernel: audit: type=1403 audit(1731444904.424:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Nov 12 20:55:05.201822 systemd[1]: Successfully loaded SELinux policy in 51.613ms. Nov 12 20:55:05.201866 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.713ms. Nov 12 20:55:05.201884 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Nov 12 20:55:05.201901 systemd[1]: Detected virtualization kvm. Nov 12 20:55:05.201937 systemd[1]: Detected architecture x86-64. Nov 12 20:55:05.201954 systemd[1]: Detected first boot. Nov 12 20:55:05.201971 systemd[1]: Initializing machine ID from VM UUID. Nov 12 20:55:05.201996 zram_generator::config[1056]: No configuration found. Nov 12 20:55:05.202032 systemd[1]: Populated /etc with preset unit settings. Nov 12 20:55:05.202055 systemd[1]: initrd-switch-root.service: Deactivated successfully. Nov 12 20:55:05.202071 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Nov 12 20:55:05.202088 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Nov 12 20:55:05.202105 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Nov 12 20:55:05.202122 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Nov 12 20:55:05.202138 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Nov 12 20:55:05.202154 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Nov 12 20:55:05.202180 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Nov 12 20:55:05.202197 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Nov 12 20:55:05.202217 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Nov 12 20:55:05.202234 systemd[1]: Created slice user.slice - User and Session Slice. Nov 12 20:55:05.202260 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Nov 12 20:55:05.202277 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Nov 12 20:55:05.202294 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Nov 12 20:55:05.202318 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Nov 12 20:55:05.202335 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Nov 12 20:55:05.202352 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Nov 12 20:55:05.202369 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Nov 12 20:55:05.202389 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Nov 12 20:55:05.202405 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Nov 12 20:55:05.202421 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Nov 12 20:55:05.202437 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Nov 12 20:55:05.202453 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Nov 12 20:55:05.202470 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Nov 12 20:55:05.202486 systemd[1]: Reached target remote-fs.target - Remote File Systems. Nov 12 20:55:05.202506 systemd[1]: Reached target slices.target - Slice Units. Nov 12 20:55:05.202522 systemd[1]: Reached target swap.target - Swaps. Nov 12 20:55:05.202538 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Nov 12 20:55:05.202555 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Nov 12 20:55:05.202572 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Nov 12 20:55:05.202589 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Nov 12 20:55:05.202605 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Nov 12 20:55:05.202624 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Nov 12 20:55:05.202640 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Nov 12 20:55:05.202658 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Nov 12 20:55:05.202677 systemd[1]: Mounting media.mount - External Media Directory... Nov 12 20:55:05.202694 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 12 20:55:05.202710 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Nov 12 20:55:05.202727 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Nov 12 20:55:05.202744 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Nov 12 20:55:05.202761 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Nov 12 20:55:05.202777 systemd[1]: Reached target machines.target - Containers. Nov 12 20:55:05.202794 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Nov 12 20:55:05.202813 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 12 20:55:05.203564 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Nov 12 20:55:05.203717 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Nov 12 20:55:05.203740 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 12 20:55:05.203799 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 12 20:55:05.203815 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 12 20:55:05.203832 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Nov 12 20:55:05.203848 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 12 20:55:05.203865 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Nov 12 20:55:05.203885 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Nov 12 20:55:05.203902 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Nov 12 20:55:05.203934 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Nov 12 20:55:05.203952 systemd[1]: Stopped systemd-fsck-usr.service. Nov 12 20:55:05.203969 systemd[1]: Starting systemd-journald.service - Journal Service... Nov 12 20:55:05.203984 kernel: loop: module loaded Nov 12 20:55:05.204000 kernel: fuse: init (API version 7.39) Nov 12 20:55:05.204016 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Nov 12 20:55:05.204033 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Nov 12 20:55:05.204077 systemd-journald[1126]: Collecting audit messages is disabled. Nov 12 20:55:05.204107 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Nov 12 20:55:05.204123 systemd-journald[1126]: Journal started Nov 12 20:55:05.204157 systemd-journald[1126]: Runtime Journal (/run/log/journal/b29ddd8cc6a24005a4ffb090905e37d8) is 6.0M, max 48.4M, 42.3M free. Nov 12 20:55:04.963579 systemd[1]: Queued start job for default target multi-user.target. Nov 12 20:55:04.993188 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Nov 12 20:55:04.993780 systemd[1]: systemd-journald.service: Deactivated successfully. Nov 12 20:55:04.994309 systemd[1]: systemd-journald.service: Consumed 1.034s CPU time. Nov 12 20:55:05.209085 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Nov 12 20:55:05.211549 systemd[1]: verity-setup.service: Deactivated successfully. Nov 12 20:55:05.211584 systemd[1]: Stopped verity-setup.service. Nov 12 20:55:05.214029 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 12 20:55:05.215933 kernel: ACPI: bus type drm_connector registered Nov 12 20:55:05.217187 systemd[1]: Started systemd-journald.service - Journal Service. Nov 12 20:55:05.218659 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Nov 12 20:55:05.219874 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Nov 12 20:55:05.221153 systemd[1]: Mounted media.mount - External Media Directory. Nov 12 20:55:05.222444 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Nov 12 20:55:05.223748 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Nov 12 20:55:05.225098 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Nov 12 20:55:05.226337 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Nov 12 20:55:05.227808 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Nov 12 20:55:05.229469 systemd[1]: modprobe@configfs.service: Deactivated successfully. Nov 12 20:55:05.229687 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Nov 12 20:55:05.231343 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 12 20:55:05.231553 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 12 20:55:05.233031 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 12 20:55:05.233259 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 12 20:55:05.234626 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 12 20:55:05.234848 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 12 20:55:05.236393 systemd[1]: modprobe@fuse.service: Deactivated successfully. Nov 12 20:55:05.236600 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Nov 12 20:55:05.238020 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 12 20:55:05.238230 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 12 20:55:05.239695 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Nov 12 20:55:05.241255 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Nov 12 20:55:05.243481 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Nov 12 20:55:05.257233 systemd[1]: Reached target network-pre.target - Preparation for Network. Nov 12 20:55:05.266049 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Nov 12 20:55:05.268837 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Nov 12 20:55:05.270147 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Nov 12 20:55:05.270199 systemd[1]: Reached target local-fs.target - Local File Systems. Nov 12 20:55:05.272410 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Nov 12 20:55:05.274965 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Nov 12 20:55:05.279825 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Nov 12 20:55:05.281154 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 12 20:55:05.288355 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Nov 12 20:55:05.298289 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Nov 12 20:55:05.299562 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 12 20:55:05.301553 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Nov 12 20:55:05.303340 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 12 20:55:05.321199 systemd-journald[1126]: Time spent on flushing to /var/log/journal/b29ddd8cc6a24005a4ffb090905e37d8 is 24.661ms for 952 entries. Nov 12 20:55:05.321199 systemd-journald[1126]: System Journal (/var/log/journal/b29ddd8cc6a24005a4ffb090905e37d8) is 8.0M, max 195.6M, 187.6M free. Nov 12 20:55:05.360118 systemd-journald[1126]: Received client request to flush runtime journal. Nov 12 20:55:05.318314 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 12 20:55:05.323574 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Nov 12 20:55:05.329179 systemd[1]: Starting systemd-sysusers.service - Create System Users... Nov 12 20:55:05.332742 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Nov 12 20:55:05.334763 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Nov 12 20:55:05.337056 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Nov 12 20:55:05.343363 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Nov 12 20:55:05.345958 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Nov 12 20:55:05.353819 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Nov 12 20:55:05.365953 kernel: loop0: detected capacity change from 0 to 142488 Nov 12 20:55:05.432024 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Nov 12 20:55:05.434718 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Nov 12 20:55:05.439110 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Nov 12 20:55:05.443565 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Nov 12 20:55:05.452657 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 12 20:55:05.458181 udevadm[1182]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Nov 12 20:55:05.464566 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Nov 12 20:55:05.465436 systemd[1]: Finished systemd-sysusers.service - Create System Users. Nov 12 20:55:05.467097 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Nov 12 20:55:05.469212 kernel: loop1: detected capacity change from 0 to 210664 Nov 12 20:55:05.488601 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Nov 12 20:55:05.522949 kernel: loop2: detected capacity change from 0 to 140768 Nov 12 20:55:05.560503 systemd-tmpfiles[1189]: ACLs are not supported, ignoring. Nov 12 20:55:05.560526 systemd-tmpfiles[1189]: ACLs are not supported, ignoring. Nov 12 20:55:05.572308 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Nov 12 20:55:05.598974 kernel: loop3: detected capacity change from 0 to 142488 Nov 12 20:55:05.619944 kernel: loop4: detected capacity change from 0 to 210664 Nov 12 20:55:05.627929 kernel: loop5: detected capacity change from 0 to 140768 Nov 12 20:55:05.634921 (sd-merge)[1194]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Nov 12 20:55:05.635697 (sd-merge)[1194]: Merged extensions into '/usr'. Nov 12 20:55:05.667472 systemd[1]: Reloading requested from client PID 1170 ('systemd-sysext') (unit systemd-sysext.service)... Nov 12 20:55:05.667633 systemd[1]: Reloading... Nov 12 20:55:05.761947 zram_generator::config[1217]: No configuration found. Nov 12 20:55:05.953164 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 12 20:55:06.022806 systemd[1]: Reloading finished in 354 ms. Nov 12 20:55:06.031880 ldconfig[1165]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Nov 12 20:55:06.066280 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Nov 12 20:55:06.068140 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Nov 12 20:55:06.083354 systemd[1]: Starting ensure-sysext.service... Nov 12 20:55:06.086380 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Nov 12 20:55:06.093135 systemd[1]: Reloading requested from client PID 1257 ('systemctl') (unit ensure-sysext.service)... Nov 12 20:55:06.093161 systemd[1]: Reloading... Nov 12 20:55:06.131764 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Nov 12 20:55:06.132162 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Nov 12 20:55:06.136741 systemd-tmpfiles[1258]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Nov 12 20:55:06.137175 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Nov 12 20:55:06.137376 systemd-tmpfiles[1258]: ACLs are not supported, ignoring. Nov 12 20:55:06.141593 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Nov 12 20:55:06.141755 systemd-tmpfiles[1258]: Skipping /boot Nov 12 20:55:06.192865 systemd-tmpfiles[1258]: Detected autofs mount point /boot during canonicalization of boot. Nov 12 20:55:06.194504 systemd-tmpfiles[1258]: Skipping /boot Nov 12 20:55:06.211931 zram_generator::config[1284]: No configuration found. Nov 12 20:55:06.348122 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 12 20:55:06.406042 systemd[1]: Reloading finished in 312 ms. Nov 12 20:55:06.440576 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Nov 12 20:55:06.449732 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Nov 12 20:55:06.493016 augenrules[1338]: No rules Nov 12 20:55:06.502544 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Nov 12 20:55:06.505983 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Nov 12 20:55:06.510165 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Nov 12 20:55:06.513713 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Nov 12 20:55:06.516008 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Nov 12 20:55:06.518419 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Nov 12 20:55:06.529335 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Nov 12 20:55:06.535466 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 12 20:55:06.535809 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Nov 12 20:55:06.544419 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Nov 12 20:55:06.547356 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Nov 12 20:55:06.549999 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Nov 12 20:55:06.553835 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Nov 12 20:55:06.555207 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Nov 12 20:55:06.558512 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Nov 12 20:55:06.562587 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Nov 12 20:55:06.563902 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Nov 12 20:55:06.565880 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Nov 12 20:55:06.568052 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Nov 12 20:55:06.568313 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Nov 12 20:55:06.570638 systemd[1]: modprobe@drm.service: Deactivated successfully. Nov 12 20:55:06.570942 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Nov 12 20:55:06.573298 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Nov 12 20:55:06.573540 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Nov 12 20:55:06.575579 systemd[1]: modprobe@loop.service: Deactivated successfully. Nov 12 20:55:06.575776 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Nov 12 20:55:06.577838 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Nov 12 20:55:06.582054 systemd[1]: Finished ensure-sysext.service. Nov 12 20:55:06.598373 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Nov 12 20:55:06.598484 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Nov 12 20:55:06.606111 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Nov 12 20:55:06.609372 systemd[1]: Starting systemd-update-done.service - Update is Completed... Nov 12 20:55:06.610572 systemd-udevd[1354]: Using default interface naming scheme 'v255'. Nov 12 20:55:06.610596 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Nov 12 20:55:06.624479 systemd[1]: Started systemd-userdbd.service - User Database Manager. Nov 12 20:55:06.627203 systemd[1]: Finished systemd-update-done.service - Update is Completed. Nov 12 20:55:06.640596 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Nov 12 20:55:06.675193 systemd[1]: Starting systemd-networkd.service - Network Configuration... Nov 12 20:55:06.725950 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1381) Nov 12 20:55:06.728932 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1381) Nov 12 20:55:06.733483 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Nov 12 20:55:06.740904 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Nov 12 20:55:06.766464 systemd[1]: Reached target time-set.target - System Time Set. Nov 12 20:55:06.779963 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1380) Nov 12 20:55:06.829612 systemd-resolved[1345]: Positive Trust Anchors: Nov 12 20:55:06.829645 systemd-resolved[1345]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Nov 12 20:55:06.829696 systemd-resolved[1345]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Nov 12 20:55:06.866588 systemd-resolved[1345]: Defaulting to hostname 'linux'. Nov 12 20:55:06.869719 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Nov 12 20:55:06.871927 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Nov 12 20:55:06.873162 systemd-networkd[1387]: lo: Link UP Nov 12 20:55:06.873174 systemd-networkd[1387]: lo: Gained carrier Nov 12 20:55:06.877014 kernel: ACPI: button: Power Button [PWRF] Nov 12 20:55:06.877662 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Nov 12 20:55:06.879430 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Nov 12 20:55:06.882114 systemd-networkd[1387]: Enumeration completed Nov 12 20:55:06.884937 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Nov 12 20:55:06.886616 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Nov 12 20:55:06.886836 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Nov 12 20:55:06.886855 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Nov 12 20:55:06.890274 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Nov 12 20:55:06.890605 systemd-networkd[1387]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 12 20:55:06.890613 systemd-networkd[1387]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Nov 12 20:55:06.891947 systemd-networkd[1387]: eth0: Link UP Nov 12 20:55:06.891954 systemd-networkd[1387]: eth0: Gained carrier Nov 12 20:55:06.891969 systemd-networkd[1387]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Nov 12 20:55:06.894046 systemd[1]: Started systemd-networkd.service - Network Configuration. Nov 12 20:55:06.896279 systemd[1]: Reached target network.target - Network. Nov 12 20:55:06.904048 systemd-networkd[1387]: eth0: DHCPv4 address 10.0.0.145/16, gateway 10.0.0.1 acquired from 10.0.0.1 Nov 12 20:55:06.906107 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Nov 12 20:55:06.907637 systemd-timesyncd[1363]: Network configuration changed, trying to establish connection. Nov 12 20:55:06.908763 systemd-timesyncd[1363]: Contacted time server 10.0.0.1:123 (10.0.0.1). Nov 12 20:55:06.908863 systemd-timesyncd[1363]: Initial clock synchronization to Tue 2024-11-12 20:55:07.254882 UTC. Nov 12 20:55:06.924225 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Nov 12 20:55:06.942811 kernel: mousedev: PS/2 mouse device common for all mice Nov 12 20:55:06.985082 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Nov 12 20:55:07.104256 kernel: kvm_amd: TSC scaling supported Nov 12 20:55:07.104336 kernel: kvm_amd: Nested Virtualization enabled Nov 12 20:55:07.104355 kernel: kvm_amd: Nested Paging enabled Nov 12 20:55:07.105655 kernel: kvm_amd: LBR virtualization supported Nov 12 20:55:07.105711 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Nov 12 20:55:07.106421 kernel: kvm_amd: Virtual GIF supported Nov 12 20:55:07.129987 kernel: EDAC MC: Ver: 3.0.0 Nov 12 20:55:07.165516 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Nov 12 20:55:07.173569 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Nov 12 20:55:07.186151 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Nov 12 20:55:07.221080 lvm[1417]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Nov 12 20:55:07.251523 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Nov 12 20:55:07.265650 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Nov 12 20:55:07.266968 systemd[1]: Reached target sysinit.target - System Initialization. Nov 12 20:55:07.268239 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Nov 12 20:55:07.269580 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Nov 12 20:55:07.271123 systemd[1]: Started logrotate.timer - Daily rotation of log files. Nov 12 20:55:07.272501 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Nov 12 20:55:07.273895 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Nov 12 20:55:07.275262 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Nov 12 20:55:07.275294 systemd[1]: Reached target paths.target - Path Units. Nov 12 20:55:07.276259 systemd[1]: Reached target timers.target - Timer Units. Nov 12 20:55:07.278224 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Nov 12 20:55:07.281993 systemd[1]: Starting docker.socket - Docker Socket for the API... Nov 12 20:55:07.299934 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Nov 12 20:55:07.302927 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Nov 12 20:55:07.347791 systemd[1]: Listening on docker.socket - Docker Socket for the API. Nov 12 20:55:07.349201 systemd[1]: Reached target sockets.target - Socket Units. Nov 12 20:55:07.350353 systemd[1]: Reached target basic.target - Basic System. Nov 12 20:55:07.351500 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Nov 12 20:55:07.351538 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Nov 12 20:55:07.351844 lvm[1421]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Nov 12 20:55:07.352921 systemd[1]: Starting containerd.service - containerd container runtime... Nov 12 20:55:07.355221 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Nov 12 20:55:07.360255 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Nov 12 20:55:07.367126 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Nov 12 20:55:07.369276 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Nov 12 20:55:07.369788 jq[1424]: false Nov 12 20:55:07.370601 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Nov 12 20:55:07.376070 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Nov 12 20:55:07.385234 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Nov 12 20:55:07.388311 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Nov 12 20:55:07.393636 extend-filesystems[1425]: Found loop3 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found loop4 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found loop5 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found sr0 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda1 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda2 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda3 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found usr Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda4 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda6 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda7 Nov 12 20:55:07.393636 extend-filesystems[1425]: Found vda9 Nov 12 20:55:07.393636 extend-filesystems[1425]: Checking size of /dev/vda9 Nov 12 20:55:07.394604 systemd[1]: Starting systemd-logind.service - User Login Management... Nov 12 20:55:07.398246 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Nov 12 20:55:07.440759 update_engine[1434]: I20241112 20:55:07.436308 1434 main.cc:92] Flatcar Update Engine starting Nov 12 20:55:07.398926 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Nov 12 20:55:07.400038 systemd[1]: Starting update-engine.service - Update Engine... Nov 12 20:55:07.402121 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Nov 12 20:55:07.409652 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Nov 12 20:55:07.409968 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Nov 12 20:55:07.423687 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Nov 12 20:55:07.428767 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Nov 12 20:55:07.428996 (ntainerd)[1444]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Nov 12 20:55:07.429062 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Nov 12 20:55:07.446994 jq[1435]: true Nov 12 20:55:07.446412 dbus-daemon[1423]: [system] SELinux support is enabled Nov 12 20:55:07.448073 systemd[1]: Started dbus.service - D-Bus System Message Bus. Nov 12 20:55:07.450356 update_engine[1434]: I20241112 20:55:07.448680 1434 update_check_scheduler.cc:74] Next update check in 5m14s Nov 12 20:55:07.455860 systemd[1]: motdgen.service: Deactivated successfully. Nov 12 20:55:07.456160 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Nov 12 20:55:07.467685 extend-filesystems[1425]: Resized partition /dev/vda9 Nov 12 20:55:07.469181 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Nov 12 20:55:07.493173 tar[1439]: linux-amd64/helm Nov 12 20:55:07.470147 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Nov 12 20:55:07.600179 extend-filesystems[1461]: resize2fs 1.47.1 (20-May-2024) Nov 12 20:55:07.615314 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1380) Nov 12 20:55:07.545443 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Nov 12 20:55:07.615407 jq[1457]: true Nov 12 20:55:07.545475 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Nov 12 20:55:07.547499 systemd[1]: Started update-engine.service - Update Engine. Nov 12 20:55:07.617041 systemd[1]: Started locksmithd.service - Cluster reboot manager. Nov 12 20:55:07.703161 sshd_keygen[1452]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Nov 12 20:55:07.764976 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Nov 12 20:55:07.776973 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Nov 12 20:55:07.815127 systemd[1]: Starting issuegen.service - Generate /run/issue... Nov 12 20:55:07.821938 systemd[1]: issuegen.service: Deactivated successfully. Nov 12 20:55:07.822181 systemd[1]: Finished issuegen.service - Generate /run/issue. Nov 12 20:55:07.824689 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Nov 12 20:55:07.934331 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Nov 12 20:55:07.950196 systemd[1]: Started getty@tty1.service - Getty on tty1. Nov 12 20:55:07.952972 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Nov 12 20:55:07.954395 systemd[1]: Reached target getty.target - Login Prompts. Nov 12 20:55:08.068396 locksmithd[1462]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Nov 12 20:55:08.087293 systemd-logind[1433]: Watching system buttons on /dev/input/event1 (Power Button) Nov 12 20:55:08.087318 systemd-logind[1433]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Nov 12 20:55:08.092077 systemd-logind[1433]: New seat seat0. Nov 12 20:55:08.097838 systemd[1]: Started systemd-logind.service - User Login Management. Nov 12 20:55:08.189983 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Nov 12 20:55:08.312353 systemd-networkd[1387]: eth0: Gained IPv6LL Nov 12 20:55:08.317380 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Nov 12 20:55:08.319445 systemd[1]: Reached target network-online.target - Network is Online. Nov 12 20:55:08.329385 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Nov 12 20:55:08.334691 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:08.340037 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Nov 12 20:55:08.380554 systemd[1]: coreos-metadata.service: Deactivated successfully. Nov 12 20:55:08.380862 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Nov 12 20:55:08.384037 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Nov 12 20:55:08.393074 extend-filesystems[1461]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Nov 12 20:55:08.393074 extend-filesystems[1461]: old_desc_blocks = 1, new_desc_blocks = 1 Nov 12 20:55:08.393074 extend-filesystems[1461]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Nov 12 20:55:08.400567 extend-filesystems[1425]: Resized filesystem in /dev/vda9 Nov 12 20:55:08.437261 systemd[1]: extend-filesystems.service: Deactivated successfully. Nov 12 20:55:08.437524 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Nov 12 20:55:08.452291 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Nov 12 20:55:08.596749 containerd[1444]: time="2024-11-12T20:55:08.596541062Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Nov 12 20:55:08.627635 containerd[1444]: time="2024-11-12T20:55:08.627518216Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630144 containerd[1444]: time="2024-11-12T20:55:08.630088012Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.60-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630144 containerd[1444]: time="2024-11-12T20:55:08.630120194Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Nov 12 20:55:08.630261 containerd[1444]: time="2024-11-12T20:55:08.630145993Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Nov 12 20:55:08.630493 containerd[1444]: time="2024-11-12T20:55:08.630462723Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Nov 12 20:55:08.630533 containerd[1444]: time="2024-11-12T20:55:08.630492773Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630594 containerd[1444]: time="2024-11-12T20:55:08.630574607Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630634 containerd[1444]: time="2024-11-12T20:55:08.630591640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630893 containerd[1444]: time="2024-11-12T20:55:08.630861203Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630893 containerd[1444]: time="2024-11-12T20:55:08.630887397Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630937 containerd[1444]: time="2024-11-12T20:55:08.630906591Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Nov 12 20:55:08.630937 containerd[1444]: time="2024-11-12T20:55:08.630917124Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.631088 containerd[1444]: time="2024-11-12T20:55:08.631060444Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.632489 containerd[1444]: time="2024-11-12T20:55:08.632422766Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Nov 12 20:55:08.632686 containerd[1444]: time="2024-11-12T20:55:08.632651507Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Nov 12 20:55:08.632686 containerd[1444]: time="2024-11-12T20:55:08.632678396Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Nov 12 20:55:08.632894 containerd[1444]: time="2024-11-12T20:55:08.632862757Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Nov 12 20:55:08.633087 containerd[1444]: time="2024-11-12T20:55:08.633038258Z" level=info msg="metadata content store policy set" policy=shared Nov 12 20:55:08.660389 bash[1484]: Updated "/home/core/.ssh/authorized_keys" Nov 12 20:55:08.661538 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Nov 12 20:55:08.664730 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Nov 12 20:55:08.714901 tar[1439]: linux-amd64/LICENSE Nov 12 20:55:08.715363 tar[1439]: linux-amd64/README.md Nov 12 20:55:08.796794 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Nov 12 20:55:08.836249 containerd[1444]: time="2024-11-12T20:55:08.836080701Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Nov 12 20:55:08.836249 containerd[1444]: time="2024-11-12T20:55:08.836190682Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Nov 12 20:55:08.836249 containerd[1444]: time="2024-11-12T20:55:08.836213725Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Nov 12 20:55:08.836249 containerd[1444]: time="2024-11-12T20:55:08.836233981Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Nov 12 20:55:08.836249 containerd[1444]: time="2024-11-12T20:55:08.836261661Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Nov 12 20:55:08.836615 containerd[1444]: time="2024-11-12T20:55:08.836580180Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Nov 12 20:55:08.837022 containerd[1444]: time="2024-11-12T20:55:08.836994913Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Nov 12 20:55:08.837190 containerd[1444]: time="2024-11-12T20:55:08.837149482Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Nov 12 20:55:08.837190 containerd[1444]: time="2024-11-12T20:55:08.837181092Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Nov 12 20:55:08.837252 containerd[1444]: time="2024-11-12T20:55:08.837201514Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Nov 12 20:55:08.837252 containerd[1444]: time="2024-11-12T20:55:08.837219223Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837252 containerd[1444]: time="2024-11-12T20:55:08.837235631Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837252 containerd[1444]: time="2024-11-12T20:55:08.837252570Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837271609Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837289193Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837305248Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837328539Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837343138Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Nov 12 20:55:08.837379 containerd[1444]: time="2024-11-12T20:55:08.837372971Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837390741Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837405279Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837424526Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837463009Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837477556Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837489982Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837507160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837520615Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837530 containerd[1444]: time="2024-11-12T20:55:08.837538043Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837551456Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837566419Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837580873Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837599881Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837625284Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837637138Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837648701Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837734798Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837764901Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837778991Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Nov 12 20:55:08.837779 containerd[1444]: time="2024-11-12T20:55:08.837792071Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Nov 12 20:55:08.838094 containerd[1444]: time="2024-11-12T20:55:08.837803270Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.838094 containerd[1444]: time="2024-11-12T20:55:08.837820416Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Nov 12 20:55:08.838094 containerd[1444]: time="2024-11-12T20:55:08.837834964Z" level=info msg="NRI interface is disabled by configuration." Nov 12 20:55:08.838094 containerd[1444]: time="2024-11-12T20:55:08.837845268Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Nov 12 20:55:08.838238 containerd[1444]: time="2024-11-12T20:55:08.838162997Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Nov 12 20:55:08.838238 containerd[1444]: time="2024-11-12T20:55:08.838237656Z" level=info msg="Connect containerd service" Nov 12 20:55:08.838478 containerd[1444]: time="2024-11-12T20:55:08.838279457Z" level=info msg="using legacy CRI server" Nov 12 20:55:08.838478 containerd[1444]: time="2024-11-12T20:55:08.838288462Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Nov 12 20:55:08.838478 containerd[1444]: time="2024-11-12T20:55:08.838413272Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Nov 12 20:55:08.839203 containerd[1444]: time="2024-11-12T20:55:08.839159937Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 12 20:55:08.839435 containerd[1444]: time="2024-11-12T20:55:08.839349039Z" level=info msg="Start subscribing containerd event" Nov 12 20:55:08.839478 containerd[1444]: time="2024-11-12T20:55:08.839461547Z" level=info msg="Start recovering state" Nov 12 20:55:08.839659 containerd[1444]: time="2024-11-12T20:55:08.839637236Z" level=info msg="Start event monitor" Nov 12 20:55:08.839713 containerd[1444]: time="2024-11-12T20:55:08.839655485Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Nov 12 20:55:08.839713 containerd[1444]: time="2024-11-12T20:55:08.839678932Z" level=info msg="Start snapshots syncer" Nov 12 20:55:08.839713 containerd[1444]: time="2024-11-12T20:55:08.839693272Z" level=info msg="Start cni network conf syncer for default" Nov 12 20:55:08.839713 containerd[1444]: time="2024-11-12T20:55:08.839705771Z" level=info msg="Start streaming server" Nov 12 20:55:08.840499 containerd[1444]: time="2024-11-12T20:55:08.839726535Z" level=info msg=serving... address=/run/containerd/containerd.sock Nov 12 20:55:08.840499 containerd[1444]: time="2024-11-12T20:55:08.839819694Z" level=info msg="containerd successfully booted in 0.248200s" Nov 12 20:55:08.839969 systemd[1]: Started containerd.service - containerd container runtime. Nov 12 20:55:10.131373 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:10.133715 systemd[1]: Reached target multi-user.target - Multi-User System. Nov 12 20:55:10.135445 systemd[1]: Startup finished in 768ms (kernel) + 8.668s (initrd) + 5.760s (userspace) = 15.197s. Nov 12 20:55:10.181236 (kubelet)[1537]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 12 20:55:10.920118 kubelet[1537]: E1112 20:55:10.920002 1537 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 12 20:55:10.925148 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 12 20:55:10.925414 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 12 20:55:10.925832 systemd[1]: kubelet.service: Consumed 2.102s CPU time. Nov 12 20:55:12.500500 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Nov 12 20:55:12.502152 systemd[1]: Started sshd@0-10.0.0.145:22-10.0.0.1:59360.service - OpenSSH per-connection server daemon (10.0.0.1:59360). Nov 12 20:55:12.553476 sshd[1551]: Accepted publickey for core from 10.0.0.1 port 59360 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:12.555740 sshd[1551]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:12.567577 systemd-logind[1433]: New session 1 of user core. Nov 12 20:55:12.569503 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Nov 12 20:55:12.586401 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Nov 12 20:55:12.601397 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Nov 12 20:55:12.609498 systemd[1]: Starting user@500.service - User Manager for UID 500... Nov 12 20:55:12.615284 (systemd)[1555]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Nov 12 20:55:12.752379 systemd[1555]: Queued start job for default target default.target. Nov 12 20:55:12.764173 systemd[1555]: Created slice app.slice - User Application Slice. Nov 12 20:55:12.764209 systemd[1555]: Reached target paths.target - Paths. Nov 12 20:55:12.764226 systemd[1555]: Reached target timers.target - Timers. Nov 12 20:55:12.766254 systemd[1555]: Starting dbus.socket - D-Bus User Message Bus Socket... Nov 12 20:55:12.780260 systemd[1555]: Listening on dbus.socket - D-Bus User Message Bus Socket. Nov 12 20:55:12.780436 systemd[1555]: Reached target sockets.target - Sockets. Nov 12 20:55:12.780457 systemd[1555]: Reached target basic.target - Basic System. Nov 12 20:55:12.780510 systemd[1555]: Reached target default.target - Main User Target. Nov 12 20:55:12.780549 systemd[1555]: Startup finished in 155ms. Nov 12 20:55:12.780984 systemd[1]: Started user@500.service - User Manager for UID 500. Nov 12 20:55:12.782636 systemd[1]: Started session-1.scope - Session 1 of User core. Nov 12 20:55:12.850044 systemd[1]: Started sshd@1-10.0.0.145:22-10.0.0.1:59364.service - OpenSSH per-connection server daemon (10.0.0.1:59364). Nov 12 20:55:12.893007 sshd[1566]: Accepted publickey for core from 10.0.0.1 port 59364 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:12.894785 sshd[1566]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:12.899257 systemd-logind[1433]: New session 2 of user core. Nov 12 20:55:12.909155 systemd[1]: Started session-2.scope - Session 2 of User core. Nov 12 20:55:12.966113 sshd[1566]: pam_unix(sshd:session): session closed for user core Nov 12 20:55:12.984202 systemd[1]: sshd@1-10.0.0.145:22-10.0.0.1:59364.service: Deactivated successfully. Nov 12 20:55:12.986086 systemd[1]: session-2.scope: Deactivated successfully. Nov 12 20:55:12.987801 systemd-logind[1433]: Session 2 logged out. Waiting for processes to exit. Nov 12 20:55:12.998282 systemd[1]: Started sshd@2-10.0.0.145:22-10.0.0.1:59368.service - OpenSSH per-connection server daemon (10.0.0.1:59368). Nov 12 20:55:12.999508 systemd-logind[1433]: Removed session 2. Nov 12 20:55:13.034252 sshd[1573]: Accepted publickey for core from 10.0.0.1 port 59368 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:13.036092 sshd[1573]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:13.040488 systemd-logind[1433]: New session 3 of user core. Nov 12 20:55:13.053131 systemd[1]: Started session-3.scope - Session 3 of User core. Nov 12 20:55:13.104038 sshd[1573]: pam_unix(sshd:session): session closed for user core Nov 12 20:55:13.119000 systemd[1]: sshd@2-10.0.0.145:22-10.0.0.1:59368.service: Deactivated successfully. Nov 12 20:55:13.120750 systemd[1]: session-3.scope: Deactivated successfully. Nov 12 20:55:13.122110 systemd-logind[1433]: Session 3 logged out. Waiting for processes to exit. Nov 12 20:55:13.136308 systemd[1]: Started sshd@3-10.0.0.145:22-10.0.0.1:59380.service - OpenSSH per-connection server daemon (10.0.0.1:59380). Nov 12 20:55:13.137509 systemd-logind[1433]: Removed session 3. Nov 12 20:55:13.172525 sshd[1580]: Accepted publickey for core from 10.0.0.1 port 59380 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:13.174401 sshd[1580]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:13.178797 systemd-logind[1433]: New session 4 of user core. Nov 12 20:55:13.192115 systemd[1]: Started session-4.scope - Session 4 of User core. Nov 12 20:55:13.248632 sshd[1580]: pam_unix(sshd:session): session closed for user core Nov 12 20:55:13.267176 systemd[1]: sshd@3-10.0.0.145:22-10.0.0.1:59380.service: Deactivated successfully. Nov 12 20:55:13.268950 systemd[1]: session-4.scope: Deactivated successfully. Nov 12 20:55:13.270592 systemd-logind[1433]: Session 4 logged out. Waiting for processes to exit. Nov 12 20:55:13.271970 systemd[1]: Started sshd@4-10.0.0.145:22-10.0.0.1:59390.service - OpenSSH per-connection server daemon (10.0.0.1:59390). Nov 12 20:55:13.272782 systemd-logind[1433]: Removed session 4. Nov 12 20:55:13.323415 sshd[1587]: Accepted publickey for core from 10.0.0.1 port 59390 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:13.325106 sshd[1587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:13.329397 systemd-logind[1433]: New session 5 of user core. Nov 12 20:55:13.339069 systemd[1]: Started session-5.scope - Session 5 of User core. Nov 12 20:55:14.053414 sudo[1590]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Nov 12 20:55:14.053786 sudo[1590]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 12 20:55:14.068888 sudo[1590]: pam_unix(sudo:session): session closed for user root Nov 12 20:55:14.071362 sshd[1587]: pam_unix(sshd:session): session closed for user core Nov 12 20:55:14.081145 systemd[1]: sshd@4-10.0.0.145:22-10.0.0.1:59390.service: Deactivated successfully. Nov 12 20:55:14.083234 systemd[1]: session-5.scope: Deactivated successfully. Nov 12 20:55:14.085064 systemd-logind[1433]: Session 5 logged out. Waiting for processes to exit. Nov 12 20:55:14.096380 systemd[1]: Started sshd@5-10.0.0.145:22-10.0.0.1:59406.service - OpenSSH per-connection server daemon (10.0.0.1:59406). Nov 12 20:55:14.097555 systemd-logind[1433]: Removed session 5. Nov 12 20:55:14.136005 sshd[1595]: Accepted publickey for core from 10.0.0.1 port 59406 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:14.138173 sshd[1595]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:14.143427 systemd-logind[1433]: New session 6 of user core. Nov 12 20:55:14.155061 systemd[1]: Started session-6.scope - Session 6 of User core. Nov 12 20:55:14.212008 sudo[1599]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Nov 12 20:55:14.212362 sudo[1599]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 12 20:55:14.217048 sudo[1599]: pam_unix(sudo:session): session closed for user root Nov 12 20:55:14.223843 sudo[1598]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Nov 12 20:55:14.224282 sudo[1598]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 12 20:55:14.244159 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Nov 12 20:55:14.245963 auditctl[1602]: No rules Nov 12 20:55:14.247339 systemd[1]: audit-rules.service: Deactivated successfully. Nov 12 20:55:14.247591 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Nov 12 20:55:14.249357 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Nov 12 20:55:14.283770 augenrules[1620]: No rules Nov 12 20:55:14.285814 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Nov 12 20:55:14.287574 sudo[1598]: pam_unix(sudo:session): session closed for user root Nov 12 20:55:14.289968 sshd[1595]: pam_unix(sshd:session): session closed for user core Nov 12 20:55:14.307735 systemd[1]: sshd@5-10.0.0.145:22-10.0.0.1:59406.service: Deactivated successfully. Nov 12 20:55:14.309863 systemd[1]: session-6.scope: Deactivated successfully. Nov 12 20:55:14.312363 systemd-logind[1433]: Session 6 logged out. Waiting for processes to exit. Nov 12 20:55:14.319341 systemd[1]: Started sshd@6-10.0.0.145:22-10.0.0.1:59410.service - OpenSSH per-connection server daemon (10.0.0.1:59410). Nov 12 20:55:14.320286 systemd-logind[1433]: Removed session 6. Nov 12 20:55:14.354708 sshd[1628]: Accepted publickey for core from 10.0.0.1 port 59410 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:55:14.356332 sshd[1628]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:55:14.360871 systemd-logind[1433]: New session 7 of user core. Nov 12 20:55:14.371241 systemd[1]: Started session-7.scope - Session 7 of User core. Nov 12 20:55:14.424683 sudo[1632]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Nov 12 20:55:14.425045 sudo[1632]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Nov 12 20:55:14.740217 systemd[1]: Starting docker.service - Docker Application Container Engine... Nov 12 20:55:14.740343 (dockerd)[1650]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Nov 12 20:55:15.052396 dockerd[1650]: time="2024-11-12T20:55:15.052228747Z" level=info msg="Starting up" Nov 12 20:55:15.813372 dockerd[1650]: time="2024-11-12T20:55:15.813291374Z" level=info msg="Loading containers: start." Nov 12 20:55:15.939959 kernel: Initializing XFRM netlink socket Nov 12 20:55:16.031245 systemd-networkd[1387]: docker0: Link UP Nov 12 20:55:16.452673 dockerd[1650]: time="2024-11-12T20:55:16.452604918Z" level=info msg="Loading containers: done." Nov 12 20:55:16.470577 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3579815838-merged.mount: Deactivated successfully. Nov 12 20:55:16.661618 dockerd[1650]: time="2024-11-12T20:55:16.661316445Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Nov 12 20:55:16.661618 dockerd[1650]: time="2024-11-12T20:55:16.661476873Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Nov 12 20:55:16.661827 dockerd[1650]: time="2024-11-12T20:55:16.661646160Z" level=info msg="Daemon has completed initialization" Nov 12 20:55:17.643807 dockerd[1650]: time="2024-11-12T20:55:17.643715263Z" level=info msg="API listen on /run/docker.sock" Nov 12 20:55:17.644013 systemd[1]: Started docker.service - Docker Application Container Engine. Nov 12 20:55:18.625683 containerd[1444]: time="2024-11-12T20:55:18.625620760Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.6\"" Nov 12 20:55:21.176525 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Nov 12 20:55:21.186099 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:21.377905 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:21.382223 (kubelet)[1814]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 12 20:55:21.495785 kubelet[1814]: E1112 20:55:21.495636 1814 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 12 20:55:21.502952 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 12 20:55:21.503159 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 12 20:55:23.029066 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1676685476.mount: Deactivated successfully. Nov 12 20:55:24.530214 containerd[1444]: time="2024-11-12T20:55:24.530127634Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:24.535975 containerd[1444]: time="2024-11-12T20:55:24.535888381Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.6: active requests=0, bytes read=32676443" Nov 12 20:55:24.537690 containerd[1444]: time="2024-11-12T20:55:24.537657260Z" level=info msg="ImageCreate event name:\"sha256:a247bfa6152e770cd36ef6fe2a8831429eb43da1fd506c30b12af93f032ee849\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:24.543642 containerd[1444]: time="2024-11-12T20:55:24.543561977Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:3a820898379831ecff7cf4ce4954bb7a6505988eefcef146fd1ee2f56a01cdbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:24.545344 containerd[1444]: time="2024-11-12T20:55:24.545280297Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.6\" with image id \"sha256:a247bfa6152e770cd36ef6fe2a8831429eb43da1fd506c30b12af93f032ee849\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.6\", repo digest \"registry.k8s.io/kube-apiserver@sha256:3a820898379831ecff7cf4ce4954bb7a6505988eefcef146fd1ee2f56a01cdbb\", size \"32673243\" in 5.919600474s" Nov 12 20:55:24.545433 containerd[1444]: time="2024-11-12T20:55:24.545347300Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.6\" returns image reference \"sha256:a247bfa6152e770cd36ef6fe2a8831429eb43da1fd506c30b12af93f032ee849\"" Nov 12 20:55:24.618332 containerd[1444]: time="2024-11-12T20:55:24.618274274Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.6\"" Nov 12 20:55:27.069480 containerd[1444]: time="2024-11-12T20:55:27.069400684Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:27.070438 containerd[1444]: time="2024-11-12T20:55:27.070357321Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.6: active requests=0, bytes read=29605796" Nov 12 20:55:27.071739 containerd[1444]: time="2024-11-12T20:55:27.071698748Z" level=info msg="ImageCreate event name:\"sha256:382949f9bfdd9da8bf555d18adac4eb0dba8264b7e3b5963e6a26ef8d412477c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:27.075934 containerd[1444]: time="2024-11-12T20:55:27.075862142Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3a412c3cdf35d39c8d37748b457a486faae7c5f2ee1d1ba2059c709bc5534686\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:27.077144 containerd[1444]: time="2024-11-12T20:55:27.077089113Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.6\" with image id \"sha256:382949f9bfdd9da8bf555d18adac4eb0dba8264b7e3b5963e6a26ef8d412477c\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.6\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3a412c3cdf35d39c8d37748b457a486faae7c5f2ee1d1ba2059c709bc5534686\", size \"31051162\" in 2.458768293s" Nov 12 20:55:27.077210 containerd[1444]: time="2024-11-12T20:55:27.077151003Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.6\" returns image reference \"sha256:382949f9bfdd9da8bf555d18adac4eb0dba8264b7e3b5963e6a26ef8d412477c\"" Nov 12 20:55:27.105451 containerd[1444]: time="2024-11-12T20:55:27.105396570Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.6\"" Nov 12 20:55:29.034430 containerd[1444]: time="2024-11-12T20:55:29.034353071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:29.092266 containerd[1444]: time="2024-11-12T20:55:29.092164590Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.6: active requests=0, bytes read=17784244" Nov 12 20:55:29.135094 containerd[1444]: time="2024-11-12T20:55:29.135028549Z" level=info msg="ImageCreate event name:\"sha256:ad5858afd532223324ff223396490f5fd8228323963b424ad7868407bd4ef1fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:29.166487 containerd[1444]: time="2024-11-12T20:55:29.166428151Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:948395c284d82c985f2dc0d99b5b51b3ca85eba97003babbc73834e0ab91fa59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:29.167643 containerd[1444]: time="2024-11-12T20:55:29.167595246Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.6\" with image id \"sha256:ad5858afd532223324ff223396490f5fd8228323963b424ad7868407bd4ef1fb\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.6\", repo digest \"registry.k8s.io/kube-scheduler@sha256:948395c284d82c985f2dc0d99b5b51b3ca85eba97003babbc73834e0ab91fa59\", size \"19229628\" in 2.06214573s" Nov 12 20:55:29.167694 containerd[1444]: time="2024-11-12T20:55:29.167647364Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.6\" returns image reference \"sha256:ad5858afd532223324ff223396490f5fd8228323963b424ad7868407bd4ef1fb\"" Nov 12 20:55:29.195685 containerd[1444]: time="2024-11-12T20:55:29.195638563Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.6\"" Nov 12 20:55:30.545400 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount383896007.mount: Deactivated successfully. Nov 12 20:55:31.575710 containerd[1444]: time="2024-11-12T20:55:31.575645543Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:31.576754 containerd[1444]: time="2024-11-12T20:55:31.576697839Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.6: active requests=0, bytes read=29054624" Nov 12 20:55:31.578166 containerd[1444]: time="2024-11-12T20:55:31.578133454Z" level=info msg="ImageCreate event name:\"sha256:2cce8902ed3ccdc78ecdb02734bd9ba32e2c7b44fc221663cf9ece2a179ff6a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:31.580623 containerd[1444]: time="2024-11-12T20:55:31.580570874Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:aaf790f611159ab21713affc2c5676f742c9b31db26dd2e61e46c4257dd11b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:31.581297 containerd[1444]: time="2024-11-12T20:55:31.581263636Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.6\" with image id \"sha256:2cce8902ed3ccdc78ecdb02734bd9ba32e2c7b44fc221663cf9ece2a179ff6a6\", repo tag \"registry.k8s.io/kube-proxy:v1.30.6\", repo digest \"registry.k8s.io/kube-proxy@sha256:aaf790f611159ab21713affc2c5676f742c9b31db26dd2e61e46c4257dd11b76\", size \"29053643\" in 2.385583435s" Nov 12 20:55:31.581354 containerd[1444]: time="2024-11-12T20:55:31.581297960Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.6\" returns image reference \"sha256:2cce8902ed3ccdc78ecdb02734bd9ba32e2c7b44fc221663cf9ece2a179ff6a6\"" Nov 12 20:55:31.596777 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Nov 12 20:55:31.604136 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:31.607502 containerd[1444]: time="2024-11-12T20:55:31.607270142Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Nov 12 20:55:31.778540 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:31.783050 (kubelet)[1925]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 12 20:55:31.913457 kubelet[1925]: E1112 20:55:31.913402 1925 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 12 20:55:31.917522 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 12 20:55:31.917717 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 12 20:55:32.749396 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2518061034.mount: Deactivated successfully. Nov 12 20:55:33.451997 containerd[1444]: time="2024-11-12T20:55:33.451937058Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:33.452687 containerd[1444]: time="2024-11-12T20:55:33.452628371Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Nov 12 20:55:33.454184 containerd[1444]: time="2024-11-12T20:55:33.454121815Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:33.457164 containerd[1444]: time="2024-11-12T20:55:33.457128828Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:33.458600 containerd[1444]: time="2024-11-12T20:55:33.458556882Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.851245403s" Nov 12 20:55:33.458651 containerd[1444]: time="2024-11-12T20:55:33.458609400Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Nov 12 20:55:33.482127 containerd[1444]: time="2024-11-12T20:55:33.482088639Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Nov 12 20:55:34.925787 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount704771529.mount: Deactivated successfully. Nov 12 20:55:34.937291 containerd[1444]: time="2024-11-12T20:55:34.937217891Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:34.938342 containerd[1444]: time="2024-11-12T20:55:34.938285908Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Nov 12 20:55:34.940019 containerd[1444]: time="2024-11-12T20:55:34.939978982Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:34.942967 containerd[1444]: time="2024-11-12T20:55:34.942890872Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:34.943566 containerd[1444]: time="2024-11-12T20:55:34.943512337Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 1.461382327s" Nov 12 20:55:34.943614 containerd[1444]: time="2024-11-12T20:55:34.943567605Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Nov 12 20:55:34.968165 containerd[1444]: time="2024-11-12T20:55:34.968123649Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Nov 12 20:55:35.579359 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3486685690.mount: Deactivated successfully. Nov 12 20:55:39.818356 containerd[1444]: time="2024-11-12T20:55:39.818275225Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:39.819382 containerd[1444]: time="2024-11-12T20:55:39.819327286Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238571" Nov 12 20:55:39.821449 containerd[1444]: time="2024-11-12T20:55:39.821379946Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:39.826867 containerd[1444]: time="2024-11-12T20:55:39.826797599Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:55:39.828243 containerd[1444]: time="2024-11-12T20:55:39.828207662Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 4.860041118s" Nov 12 20:55:39.828298 containerd[1444]: time="2024-11-12T20:55:39.828246844Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" Nov 12 20:55:42.096814 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Nov 12 20:55:42.110221 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:42.280684 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:42.286188 (kubelet)[2125]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Nov 12 20:55:42.381131 kubelet[2125]: E1112 20:55:42.380967 2125 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Nov 12 20:55:42.386375 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Nov 12 20:55:42.386569 systemd[1]: kubelet.service: Failed with result 'exit-code'. Nov 12 20:55:42.438904 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:42.456198 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:42.474603 systemd[1]: Reloading requested from client PID 2141 ('systemctl') (unit session-7.scope)... Nov 12 20:55:42.474620 systemd[1]: Reloading... Nov 12 20:55:42.559944 zram_generator::config[2183]: No configuration found. Nov 12 20:55:43.870713 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 12 20:55:43.952544 systemd[1]: Reloading finished in 1477 ms. Nov 12 20:55:44.005667 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:44.010248 systemd[1]: kubelet.service: Deactivated successfully. Nov 12 20:55:44.010548 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:44.012641 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:55:44.546805 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:55:44.551306 (kubelet)[2230]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 12 20:55:44.585573 kubelet[2230]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 12 20:55:44.585573 kubelet[2230]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 12 20:55:44.585573 kubelet[2230]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 12 20:55:44.585996 kubelet[2230]: I1112 20:55:44.585621 2230 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 12 20:55:47.267977 kubelet[2230]: I1112 20:55:47.267928 2230 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Nov 12 20:55:47.267977 kubelet[2230]: I1112 20:55:47.267962 2230 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 12 20:55:47.268528 kubelet[2230]: I1112 20:55:47.268173 2230 server.go:927] "Client rotation is on, will bootstrap in background" Nov 12 20:55:47.380200 kubelet[2230]: I1112 20:55:47.380131 2230 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 12 20:55:47.390982 kubelet[2230]: E1112 20:55:47.390891 2230 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.145:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.408337 kubelet[2230]: I1112 20:55:47.408281 2230 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 12 20:55:47.412366 kubelet[2230]: I1112 20:55:47.412294 2230 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 12 20:55:47.412574 kubelet[2230]: I1112 20:55:47.412351 2230 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Nov 12 20:55:47.413301 kubelet[2230]: I1112 20:55:47.413268 2230 topology_manager.go:138] "Creating topology manager with none policy" Nov 12 20:55:47.413301 kubelet[2230]: I1112 20:55:47.413296 2230 container_manager_linux.go:301] "Creating device plugin manager" Nov 12 20:55:47.413491 kubelet[2230]: I1112 20:55:47.413467 2230 state_mem.go:36] "Initialized new in-memory state store" Nov 12 20:55:47.420562 kubelet[2230]: I1112 20:55:47.420500 2230 kubelet.go:400] "Attempting to sync node with API server" Nov 12 20:55:47.420562 kubelet[2230]: I1112 20:55:47.420555 2230 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 12 20:55:47.420639 kubelet[2230]: I1112 20:55:47.420605 2230 kubelet.go:312] "Adding apiserver pod source" Nov 12 20:55:47.420662 kubelet[2230]: I1112 20:55:47.420641 2230 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 12 20:55:47.421225 kubelet[2230]: W1112 20:55:47.421153 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.145:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.421263 kubelet[2230]: E1112 20:55:47.421231 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.145:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.424802 kubelet[2230]: W1112 20:55:47.424769 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.145:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.424853 kubelet[2230]: E1112 20:55:47.424803 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.145:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.432896 kubelet[2230]: I1112 20:55:47.432874 2230 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Nov 12 20:55:47.436697 kubelet[2230]: I1112 20:55:47.436665 2230 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 12 20:55:47.436765 kubelet[2230]: W1112 20:55:47.436757 2230 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Nov 12 20:55:47.437591 kubelet[2230]: I1112 20:55:47.437549 2230 server.go:1264] "Started kubelet" Nov 12 20:55:47.439253 kubelet[2230]: I1112 20:55:47.439222 2230 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 12 20:55:47.445420 kubelet[2230]: I1112 20:55:47.445369 2230 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 12 20:55:47.446479 kubelet[2230]: I1112 20:55:47.446456 2230 server.go:455] "Adding debug handlers to kubelet server" Nov 12 20:55:47.447436 kubelet[2230]: I1112 20:55:47.447390 2230 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 12 20:55:47.447650 kubelet[2230]: I1112 20:55:47.447631 2230 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 12 20:55:47.728809 kubelet[2230]: I1112 20:55:47.727867 2230 volume_manager.go:291] "Starting Kubelet Volume Manager" Nov 12 20:55:47.728809 kubelet[2230]: I1112 20:55:47.728018 2230 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Nov 12 20:55:47.728809 kubelet[2230]: I1112 20:55:47.728094 2230 reconciler.go:26] "Reconciler: start to sync state" Nov 12 20:55:47.728809 kubelet[2230]: W1112 20:55:47.728480 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.145:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.728809 kubelet[2230]: E1112 20:55:47.728535 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.145:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.734026 kubelet[2230]: E1112 20:55:47.733971 2230 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.145:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.145:6443: connect: connection refused" interval="200ms" Nov 12 20:55:47.734149 kubelet[2230]: I1112 20:55:47.734128 2230 factory.go:221] Registration of the systemd container factory successfully Nov 12 20:55:47.734208 kubelet[2230]: I1112 20:55:47.734191 2230 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 12 20:55:47.736164 kubelet[2230]: I1112 20:55:47.736127 2230 factory.go:221] Registration of the containerd container factory successfully Nov 12 20:55:47.793303 kubelet[2230]: I1112 20:55:47.793110 2230 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 12 20:55:47.793303 kubelet[2230]: E1112 20:55:47.793219 2230 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.145:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.145:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1807540190288030 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2024-11-12 20:55:47.437518896 +0000 UTC m=+2.882048948,LastTimestamp:2024-11-12 20:55:47.437518896 +0000 UTC m=+2.882048948,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Nov 12 20:55:47.795450 kubelet[2230]: I1112 20:55:47.795060 2230 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 12 20:55:47.795450 kubelet[2230]: I1112 20:55:47.795100 2230 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 12 20:55:47.795450 kubelet[2230]: I1112 20:55:47.795123 2230 kubelet.go:2337] "Starting kubelet main sync loop" Nov 12 20:55:47.795450 kubelet[2230]: E1112 20:55:47.795176 2230 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 12 20:55:47.796657 kubelet[2230]: W1112 20:55:47.796599 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.145:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.796717 kubelet[2230]: E1112 20:55:47.796671 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.145:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:47.806149 kubelet[2230]: I1112 20:55:47.806117 2230 cpu_manager.go:214] "Starting CPU manager" policy="none" Nov 12 20:55:47.806149 kubelet[2230]: I1112 20:55:47.806132 2230 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Nov 12 20:55:47.806149 kubelet[2230]: I1112 20:55:47.806151 2230 state_mem.go:36] "Initialized new in-memory state store" Nov 12 20:55:47.829672 kubelet[2230]: I1112 20:55:47.829631 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:55:47.830153 kubelet[2230]: E1112 20:55:47.830122 2230 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.145:6443/api/v1/nodes\": dial tcp 10.0.0.145:6443: connect: connection refused" node="localhost" Nov 12 20:55:47.895782 kubelet[2230]: E1112 20:55:47.895678 2230 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 12 20:55:47.934923 kubelet[2230]: E1112 20:55:47.934832 2230 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.145:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.145:6443: connect: connection refused" interval="400ms" Nov 12 20:55:47.949677 kubelet[2230]: I1112 20:55:47.949609 2230 policy_none.go:49] "None policy: Start" Nov 12 20:55:47.950895 kubelet[2230]: I1112 20:55:47.950855 2230 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 12 20:55:47.950895 kubelet[2230]: I1112 20:55:47.950895 2230 state_mem.go:35] "Initializing new in-memory state store" Nov 12 20:55:47.977993 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Nov 12 20:55:47.992373 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Nov 12 20:55:47.996466 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Nov 12 20:55:48.009426 kubelet[2230]: I1112 20:55:48.009362 2230 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 12 20:55:48.009699 kubelet[2230]: I1112 20:55:48.009645 2230 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 12 20:55:48.009854 kubelet[2230]: I1112 20:55:48.009784 2230 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 12 20:55:48.010989 kubelet[2230]: E1112 20:55:48.010954 2230 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Nov 12 20:55:48.031611 kubelet[2230]: I1112 20:55:48.031578 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:55:48.032104 kubelet[2230]: E1112 20:55:48.032055 2230 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.145:6443/api/v1/nodes\": dial tcp 10.0.0.145:6443: connect: connection refused" node="localhost" Nov 12 20:55:48.096351 kubelet[2230]: I1112 20:55:48.096269 2230 topology_manager.go:215] "Topology Admit Handler" podUID="e7d19f112e2b5d4d87fec04d3ab56a3d" podNamespace="kube-system" podName="kube-apiserver-localhost" Nov 12 20:55:48.098086 kubelet[2230]: I1112 20:55:48.098011 2230 topology_manager.go:215] "Topology Admit Handler" podUID="35a50a3f0f14abbdd3fae477f39e6e18" podNamespace="kube-system" podName="kube-controller-manager-localhost" Nov 12 20:55:48.099745 kubelet[2230]: I1112 20:55:48.099353 2230 topology_manager.go:215] "Topology Admit Handler" podUID="c95384ce7f39fb5cff38cd36dacf8a69" podNamespace="kube-system" podName="kube-scheduler-localhost" Nov 12 20:55:48.104603 systemd[1]: Created slice kubepods-burstable-pode7d19f112e2b5d4d87fec04d3ab56a3d.slice - libcontainer container kubepods-burstable-pode7d19f112e2b5d4d87fec04d3ab56a3d.slice. Nov 12 20:55:48.127793 systemd[1]: Created slice kubepods-burstable-pod35a50a3f0f14abbdd3fae477f39e6e18.slice - libcontainer container kubepods-burstable-pod35a50a3f0f14abbdd3fae477f39e6e18.slice. Nov 12 20:55:48.129751 kubelet[2230]: I1112 20:55:48.129723 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:55:48.129831 kubelet[2230]: I1112 20:55:48.129755 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:55:48.129831 kubelet[2230]: I1112 20:55:48.129780 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:55:48.129831 kubelet[2230]: I1112 20:55:48.129798 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:55:48.129831 kubelet[2230]: I1112 20:55:48.129817 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:55:48.129997 kubelet[2230]: I1112 20:55:48.129885 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c95384ce7f39fb5cff38cd36dacf8a69-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"c95384ce7f39fb5cff38cd36dacf8a69\") " pod="kube-system/kube-scheduler-localhost" Nov 12 20:55:48.129997 kubelet[2230]: I1112 20:55:48.129942 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:55:48.129997 kubelet[2230]: I1112 20:55:48.129981 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:55:48.130093 kubelet[2230]: I1112 20:55:48.130005 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:55:48.135880 systemd[1]: Created slice kubepods-burstable-podc95384ce7f39fb5cff38cd36dacf8a69.slice - libcontainer container kubepods-burstable-podc95384ce7f39fb5cff38cd36dacf8a69.slice. Nov 12 20:55:48.336508 kubelet[2230]: E1112 20:55:48.336359 2230 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.145:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.145:6443: connect: connection refused" interval="800ms" Nov 12 20:55:48.426254 kubelet[2230]: E1112 20:55:48.426212 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:48.426997 containerd[1444]: time="2024-11-12T20:55:48.426963884Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e7d19f112e2b5d4d87fec04d3ab56a3d,Namespace:kube-system,Attempt:0,}" Nov 12 20:55:48.433516 kubelet[2230]: I1112 20:55:48.433483 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:55:48.433667 kubelet[2230]: E1112 20:55:48.433649 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:48.433839 kubelet[2230]: E1112 20:55:48.433808 2230 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.145:6443/api/v1/nodes\": dial tcp 10.0.0.145:6443: connect: connection refused" node="localhost" Nov 12 20:55:48.434108 containerd[1444]: time="2024-11-12T20:55:48.434065406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:35a50a3f0f14abbdd3fae477f39e6e18,Namespace:kube-system,Attempt:0,}" Nov 12 20:55:48.442500 kubelet[2230]: E1112 20:55:48.442451 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:48.443104 containerd[1444]: time="2024-11-12T20:55:48.443060471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:c95384ce7f39fb5cff38cd36dacf8a69,Namespace:kube-system,Attempt:0,}" Nov 12 20:55:48.459714 kubelet[2230]: W1112 20:55:48.459638 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.145:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:48.459714 kubelet[2230]: E1112 20:55:48.459714 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.145:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:48.841926 kubelet[2230]: W1112 20:55:48.841871 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.145:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:48.841926 kubelet[2230]: E1112 20:55:48.841940 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.145:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:48.851629 kubelet[2230]: W1112 20:55:48.851548 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.145:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:48.851629 kubelet[2230]: E1112 20:55:48.851622 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.145:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:49.137297 kubelet[2230]: E1112 20:55:49.137225 2230 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.145:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.145:6443: connect: connection refused" interval="1.6s" Nov 12 20:55:49.236027 kubelet[2230]: I1112 20:55:49.235984 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:55:49.236470 kubelet[2230]: E1112 20:55:49.236417 2230 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.145:6443/api/v1/nodes\": dial tcp 10.0.0.145:6443: connect: connection refused" node="localhost" Nov 12 20:55:49.268293 kubelet[2230]: W1112 20:55:49.268203 2230 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.145:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:49.268293 kubelet[2230]: E1112 20:55:49.268294 2230 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.145:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:49.457252 kubelet[2230]: E1112 20:55:49.457080 2230 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.145:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.145:6443: connect: connection refused Nov 12 20:55:49.476506 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3931466145.mount: Deactivated successfully. Nov 12 20:55:49.487713 containerd[1444]: time="2024-11-12T20:55:49.487612886Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 12 20:55:49.489604 containerd[1444]: time="2024-11-12T20:55:49.489521970Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 12 20:55:49.490531 containerd[1444]: time="2024-11-12T20:55:49.490492085Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Nov 12 20:55:49.491856 containerd[1444]: time="2024-11-12T20:55:49.491808176Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 12 20:55:49.493397 containerd[1444]: time="2024-11-12T20:55:49.493332640Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Nov 12 20:55:49.495242 containerd[1444]: time="2024-11-12T20:55:49.495144148Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Nov 12 20:55:49.496579 containerd[1444]: time="2024-11-12T20:55:49.496513960Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 12 20:55:49.500931 containerd[1444]: time="2024-11-12T20:55:49.500832617Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Nov 12 20:55:49.502036 containerd[1444]: time="2024-11-12T20:55:49.501960192Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.058777623s" Nov 12 20:55:49.503490 containerd[1444]: time="2024-11-12T20:55:49.503446284Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.07635163s" Nov 12 20:55:49.504904 containerd[1444]: time="2024-11-12T20:55:49.504863319Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.070696876s" Nov 12 20:55:49.718959 containerd[1444]: time="2024-11-12T20:55:49.718598367Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:55:49.718959 containerd[1444]: time="2024-11-12T20:55:49.718742585Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:55:49.718959 containerd[1444]: time="2024-11-12T20:55:49.718768237Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.719865 containerd[1444]: time="2024-11-12T20:55:49.718927913Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.736106 containerd[1444]: time="2024-11-12T20:55:49.735448612Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:55:49.736106 containerd[1444]: time="2024-11-12T20:55:49.735558617Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:55:49.736106 containerd[1444]: time="2024-11-12T20:55:49.735580811Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.737349 containerd[1444]: time="2024-11-12T20:55:49.735656454Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.744107 systemd[1]: Started cri-containerd-030c48f7d6ac5fdc2e4cafccd675e2a4d79b7bef519fc1793a0300a4b76ea312.scope - libcontainer container 030c48f7d6ac5fdc2e4cafccd675e2a4d79b7bef519fc1793a0300a4b76ea312. Nov 12 20:55:49.745253 containerd[1444]: time="2024-11-12T20:55:49.745069254Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:55:49.745253 containerd[1444]: time="2024-11-12T20:55:49.745120749Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:55:49.745253 containerd[1444]: time="2024-11-12T20:55:49.745138722Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.745372 containerd[1444]: time="2024-11-12T20:55:49.745272265Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:55:49.767130 systemd[1]: Started cri-containerd-97fec2c0ef023f8b579fa16866476277364fd9118aa019f15bff3d31e0e53338.scope - libcontainer container 97fec2c0ef023f8b579fa16866476277364fd9118aa019f15bff3d31e0e53338. Nov 12 20:55:49.773475 systemd[1]: Started cri-containerd-17e08fae114bca8280aac4d85872e729deb06ee276c7a19e5b4366c88bc8a34b.scope - libcontainer container 17e08fae114bca8280aac4d85872e729deb06ee276c7a19e5b4366c88bc8a34b. Nov 12 20:55:49.825813 containerd[1444]: time="2024-11-12T20:55:49.825764547Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:c95384ce7f39fb5cff38cd36dacf8a69,Namespace:kube-system,Attempt:0,} returns sandbox id \"030c48f7d6ac5fdc2e4cafccd675e2a4d79b7bef519fc1793a0300a4b76ea312\"" Nov 12 20:55:49.827385 kubelet[2230]: E1112 20:55:49.827362 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:49.830415 containerd[1444]: time="2024-11-12T20:55:49.830271810Z" level=info msg="CreateContainer within sandbox \"030c48f7d6ac5fdc2e4cafccd675e2a4d79b7bef519fc1793a0300a4b76ea312\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Nov 12 20:55:49.852612 containerd[1444]: time="2024-11-12T20:55:49.852563898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e7d19f112e2b5d4d87fec04d3ab56a3d,Namespace:kube-system,Attempt:0,} returns sandbox id \"97fec2c0ef023f8b579fa16866476277364fd9118aa019f15bff3d31e0e53338\"" Nov 12 20:55:49.853826 kubelet[2230]: E1112 20:55:49.853787 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:49.857232 containerd[1444]: time="2024-11-12T20:55:49.857197598Z" level=info msg="CreateContainer within sandbox \"97fec2c0ef023f8b579fa16866476277364fd9118aa019f15bff3d31e0e53338\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Nov 12 20:55:49.859369 containerd[1444]: time="2024-11-12T20:55:49.859245207Z" level=info msg="CreateContainer within sandbox \"030c48f7d6ac5fdc2e4cafccd675e2a4d79b7bef519fc1793a0300a4b76ea312\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"382238562847591b0c9d6c4c80864154ba5a4caac4f81baa3aa470c4b34c6cb8\"" Nov 12 20:55:49.860079 containerd[1444]: time="2024-11-12T20:55:49.860054924Z" level=info msg="StartContainer for \"382238562847591b0c9d6c4c80864154ba5a4caac4f81baa3aa470c4b34c6cb8\"" Nov 12 20:55:49.870403 containerd[1444]: time="2024-11-12T20:55:49.870354908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:35a50a3f0f14abbdd3fae477f39e6e18,Namespace:kube-system,Attempt:0,} returns sandbox id \"17e08fae114bca8280aac4d85872e729deb06ee276c7a19e5b4366c88bc8a34b\"" Nov 12 20:55:49.871323 kubelet[2230]: E1112 20:55:49.871299 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:49.873548 containerd[1444]: time="2024-11-12T20:55:49.873422763Z" level=info msg="CreateContainer within sandbox \"17e08fae114bca8280aac4d85872e729deb06ee276c7a19e5b4366c88bc8a34b\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Nov 12 20:55:49.877728 containerd[1444]: time="2024-11-12T20:55:49.877695690Z" level=info msg="CreateContainer within sandbox \"97fec2c0ef023f8b579fa16866476277364fd9118aa019f15bff3d31e0e53338\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"075cb148a579155efe3e152ae840dc29d21a49cbbde7809a8b1afb28034aac11\"" Nov 12 20:55:49.878835 containerd[1444]: time="2024-11-12T20:55:49.878792982Z" level=info msg="StartContainer for \"075cb148a579155efe3e152ae840dc29d21a49cbbde7809a8b1afb28034aac11\"" Nov 12 20:55:49.896135 systemd[1]: Started cri-containerd-382238562847591b0c9d6c4c80864154ba5a4caac4f81baa3aa470c4b34c6cb8.scope - libcontainer container 382238562847591b0c9d6c4c80864154ba5a4caac4f81baa3aa470c4b34c6cb8. Nov 12 20:55:49.914363 systemd[1]: Started cri-containerd-075cb148a579155efe3e152ae840dc29d21a49cbbde7809a8b1afb28034aac11.scope - libcontainer container 075cb148a579155efe3e152ae840dc29d21a49cbbde7809a8b1afb28034aac11. Nov 12 20:55:49.925146 containerd[1444]: time="2024-11-12T20:55:49.925066430Z" level=info msg="CreateContainer within sandbox \"17e08fae114bca8280aac4d85872e729deb06ee276c7a19e5b4366c88bc8a34b\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"178c7d3c4c749512b79b81a86084477ea026d351c963b4587580f88293e47455\"" Nov 12 20:55:49.926264 containerd[1444]: time="2024-11-12T20:55:49.925952983Z" level=info msg="StartContainer for \"178c7d3c4c749512b79b81a86084477ea026d351c963b4587580f88293e47455\"" Nov 12 20:55:49.952289 containerd[1444]: time="2024-11-12T20:55:49.952230394Z" level=info msg="StartContainer for \"382238562847591b0c9d6c4c80864154ba5a4caac4f81baa3aa470c4b34c6cb8\" returns successfully" Nov 12 20:55:49.971123 systemd[1]: Started cri-containerd-178c7d3c4c749512b79b81a86084477ea026d351c963b4587580f88293e47455.scope - libcontainer container 178c7d3c4c749512b79b81a86084477ea026d351c963b4587580f88293e47455. Nov 12 20:55:49.976321 containerd[1444]: time="2024-11-12T20:55:49.976257395Z" level=info msg="StartContainer for \"075cb148a579155efe3e152ae840dc29d21a49cbbde7809a8b1afb28034aac11\" returns successfully" Nov 12 20:55:50.028333 containerd[1444]: time="2024-11-12T20:55:50.028267654Z" level=info msg="StartContainer for \"178c7d3c4c749512b79b81a86084477ea026d351c963b4587580f88293e47455\" returns successfully" Nov 12 20:55:50.804700 kubelet[2230]: E1112 20:55:50.804622 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:50.806331 kubelet[2230]: E1112 20:55:50.806294 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:50.807935 kubelet[2230]: E1112 20:55:50.807903 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:50.838807 kubelet[2230]: I1112 20:55:50.838719 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:55:51.242752 kubelet[2230]: E1112 20:55:51.241095 2230 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Nov 12 20:55:51.423217 kubelet[2230]: I1112 20:55:51.423142 2230 apiserver.go:52] "Watching apiserver" Nov 12 20:55:51.429021 kubelet[2230]: I1112 20:55:51.428985 2230 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Nov 12 20:55:51.437671 kubelet[2230]: I1112 20:55:51.437612 2230 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Nov 12 20:55:51.820383 kubelet[2230]: E1112 20:55:51.820327 2230 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Nov 12 20:55:51.820897 kubelet[2230]: E1112 20:55:51.820868 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:52.941430 update_engine[1434]: I20241112 20:55:52.941320 1434 update_attempter.cc:509] Updating boot flags... Nov 12 20:55:53.173961 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2512) Nov 12 20:55:53.312950 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2514) Nov 12 20:55:53.478079 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2514) Nov 12 20:55:55.918222 kubelet[2230]: E1112 20:55:55.918182 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:56.272264 kubelet[2230]: E1112 20:55:56.272144 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:56.816294 kubelet[2230]: E1112 20:55:56.816262 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:56.816445 kubelet[2230]: E1112 20:55:56.816397 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:58.246566 kubelet[2230]: I1112 20:55:58.246480 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.246449067 podStartE2EDuration="3.246449067s" podCreationTimestamp="2024-11-12 20:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:55:57.96071106 +0000 UTC m=+13.405241112" watchObservedRunningTime="2024-11-12 20:55:58.246449067 +0000 UTC m=+13.690979119" Nov 12 20:55:58.247083 kubelet[2230]: I1112 20:55:58.246604 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=3.246600704 podStartE2EDuration="3.246600704s" podCreationTimestamp="2024-11-12 20:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:55:58.246392171 +0000 UTC m=+13.690922223" watchObservedRunningTime="2024-11-12 20:55:58.246600704 +0000 UTC m=+13.691130746" Nov 12 20:55:58.842184 kubelet[2230]: E1112 20:55:58.842142 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:55:59.821842 kubelet[2230]: E1112 20:55:59.821770 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:01.370556 systemd[1]: Reloading requested from client PID 2526 ('systemctl') (unit session-7.scope)... Nov 12 20:56:01.370573 systemd[1]: Reloading... Nov 12 20:56:01.444971 zram_generator::config[2563]: No configuration found. Nov 12 20:56:01.567836 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Nov 12 20:56:01.660421 systemd[1]: Reloading finished in 289 ms. Nov 12 20:56:01.707639 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:56:01.723207 systemd[1]: kubelet.service: Deactivated successfully. Nov 12 20:56:01.723473 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:56:01.723527 systemd[1]: kubelet.service: Consumed 1.195s CPU time, 120.3M memory peak, 0B memory swap peak. Nov 12 20:56:01.734105 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Nov 12 20:56:01.915017 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Nov 12 20:56:01.919817 (kubelet)[2610]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Nov 12 20:56:01.986766 kubelet[2610]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 12 20:56:01.986766 kubelet[2610]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 12 20:56:01.986766 kubelet[2610]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 12 20:56:01.987112 kubelet[2610]: I1112 20:56:01.986802 2610 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 12 20:56:01.991366 kubelet[2610]: I1112 20:56:01.991331 2610 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Nov 12 20:56:01.991366 kubelet[2610]: I1112 20:56:01.991360 2610 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 12 20:56:01.991596 kubelet[2610]: I1112 20:56:01.991574 2610 server.go:927] "Client rotation is on, will bootstrap in background" Nov 12 20:56:01.993079 kubelet[2610]: I1112 20:56:01.993010 2610 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 12 20:56:01.994333 kubelet[2610]: I1112 20:56:01.994294 2610 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Nov 12 20:56:02.002993 kubelet[2610]: I1112 20:56:02.002646 2610 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Nov 12 20:56:02.003190 kubelet[2610]: I1112 20:56:02.003120 2610 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 12 20:56:02.005134 kubelet[2610]: I1112 20:56:02.003153 2610 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Nov 12 20:56:02.005134 kubelet[2610]: I1112 20:56:02.003497 2610 topology_manager.go:138] "Creating topology manager with none policy" Nov 12 20:56:02.005134 kubelet[2610]: I1112 20:56:02.003510 2610 container_manager_linux.go:301] "Creating device plugin manager" Nov 12 20:56:02.005134 kubelet[2610]: I1112 20:56:02.003562 2610 state_mem.go:36] "Initialized new in-memory state store" Nov 12 20:56:02.005134 kubelet[2610]: I1112 20:56:02.003675 2610 kubelet.go:400] "Attempting to sync node with API server" Nov 12 20:56:02.005378 kubelet[2610]: I1112 20:56:02.003688 2610 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 12 20:56:02.005378 kubelet[2610]: I1112 20:56:02.003713 2610 kubelet.go:312] "Adding apiserver pod source" Nov 12 20:56:02.005378 kubelet[2610]: I1112 20:56:02.003738 2610 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 12 20:56:02.006101 kubelet[2610]: I1112 20:56:02.006078 2610 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Nov 12 20:56:02.006329 kubelet[2610]: I1112 20:56:02.006313 2610 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 12 20:56:02.006802 kubelet[2610]: I1112 20:56:02.006784 2610 server.go:1264] "Started kubelet" Nov 12 20:56:02.008511 kubelet[2610]: I1112 20:56:02.008480 2610 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 12 20:56:02.010202 kubelet[2610]: I1112 20:56:02.010180 2610 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 12 20:56:02.010982 kubelet[2610]: I1112 20:56:02.010967 2610 volume_manager.go:291] "Starting Kubelet Volume Manager" Nov 12 20:56:02.011163 kubelet[2610]: I1112 20:56:02.011144 2610 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Nov 12 20:56:02.011290 kubelet[2610]: I1112 20:56:02.011278 2610 reconciler.go:26] "Reconciler: start to sync state" Nov 12 20:56:02.015617 kubelet[2610]: I1112 20:56:02.015589 2610 factory.go:221] Registration of the systemd container factory successfully Nov 12 20:56:02.017752 kubelet[2610]: I1112 20:56:02.016924 2610 server.go:455] "Adding debug handlers to kubelet server" Nov 12 20:56:02.018304 kubelet[2610]: I1112 20:56:02.018240 2610 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 12 20:56:02.018573 kubelet[2610]: I1112 20:56:02.018546 2610 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 12 20:56:02.018864 kubelet[2610]: I1112 20:56:02.018838 2610 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Nov 12 20:56:02.024110 kubelet[2610]: I1112 20:56:02.024048 2610 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 12 20:56:02.028089 kubelet[2610]: I1112 20:56:02.028058 2610 factory.go:221] Registration of the containerd container factory successfully Nov 12 20:56:02.030379 kubelet[2610]: I1112 20:56:02.030337 2610 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 12 20:56:02.030476 kubelet[2610]: I1112 20:56:02.030388 2610 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 12 20:56:02.030476 kubelet[2610]: I1112 20:56:02.030420 2610 kubelet.go:2337] "Starting kubelet main sync loop" Nov 12 20:56:02.030530 kubelet[2610]: E1112 20:56:02.030475 2610 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 12 20:56:02.035314 kubelet[2610]: E1112 20:56:02.035280 2610 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Nov 12 20:56:02.066216 kubelet[2610]: I1112 20:56:02.066171 2610 cpu_manager.go:214] "Starting CPU manager" policy="none" Nov 12 20:56:02.066216 kubelet[2610]: I1112 20:56:02.066191 2610 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Nov 12 20:56:02.066216 kubelet[2610]: I1112 20:56:02.066213 2610 state_mem.go:36] "Initialized new in-memory state store" Nov 12 20:56:02.066460 kubelet[2610]: I1112 20:56:02.066439 2610 state_mem.go:88] "Updated default CPUSet" cpuSet="" Nov 12 20:56:02.066511 kubelet[2610]: I1112 20:56:02.066456 2610 state_mem.go:96] "Updated CPUSet assignments" assignments={} Nov 12 20:56:02.066511 kubelet[2610]: I1112 20:56:02.066479 2610 policy_none.go:49] "None policy: Start" Nov 12 20:56:02.067098 kubelet[2610]: I1112 20:56:02.067023 2610 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 12 20:56:02.067098 kubelet[2610]: I1112 20:56:02.067041 2610 state_mem.go:35] "Initializing new in-memory state store" Nov 12 20:56:02.067225 kubelet[2610]: I1112 20:56:02.067187 2610 state_mem.go:75] "Updated machine memory state" Nov 12 20:56:02.071313 kubelet[2610]: I1112 20:56:02.071281 2610 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 12 20:56:02.071626 kubelet[2610]: I1112 20:56:02.071495 2610 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 12 20:56:02.071626 kubelet[2610]: I1112 20:56:02.071615 2610 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 12 20:56:02.115364 kubelet[2610]: I1112 20:56:02.115306 2610 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Nov 12 20:56:02.125466 kubelet[2610]: I1112 20:56:02.125426 2610 kubelet_node_status.go:112] "Node was previously registered" node="localhost" Nov 12 20:56:02.125596 kubelet[2610]: I1112 20:56:02.125530 2610 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Nov 12 20:56:02.131646 kubelet[2610]: I1112 20:56:02.131608 2610 topology_manager.go:215] "Topology Admit Handler" podUID="e7d19f112e2b5d4d87fec04d3ab56a3d" podNamespace="kube-system" podName="kube-apiserver-localhost" Nov 12 20:56:02.131791 kubelet[2610]: I1112 20:56:02.131711 2610 topology_manager.go:215] "Topology Admit Handler" podUID="35a50a3f0f14abbdd3fae477f39e6e18" podNamespace="kube-system" podName="kube-controller-manager-localhost" Nov 12 20:56:02.131791 kubelet[2610]: I1112 20:56:02.131766 2610 topology_manager.go:215] "Topology Admit Handler" podUID="c95384ce7f39fb5cff38cd36dacf8a69" podNamespace="kube-system" podName="kube-scheduler-localhost" Nov 12 20:56:02.137490 kubelet[2610]: E1112 20:56:02.137258 2610 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Nov 12 20:56:02.137872 kubelet[2610]: E1112 20:56:02.137534 2610 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Nov 12 20:56:02.137872 kubelet[2610]: E1112 20:56:02.137804 2610 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.159785 sudo[2643]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Nov 12 20:56:02.160145 sudo[2643]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Nov 12 20:56:02.312970 kubelet[2610]: I1112 20:56:02.312795 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.312970 kubelet[2610]: I1112 20:56:02.312844 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.312970 kubelet[2610]: I1112 20:56:02.312869 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c95384ce7f39fb5cff38cd36dacf8a69-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"c95384ce7f39fb5cff38cd36dacf8a69\") " pod="kube-system/kube-scheduler-localhost" Nov 12 20:56:02.312970 kubelet[2610]: I1112 20:56:02.312885 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:56:02.312970 kubelet[2610]: I1112 20:56:02.312899 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:56:02.313219 kubelet[2610]: I1112 20:56:02.312941 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.313219 kubelet[2610]: I1112 20:56:02.312955 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e7d19f112e2b5d4d87fec04d3ab56a3d-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e7d19f112e2b5d4d87fec04d3ab56a3d\") " pod="kube-system/kube-apiserver-localhost" Nov 12 20:56:02.313219 kubelet[2610]: I1112 20:56:02.312977 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.313219 kubelet[2610]: I1112 20:56:02.313014 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/35a50a3f0f14abbdd3fae477f39e6e18-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"35a50a3f0f14abbdd3fae477f39e6e18\") " pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:02.438947 kubelet[2610]: E1112 20:56:02.438876 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:02.441932 kubelet[2610]: E1112 20:56:02.439390 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:02.442561 kubelet[2610]: E1112 20:56:02.442515 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:02.636696 sudo[2643]: pam_unix(sudo:session): session closed for user root Nov 12 20:56:03.005569 kubelet[2610]: I1112 20:56:03.005430 2610 apiserver.go:52] "Watching apiserver" Nov 12 20:56:03.011439 kubelet[2610]: I1112 20:56:03.011401 2610 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Nov 12 20:56:03.045716 kubelet[2610]: E1112 20:56:03.045671 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:03.053562 kubelet[2610]: E1112 20:56:03.053522 2610 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Nov 12 20:56:03.054084 kubelet[2610]: E1112 20:56:03.054015 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:03.054566 kubelet[2610]: E1112 20:56:03.054512 2610 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Nov 12 20:56:03.055192 kubelet[2610]: E1112 20:56:03.055161 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:03.081660 kubelet[2610]: I1112 20:56:03.081574 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=5.081551897 podStartE2EDuration="5.081551897s" podCreationTimestamp="2024-11-12 20:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:56:03.073015366 +0000 UTC m=+1.148692170" watchObservedRunningTime="2024-11-12 20:56:03.081551897 +0000 UTC m=+1.157228701" Nov 12 20:56:04.047218 kubelet[2610]: E1112 20:56:04.047164 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:04.047859 kubelet[2610]: E1112 20:56:04.047338 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:04.387939 sudo[1632]: pam_unix(sudo:session): session closed for user root Nov 12 20:56:04.397594 sshd[1628]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:04.402903 systemd[1]: sshd@6-10.0.0.145:22-10.0.0.1:59410.service: Deactivated successfully. Nov 12 20:56:04.405113 systemd[1]: session-7.scope: Deactivated successfully. Nov 12 20:56:04.405314 systemd[1]: session-7.scope: Consumed 5.618s CPU time, 193.7M memory peak, 0B memory swap peak. Nov 12 20:56:04.405904 systemd-logind[1433]: Session 7 logged out. Waiting for processes to exit. Nov 12 20:56:04.407193 systemd-logind[1433]: Removed session 7. Nov 12 20:56:05.048386 kubelet[2610]: E1112 20:56:05.048338 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:06.173988 kubelet[2610]: E1112 20:56:06.173949 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:07.050888 kubelet[2610]: E1112 20:56:07.050840 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:08.051482 kubelet[2610]: E1112 20:56:08.051439 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:10.400991 kubelet[2610]: E1112 20:56:10.400886 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:11.055612 kubelet[2610]: E1112 20:56:11.055575 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:12.803344 kubelet[2610]: E1112 20:56:12.803313 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:13.912279 kubelet[2610]: I1112 20:56:13.912243 2610 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Nov 12 20:56:13.912814 containerd[1444]: time="2024-11-12T20:56:13.912770631Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Nov 12 20:56:13.913095 kubelet[2610]: I1112 20:56:13.912984 2610 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Nov 12 20:56:14.836722 kubelet[2610]: I1112 20:56:14.836605 2610 topology_manager.go:215] "Topology Admit Handler" podUID="f417285e-856b-43a2-aa17-48432b9b5631" podNamespace="kube-system" podName="kube-proxy-xw97b" Nov 12 20:56:14.845018 kubelet[2610]: I1112 20:56:14.843991 2610 topology_manager.go:215] "Topology Admit Handler" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" podNamespace="kube-system" podName="cilium-wh5nj" Nov 12 20:56:14.853128 systemd[1]: Created slice kubepods-besteffort-podf417285e_856b_43a2_aa17_48432b9b5631.slice - libcontainer container kubepods-besteffort-podf417285e_856b_43a2_aa17_48432b9b5631.slice. Nov 12 20:56:14.878100 systemd[1]: Created slice kubepods-burstable-pod224c5d84_0aec_401b_bfd3_b3f282e69424.slice - libcontainer container kubepods-burstable-pod224c5d84_0aec_401b_bfd3_b3f282e69424.slice. Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884485 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f417285e-856b-43a2-aa17-48432b9b5631-lib-modules\") pod \"kube-proxy-xw97b\" (UID: \"f417285e-856b-43a2-aa17-48432b9b5631\") " pod="kube-system/kube-proxy-xw97b" Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884526 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-config-path\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884543 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-net\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884556 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f417285e-856b-43a2-aa17-48432b9b5631-kube-proxy\") pod \"kube-proxy-xw97b\" (UID: \"f417285e-856b-43a2-aa17-48432b9b5631\") " pod="kube-system/kube-proxy-xw97b" Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884568 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-run\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.884577 kubelet[2610]: I1112 20:56:14.884587 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-xtables-lock\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885071 kubelet[2610]: I1112 20:56:14.884600 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-kernel\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885071 kubelet[2610]: I1112 20:56:14.884614 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchjp\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-kube-api-access-dchjp\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885071 kubelet[2610]: I1112 20:56:14.884630 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8kvw\" (UniqueName: \"kubernetes.io/projected/f417285e-856b-43a2-aa17-48432b9b5631-kube-api-access-d8kvw\") pod \"kube-proxy-xw97b\" (UID: \"f417285e-856b-43a2-aa17-48432b9b5631\") " pod="kube-system/kube-proxy-xw97b" Nov 12 20:56:14.885071 kubelet[2610]: I1112 20:56:14.884647 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-lib-modules\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885071 kubelet[2610]: I1112 20:56:14.884660 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-hubble-tls\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884674 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/224c5d84-0aec-401b-bfd3-b3f282e69424-clustermesh-secrets\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884688 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-hostproc\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884700 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cni-path\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884717 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-etc-cni-netd\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884730 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-cgroup\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:14.885246 kubelet[2610]: I1112 20:56:14.884742 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f417285e-856b-43a2-aa17-48432b9b5631-xtables-lock\") pod \"kube-proxy-xw97b\" (UID: \"f417285e-856b-43a2-aa17-48432b9b5631\") " pod="kube-system/kube-proxy-xw97b" Nov 12 20:56:14.885441 kubelet[2610]: I1112 20:56:14.884758 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-bpf-maps\") pod \"cilium-wh5nj\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " pod="kube-system/cilium-wh5nj" Nov 12 20:56:15.196655 kubelet[2610]: I1112 20:56:15.196601 2610 topology_manager.go:215] "Topology Admit Handler" podUID="69dc6592-b201-4942-a38d-29a1e86e8cc9" podNamespace="kube-system" podName="cilium-operator-599987898-82wgj" Nov 12 20:56:15.203267 systemd[1]: Created slice kubepods-besteffort-pod69dc6592_b201_4942_a38d_29a1e86e8cc9.slice - libcontainer container kubepods-besteffort-pod69dc6592_b201_4942_a38d_29a1e86e8cc9.slice. Nov 12 20:56:15.387735 kubelet[2610]: I1112 20:56:15.387677 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wmld\" (UniqueName: \"kubernetes.io/projected/69dc6592-b201-4942-a38d-29a1e86e8cc9-kube-api-access-6wmld\") pod \"cilium-operator-599987898-82wgj\" (UID: \"69dc6592-b201-4942-a38d-29a1e86e8cc9\") " pod="kube-system/cilium-operator-599987898-82wgj" Nov 12 20:56:15.387735 kubelet[2610]: I1112 20:56:15.387741 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69dc6592-b201-4942-a38d-29a1e86e8cc9-cilium-config-path\") pod \"cilium-operator-599987898-82wgj\" (UID: \"69dc6592-b201-4942-a38d-29a1e86e8cc9\") " pod="kube-system/cilium-operator-599987898-82wgj" Nov 12 20:56:15.772057 kubelet[2610]: E1112 20:56:15.772016 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:15.772644 containerd[1444]: time="2024-11-12T20:56:15.772580828Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xw97b,Uid:f417285e-856b-43a2-aa17-48432b9b5631,Namespace:kube-system,Attempt:0,}" Nov 12 20:56:15.783114 kubelet[2610]: E1112 20:56:15.783075 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:15.783614 containerd[1444]: time="2024-11-12T20:56:15.783581797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wh5nj,Uid:224c5d84-0aec-401b-bfd3-b3f282e69424,Namespace:kube-system,Attempt:0,}" Nov 12 20:56:15.806386 kubelet[2610]: E1112 20:56:15.806352 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:15.807027 containerd[1444]: time="2024-11-12T20:56:15.806972654Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-82wgj,Uid:69dc6592-b201-4942-a38d-29a1e86e8cc9,Namespace:kube-system,Attempt:0,}" Nov 12 20:56:16.562831 containerd[1444]: time="2024-11-12T20:56:16.561694553Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:56:16.562831 containerd[1444]: time="2024-11-12T20:56:16.562761744Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:56:16.562831 containerd[1444]: time="2024-11-12T20:56:16.562777145Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.563206 containerd[1444]: time="2024-11-12T20:56:16.562900318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.574292 containerd[1444]: time="2024-11-12T20:56:16.574207550Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:56:16.574870 containerd[1444]: time="2024-11-12T20:56:16.574274788Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:56:16.574870 containerd[1444]: time="2024-11-12T20:56:16.574297214Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.574870 containerd[1444]: time="2024-11-12T20:56:16.574396217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.586942 containerd[1444]: time="2024-11-12T20:56:16.586310555Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:56:16.586942 containerd[1444]: time="2024-11-12T20:56:16.586417985Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:56:16.586942 containerd[1444]: time="2024-11-12T20:56:16.586432656Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.586942 containerd[1444]: time="2024-11-12T20:56:16.586568765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:16.591153 systemd[1]: Started cri-containerd-f82fe13e34524771c01bee2029e6c74e1be9df4420711ad037d5f389a9da0b0a.scope - libcontainer container f82fe13e34524771c01bee2029e6c74e1be9df4420711ad037d5f389a9da0b0a. Nov 12 20:56:16.601194 systemd[1]: Started cri-containerd-d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4.scope - libcontainer container d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4. Nov 12 20:56:16.618283 systemd[1]: Started cri-containerd-6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb.scope - libcontainer container 6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb. Nov 12 20:56:16.626968 containerd[1444]: time="2024-11-12T20:56:16.626921631Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xw97b,Uid:f417285e-856b-43a2-aa17-48432b9b5631,Namespace:kube-system,Attempt:0,} returns sandbox id \"f82fe13e34524771c01bee2029e6c74e1be9df4420711ad037d5f389a9da0b0a\"" Nov 12 20:56:16.627893 kubelet[2610]: E1112 20:56:16.627784 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:16.631700 containerd[1444]: time="2024-11-12T20:56:16.631590294Z" level=info msg="CreateContainer within sandbox \"f82fe13e34524771c01bee2029e6c74e1be9df4420711ad037d5f389a9da0b0a\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Nov 12 20:56:16.641185 containerd[1444]: time="2024-11-12T20:56:16.641119015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wh5nj,Uid:224c5d84-0aec-401b-bfd3-b3f282e69424,Namespace:kube-system,Attempt:0,} returns sandbox id \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\"" Nov 12 20:56:16.642741 kubelet[2610]: E1112 20:56:16.642121 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:16.644348 containerd[1444]: time="2024-11-12T20:56:16.644284383Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Nov 12 20:56:16.661326 containerd[1444]: time="2024-11-12T20:56:16.661263958Z" level=info msg="CreateContainer within sandbox \"f82fe13e34524771c01bee2029e6c74e1be9df4420711ad037d5f389a9da0b0a\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"b2b47cc56b510b1aa11f5ab46113ec624cc6905095a848d535f473e0dce5029c\"" Nov 12 20:56:16.662214 containerd[1444]: time="2024-11-12T20:56:16.662142161Z" level=info msg="StartContainer for \"b2b47cc56b510b1aa11f5ab46113ec624cc6905095a848d535f473e0dce5029c\"" Nov 12 20:56:16.675129 containerd[1444]: time="2024-11-12T20:56:16.675056261Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-82wgj,Uid:69dc6592-b201-4942-a38d-29a1e86e8cc9,Namespace:kube-system,Attempt:0,} returns sandbox id \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\"" Nov 12 20:56:16.676496 kubelet[2610]: E1112 20:56:16.676468 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:16.703331 systemd[1]: Started cri-containerd-b2b47cc56b510b1aa11f5ab46113ec624cc6905095a848d535f473e0dce5029c.scope - libcontainer container b2b47cc56b510b1aa11f5ab46113ec624cc6905095a848d535f473e0dce5029c. Nov 12 20:56:16.743307 containerd[1444]: time="2024-11-12T20:56:16.743240659Z" level=info msg="StartContainer for \"b2b47cc56b510b1aa11f5ab46113ec624cc6905095a848d535f473e0dce5029c\" returns successfully" Nov 12 20:56:17.069611 kubelet[2610]: E1112 20:56:17.069556 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:27.783097 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4133039132.mount: Deactivated successfully. Nov 12 20:56:31.799027 systemd[1]: Started sshd@7-10.0.0.145:22-10.0.0.1:42148.service - OpenSSH per-connection server daemon (10.0.0.1:42148). Nov 12 20:56:33.460116 containerd[1444]: time="2024-11-12T20:56:33.460014174Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:56:33.461471 containerd[1444]: time="2024-11-12T20:56:33.460662331Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735315" Nov 12 20:56:33.461643 sshd[3000]: Accepted publickey for core from 10.0.0.1 port 42148 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:33.462621 containerd[1444]: time="2024-11-12T20:56:33.462573335Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:56:33.463982 sshd[3000]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:33.464687 containerd[1444]: time="2024-11-12T20:56:33.464607326Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 16.820222508s" Nov 12 20:56:33.464687 containerd[1444]: time="2024-11-12T20:56:33.464665212Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Nov 12 20:56:33.466311 containerd[1444]: time="2024-11-12T20:56:33.466277779Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Nov 12 20:56:33.469254 containerd[1444]: time="2024-11-12T20:56:33.469206551Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Nov 12 20:56:33.470037 systemd-logind[1433]: New session 8 of user core. Nov 12 20:56:33.476138 systemd[1]: Started session-8.scope - Session 8 of User core. Nov 12 20:56:33.483153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2507854461.mount: Deactivated successfully. Nov 12 20:56:33.486230 containerd[1444]: time="2024-11-12T20:56:33.486182384Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\"" Nov 12 20:56:33.486935 containerd[1444]: time="2024-11-12T20:56:33.486880511Z" level=info msg="StartContainer for \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\"" Nov 12 20:56:33.525213 systemd[1]: Started cri-containerd-a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98.scope - libcontainer container a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98. Nov 12 20:56:33.560259 containerd[1444]: time="2024-11-12T20:56:33.560194523Z" level=info msg="StartContainer for \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\" returns successfully" Nov 12 20:56:33.572340 systemd[1]: cri-containerd-a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98.scope: Deactivated successfully. Nov 12 20:56:33.637215 sshd[3000]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:33.641476 systemd[1]: sshd@7-10.0.0.145:22-10.0.0.1:42148.service: Deactivated successfully. Nov 12 20:56:33.643811 systemd[1]: session-8.scope: Deactivated successfully. Nov 12 20:56:33.644387 systemd-logind[1433]: Session 8 logged out. Waiting for processes to exit. Nov 12 20:56:33.645259 systemd-logind[1433]: Removed session 8. Nov 12 20:56:34.408467 kubelet[2610]: E1112 20:56:34.408405 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:34.480440 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98-rootfs.mount: Deactivated successfully. Nov 12 20:56:34.670303 kubelet[2610]: I1112 20:56:34.670136 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-xw97b" podStartSLOduration=20.670076606 podStartE2EDuration="20.670076606s" podCreationTimestamp="2024-11-12 20:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:56:17.085510708 +0000 UTC m=+15.161187542" watchObservedRunningTime="2024-11-12 20:56:34.670076606 +0000 UTC m=+32.745753420" Nov 12 20:56:34.741368 containerd[1444]: time="2024-11-12T20:56:34.736684110Z" level=info msg="shim disconnected" id=a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98 namespace=k8s.io Nov 12 20:56:34.741975 containerd[1444]: time="2024-11-12T20:56:34.741371142Z" level=warning msg="cleaning up after shim disconnected" id=a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98 namespace=k8s.io Nov 12 20:56:34.741975 containerd[1444]: time="2024-11-12T20:56:34.741401061Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:56:35.412270 kubelet[2610]: E1112 20:56:35.412169 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:35.415056 containerd[1444]: time="2024-11-12T20:56:35.415012446Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Nov 12 20:56:35.601306 containerd[1444]: time="2024-11-12T20:56:35.601224397Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\"" Nov 12 20:56:35.601979 containerd[1444]: time="2024-11-12T20:56:35.601957973Z" level=info msg="StartContainer for \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\"" Nov 12 20:56:35.641526 systemd[1]: Started cri-containerd-acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180.scope - libcontainer container acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180. Nov 12 20:56:35.672125 containerd[1444]: time="2024-11-12T20:56:35.671992954Z" level=info msg="StartContainer for \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\" returns successfully" Nov 12 20:56:35.685042 systemd[1]: systemd-sysctl.service: Deactivated successfully. Nov 12 20:56:35.685561 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Nov 12 20:56:35.685663 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Nov 12 20:56:35.694521 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Nov 12 20:56:35.694883 systemd[1]: cri-containerd-acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180.scope: Deactivated successfully. Nov 12 20:56:35.708983 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180-rootfs.mount: Deactivated successfully. Nov 12 20:56:35.713015 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Nov 12 20:56:35.728889 containerd[1444]: time="2024-11-12T20:56:35.728815564Z" level=info msg="shim disconnected" id=acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180 namespace=k8s.io Nov 12 20:56:35.728889 containerd[1444]: time="2024-11-12T20:56:35.728887579Z" level=warning msg="cleaning up after shim disconnected" id=acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180 namespace=k8s.io Nov 12 20:56:35.729088 containerd[1444]: time="2024-11-12T20:56:35.728897959Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:56:36.415672 kubelet[2610]: E1112 20:56:36.415624 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:36.418362 containerd[1444]: time="2024-11-12T20:56:36.418316879Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Nov 12 20:56:36.452967 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1872070238.mount: Deactivated successfully. Nov 12 20:56:36.458316 containerd[1444]: time="2024-11-12T20:56:36.458240476Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\"" Nov 12 20:56:36.459054 containerd[1444]: time="2024-11-12T20:56:36.459023379Z" level=info msg="StartContainer for \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\"" Nov 12 20:56:36.490166 systemd[1]: Started cri-containerd-c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0.scope - libcontainer container c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0. Nov 12 20:56:36.526152 containerd[1444]: time="2024-11-12T20:56:36.525638623Z" level=info msg="StartContainer for \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\" returns successfully" Nov 12 20:56:36.531178 systemd[1]: cri-containerd-c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0.scope: Deactivated successfully. Nov 12 20:56:36.585676 containerd[1444]: time="2024-11-12T20:56:36.585583267Z" level=info msg="shim disconnected" id=c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0 namespace=k8s.io Nov 12 20:56:36.585676 containerd[1444]: time="2024-11-12T20:56:36.585669859Z" level=warning msg="cleaning up after shim disconnected" id=c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0 namespace=k8s.io Nov 12 20:56:36.585676 containerd[1444]: time="2024-11-12T20:56:36.585683256Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:56:36.856356 containerd[1444]: time="2024-11-12T20:56:36.856292693Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:56:36.857056 containerd[1444]: time="2024-11-12T20:56:36.856993783Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18907193" Nov 12 20:56:36.858574 containerd[1444]: time="2024-11-12T20:56:36.858531093Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Nov 12 20:56:36.859970 containerd[1444]: time="2024-11-12T20:56:36.859926499Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.393491856s" Nov 12 20:56:36.860031 containerd[1444]: time="2024-11-12T20:56:36.859971048Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Nov 12 20:56:36.878661 containerd[1444]: time="2024-11-12T20:56:36.878606127Z" level=info msg="CreateContainer within sandbox \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Nov 12 20:56:36.894864 containerd[1444]: time="2024-11-12T20:56:36.894808870Z" level=info msg="CreateContainer within sandbox \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\"" Nov 12 20:56:36.898227 containerd[1444]: time="2024-11-12T20:56:36.898189590Z" level=info msg="StartContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\"" Nov 12 20:56:36.940205 systemd[1]: Started cri-containerd-a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f.scope - libcontainer container a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f. Nov 12 20:56:36.968456 containerd[1444]: time="2024-11-12T20:56:36.968384070Z" level=info msg="StartContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" returns successfully" Nov 12 20:56:37.425624 kubelet[2610]: E1112 20:56:37.425575 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:37.429105 kubelet[2610]: E1112 20:56:37.428851 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:37.434518 containerd[1444]: time="2024-11-12T20:56:37.434460578Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Nov 12 20:56:37.733300 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3913851462.mount: Deactivated successfully. Nov 12 20:56:37.843789 containerd[1444]: time="2024-11-12T20:56:37.843730741Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\"" Nov 12 20:56:37.844515 containerd[1444]: time="2024-11-12T20:56:37.844461830Z" level=info msg="StartContainer for \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\"" Nov 12 20:56:37.851401 kubelet[2610]: I1112 20:56:37.850519 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-82wgj" podStartSLOduration=3.661876276 podStartE2EDuration="23.850492974s" podCreationTimestamp="2024-11-12 20:56:14 +0000 UTC" firstStartedPulling="2024-11-12 20:56:16.677217437 +0000 UTC m=+14.752894241" lastFinishedPulling="2024-11-12 20:56:36.865834135 +0000 UTC m=+34.941510939" observedRunningTime="2024-11-12 20:56:37.699193169 +0000 UTC m=+35.774869973" watchObservedRunningTime="2024-11-12 20:56:37.850492974 +0000 UTC m=+35.926169778" Nov 12 20:56:37.910194 systemd[1]: Started cri-containerd-e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6.scope - libcontainer container e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6. Nov 12 20:56:37.941948 systemd[1]: cri-containerd-e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6.scope: Deactivated successfully. Nov 12 20:56:37.945206 containerd[1444]: time="2024-11-12T20:56:37.944962069Z" level=info msg="StartContainer for \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\" returns successfully" Nov 12 20:56:37.977031 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6-rootfs.mount: Deactivated successfully. Nov 12 20:56:38.117860 containerd[1444]: time="2024-11-12T20:56:38.117774016Z" level=info msg="shim disconnected" id=e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6 namespace=k8s.io Nov 12 20:56:38.117860 containerd[1444]: time="2024-11-12T20:56:38.117849976Z" level=warning msg="cleaning up after shim disconnected" id=e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6 namespace=k8s.io Nov 12 20:56:38.117860 containerd[1444]: time="2024-11-12T20:56:38.117859155Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:56:38.434815 kubelet[2610]: E1112 20:56:38.434674 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:38.434815 kubelet[2610]: E1112 20:56:38.434727 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:38.438606 containerd[1444]: time="2024-11-12T20:56:38.438555176Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Nov 12 20:56:38.482497 containerd[1444]: time="2024-11-12T20:56:38.482440059Z" level=info msg="CreateContainer within sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\"" Nov 12 20:56:38.483220 containerd[1444]: time="2024-11-12T20:56:38.483180156Z" level=info msg="StartContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\"" Nov 12 20:56:38.513062 systemd[1]: Started cri-containerd-4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de.scope - libcontainer container 4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de. Nov 12 20:56:38.548945 containerd[1444]: time="2024-11-12T20:56:38.548862547Z" level=info msg="StartContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" returns successfully" Nov 12 20:56:38.663059 systemd[1]: Started sshd@8-10.0.0.145:22-10.0.0.1:45516.service - OpenSSH per-connection server daemon (10.0.0.1:45516). Nov 12 20:56:38.690207 kubelet[2610]: I1112 20:56:38.690076 2610 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Nov 12 20:56:38.701179 sshd[3385]: Accepted publickey for core from 10.0.0.1 port 45516 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:38.702853 sshd[3385]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:38.707619 systemd-logind[1433]: New session 9 of user core. Nov 12 20:56:38.713108 systemd[1]: Started session-9.scope - Session 9 of User core. Nov 12 20:56:38.856786 kubelet[2610]: I1112 20:56:38.856716 2610 topology_manager.go:215] "Topology Admit Handler" podUID="bb6d2776-456c-428c-b68a-367f9310df6a" podNamespace="kube-system" podName="coredns-7db6d8ff4d-9h46f" Nov 12 20:56:38.866150 systemd[1]: Created slice kubepods-burstable-podbb6d2776_456c_428c_b68a_367f9310df6a.slice - libcontainer container kubepods-burstable-podbb6d2776_456c_428c_b68a_367f9310df6a.slice. Nov 12 20:56:39.017422 kubelet[2610]: I1112 20:56:39.017150 2610 topology_manager.go:215] "Topology Admit Handler" podUID="b29259a7-d909-44f7-8bb7-fea845270210" podNamespace="kube-system" podName="coredns-7db6d8ff4d-4djqm" Nov 12 20:56:39.024829 systemd[1]: Created slice kubepods-burstable-podb29259a7_d909_44f7_8bb7_fea845270210.slice - libcontainer container kubepods-burstable-podb29259a7_d909_44f7_8bb7_fea845270210.slice. Nov 12 20:56:39.027347 sshd[3385]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:39.031161 systemd[1]: sshd@8-10.0.0.145:22-10.0.0.1:45516.service: Deactivated successfully. Nov 12 20:56:39.033599 systemd[1]: session-9.scope: Deactivated successfully. Nov 12 20:56:39.035547 systemd-logind[1433]: Session 9 logged out. Waiting for processes to exit. Nov 12 20:56:39.036755 systemd-logind[1433]: Removed session 9. Nov 12 20:56:39.043569 kubelet[2610]: I1112 20:56:39.043532 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwxlj\" (UniqueName: \"kubernetes.io/projected/bb6d2776-456c-428c-b68a-367f9310df6a-kube-api-access-kwxlj\") pod \"coredns-7db6d8ff4d-9h46f\" (UID: \"bb6d2776-456c-428c-b68a-367f9310df6a\") " pod="kube-system/coredns-7db6d8ff4d-9h46f" Nov 12 20:56:39.043569 kubelet[2610]: I1112 20:56:39.043569 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb6d2776-456c-428c-b68a-367f9310df6a-config-volume\") pod \"coredns-7db6d8ff4d-9h46f\" (UID: \"bb6d2776-456c-428c-b68a-367f9310df6a\") " pod="kube-system/coredns-7db6d8ff4d-9h46f" Nov 12 20:56:39.144248 kubelet[2610]: I1112 20:56:39.144142 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9v6h\" (UniqueName: \"kubernetes.io/projected/b29259a7-d909-44f7-8bb7-fea845270210-kube-api-access-b9v6h\") pod \"coredns-7db6d8ff4d-4djqm\" (UID: \"b29259a7-d909-44f7-8bb7-fea845270210\") " pod="kube-system/coredns-7db6d8ff4d-4djqm" Nov 12 20:56:39.144248 kubelet[2610]: I1112 20:56:39.144198 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b29259a7-d909-44f7-8bb7-fea845270210-config-volume\") pod \"coredns-7db6d8ff4d-4djqm\" (UID: \"b29259a7-d909-44f7-8bb7-fea845270210\") " pod="kube-system/coredns-7db6d8ff4d-4djqm" Nov 12 20:56:39.440079 kubelet[2610]: E1112 20:56:39.440044 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:39.469997 kubelet[2610]: E1112 20:56:39.469955 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:39.474156 containerd[1444]: time="2024-11-12T20:56:39.474118398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-9h46f,Uid:bb6d2776-456c-428c-b68a-367f9310df6a,Namespace:kube-system,Attempt:0,}" Nov 12 20:56:39.628813 kubelet[2610]: E1112 20:56:39.628773 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:39.629498 containerd[1444]: time="2024-11-12T20:56:39.629457698Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-4djqm,Uid:b29259a7-d909-44f7-8bb7-fea845270210,Namespace:kube-system,Attempt:0,}" Nov 12 20:56:39.782154 kubelet[2610]: I1112 20:56:39.782004 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-wh5nj" podStartSLOduration=8.959108912 podStartE2EDuration="25.781969364s" podCreationTimestamp="2024-11-12 20:56:14 +0000 UTC" firstStartedPulling="2024-11-12 20:56:16.643103619 +0000 UTC m=+14.718780423" lastFinishedPulling="2024-11-12 20:56:33.465964071 +0000 UTC m=+31.541640875" observedRunningTime="2024-11-12 20:56:39.781838142 +0000 UTC m=+37.857514966" watchObservedRunningTime="2024-11-12 20:56:39.781969364 +0000 UTC m=+37.857646168" Nov 12 20:56:40.442899 kubelet[2610]: E1112 20:56:40.442845 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:40.899200 systemd-networkd[1387]: cilium_host: Link UP Nov 12 20:56:40.899408 systemd-networkd[1387]: cilium_net: Link UP Nov 12 20:56:40.900334 systemd-networkd[1387]: cilium_net: Gained carrier Nov 12 20:56:40.900809 systemd-networkd[1387]: cilium_host: Gained carrier Nov 12 20:56:40.901275 systemd-networkd[1387]: cilium_net: Gained IPv6LL Nov 12 20:56:40.902023 systemd-networkd[1387]: cilium_host: Gained IPv6LL Nov 12 20:56:41.005705 systemd-networkd[1387]: cilium_vxlan: Link UP Nov 12 20:56:41.005867 systemd-networkd[1387]: cilium_vxlan: Gained carrier Nov 12 20:56:41.226951 kernel: NET: Registered PF_ALG protocol family Nov 12 20:56:41.444206 kubelet[2610]: E1112 20:56:41.444178 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:41.952831 systemd-networkd[1387]: lxc_health: Link UP Nov 12 20:56:41.962094 systemd-networkd[1387]: lxc_health: Gained carrier Nov 12 20:56:42.516005 systemd-networkd[1387]: lxcfff5c87c346b: Link UP Nov 12 20:56:42.521954 kernel: eth0: renamed from tmpfb407 Nov 12 20:56:42.522198 systemd-networkd[1387]: lxc997af89ccab5: Link UP Nov 12 20:56:42.543375 systemd-networkd[1387]: lxcfff5c87c346b: Gained carrier Nov 12 20:56:42.544928 kernel: eth0: renamed from tmpdc59a Nov 12 20:56:42.553384 systemd-networkd[1387]: lxc997af89ccab5: Gained carrier Nov 12 20:56:43.029179 systemd-networkd[1387]: cilium_vxlan: Gained IPv6LL Nov 12 20:56:43.479055 systemd-networkd[1387]: lxc_health: Gained IPv6LL Nov 12 20:56:43.789024 kubelet[2610]: E1112 20:56:43.787463 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:44.038588 systemd[1]: Started sshd@9-10.0.0.145:22-10.0.0.1:45522.service - OpenSSH per-connection server daemon (10.0.0.1:45522). Nov 12 20:56:44.054093 systemd-networkd[1387]: lxc997af89ccab5: Gained IPv6LL Nov 12 20:56:44.083297 sshd[3848]: Accepted publickey for core from 10.0.0.1 port 45522 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:44.085121 sshd[3848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:44.089275 systemd-logind[1433]: New session 10 of user core. Nov 12 20:56:44.098059 systemd[1]: Started session-10.scope - Session 10 of User core. Nov 12 20:56:44.226965 sshd[3848]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:44.232028 systemd[1]: sshd@9-10.0.0.145:22-10.0.0.1:45522.service: Deactivated successfully. Nov 12 20:56:44.234474 systemd[1]: session-10.scope: Deactivated successfully. Nov 12 20:56:44.235198 systemd-logind[1433]: Session 10 logged out. Waiting for processes to exit. Nov 12 20:56:44.236139 systemd-logind[1433]: Removed session 10. Nov 12 20:56:44.449709 kubelet[2610]: E1112 20:56:44.449677 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:44.565067 systemd-networkd[1387]: lxcfff5c87c346b: Gained IPv6LL Nov 12 20:56:45.452093 kubelet[2610]: E1112 20:56:45.452053 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:46.198369 containerd[1444]: time="2024-11-12T20:56:46.197938191Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:56:46.198369 containerd[1444]: time="2024-11-12T20:56:46.198031980Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:56:46.198369 containerd[1444]: time="2024-11-12T20:56:46.198069848Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:46.198369 containerd[1444]: time="2024-11-12T20:56:46.198189565Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:46.227801 systemd[1]: Started cri-containerd-dc59a055bf0ad96b025fa70204ce61aa31f8bba4eaaacc44c2d4a49618e5ff9b.scope - libcontainer container dc59a055bf0ad96b025fa70204ce61aa31f8bba4eaaacc44c2d4a49618e5ff9b. Nov 12 20:56:46.240761 containerd[1444]: time="2024-11-12T20:56:46.240630387Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:56:46.240761 containerd[1444]: time="2024-11-12T20:56:46.240702827Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:56:46.240761 containerd[1444]: time="2024-11-12T20:56:46.240714670Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:46.240993 containerd[1444]: time="2024-11-12T20:56:46.240801206Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:56:46.245006 systemd-resolved[1345]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 12 20:56:46.263279 systemd[1]: Started cri-containerd-fb407a164d1e3b22d475a6f53982a32e3a8b93d01e68146df937cbfcb20139f1.scope - libcontainer container fb407a164d1e3b22d475a6f53982a32e3a8b93d01e68146df937cbfcb20139f1. Nov 12 20:56:46.277570 containerd[1444]: time="2024-11-12T20:56:46.277514333Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-9h46f,Uid:bb6d2776-456c-428c-b68a-367f9310df6a,Namespace:kube-system,Attempt:0,} returns sandbox id \"dc59a055bf0ad96b025fa70204ce61aa31f8bba4eaaacc44c2d4a49618e5ff9b\"" Nov 12 20:56:46.277769 systemd-resolved[1345]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Nov 12 20:56:46.278868 kubelet[2610]: E1112 20:56:46.278845 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:46.282560 containerd[1444]: time="2024-11-12T20:56:46.282525302Z" level=info msg="CreateContainer within sandbox \"dc59a055bf0ad96b025fa70204ce61aa31f8bba4eaaacc44c2d4a49618e5ff9b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 12 20:56:46.304473 containerd[1444]: time="2024-11-12T20:56:46.303833156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-4djqm,Uid:b29259a7-d909-44f7-8bb7-fea845270210,Namespace:kube-system,Attempt:0,} returns sandbox id \"fb407a164d1e3b22d475a6f53982a32e3a8b93d01e68146df937cbfcb20139f1\"" Nov 12 20:56:46.304746 kubelet[2610]: E1112 20:56:46.304711 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:46.306643 containerd[1444]: time="2024-11-12T20:56:46.306611630Z" level=info msg="CreateContainer within sandbox \"fb407a164d1e3b22d475a6f53982a32e3a8b93d01e68146df937cbfcb20139f1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Nov 12 20:56:47.377241 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1874891566.mount: Deactivated successfully. Nov 12 20:56:47.379731 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3380828341.mount: Deactivated successfully. Nov 12 20:56:47.451658 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount685265401.mount: Deactivated successfully. Nov 12 20:56:48.467178 containerd[1444]: time="2024-11-12T20:56:48.467122355Z" level=info msg="CreateContainer within sandbox \"fb407a164d1e3b22d475a6f53982a32e3a8b93d01e68146df937cbfcb20139f1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"94fe56adc18b7a2b53bb058142a1933f294ea0ddd4ba223fe4c70f103aac2e58\"" Nov 12 20:56:48.467812 containerd[1444]: time="2024-11-12T20:56:48.467787143Z" level=info msg="StartContainer for \"94fe56adc18b7a2b53bb058142a1933f294ea0ddd4ba223fe4c70f103aac2e58\"" Nov 12 20:56:48.500051 systemd[1]: Started cri-containerd-94fe56adc18b7a2b53bb058142a1933f294ea0ddd4ba223fe4c70f103aac2e58.scope - libcontainer container 94fe56adc18b7a2b53bb058142a1933f294ea0ddd4ba223fe4c70f103aac2e58. Nov 12 20:56:48.632555 containerd[1444]: time="2024-11-12T20:56:48.632483207Z" level=info msg="CreateContainer within sandbox \"dc59a055bf0ad96b025fa70204ce61aa31f8bba4eaaacc44c2d4a49618e5ff9b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4\"" Nov 12 20:56:48.633963 containerd[1444]: time="2024-11-12T20:56:48.633792286Z" level=info msg="StartContainer for \"21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4\"" Nov 12 20:56:48.668069 systemd[1]: Started cri-containerd-21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4.scope - libcontainer container 21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4. Nov 12 20:56:49.021843 containerd[1444]: time="2024-11-12T20:56:49.021762692Z" level=info msg="StartContainer for \"94fe56adc18b7a2b53bb058142a1933f294ea0ddd4ba223fe4c70f103aac2e58\" returns successfully" Nov 12 20:56:49.022053 containerd[1444]: time="2024-11-12T20:56:49.021786706Z" level=info msg="StartContainer for \"21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4\" returns successfully" Nov 12 20:56:49.240290 systemd[1]: Started sshd@10-10.0.0.145:22-10.0.0.1:39636.service - OpenSSH per-connection server daemon (10.0.0.1:39636). Nov 12 20:56:49.287329 sshd[4035]: Accepted publickey for core from 10.0.0.1 port 39636 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:49.289354 sshd[4035]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:49.294069 systemd-logind[1433]: New session 11 of user core. Nov 12 20:56:49.303196 systemd[1]: Started session-11.scope - Session 11 of User core. Nov 12 20:56:49.425552 sshd[4035]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:49.429709 systemd[1]: sshd@10-10.0.0.145:22-10.0.0.1:39636.service: Deactivated successfully. Nov 12 20:56:49.431495 systemd[1]: session-11.scope: Deactivated successfully. Nov 12 20:56:49.432118 systemd-logind[1433]: Session 11 logged out. Waiting for processes to exit. Nov 12 20:56:49.432968 systemd-logind[1433]: Removed session 11. Nov 12 20:56:49.463405 kubelet[2610]: E1112 20:56:49.463378 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:49.467581 kubelet[2610]: E1112 20:56:49.467380 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:49.474600 kubelet[2610]: I1112 20:56:49.474543 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-9h46f" podStartSLOduration=35.474526148 podStartE2EDuration="35.474526148s" podCreationTimestamp="2024-11-12 20:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:56:49.474054408 +0000 UTC m=+47.549731213" watchObservedRunningTime="2024-11-12 20:56:49.474526148 +0000 UTC m=+47.550202952" Nov 12 20:56:49.479650 systemd[1]: run-containerd-runc-k8s.io-21f782a06ed7f92537af4adf8708c31c83af765c8fa3bdd9860db7da226134b4-runc.IDU2av.mount: Deactivated successfully. Nov 12 20:56:50.469300 kubelet[2610]: E1112 20:56:50.469254 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:50.469300 kubelet[2610]: E1112 20:56:50.469272 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:51.470714 kubelet[2610]: E1112 20:56:51.470663 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:51.471189 kubelet[2610]: E1112 20:56:51.470754 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:56:54.439804 systemd[1]: Started sshd@11-10.0.0.145:22-10.0.0.1:39650.service - OpenSSH per-connection server daemon (10.0.0.1:39650). Nov 12 20:56:54.480297 sshd[4060]: Accepted publickey for core from 10.0.0.1 port 39650 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:54.482178 sshd[4060]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:54.486520 systemd-logind[1433]: New session 12 of user core. Nov 12 20:56:54.494062 systemd[1]: Started session-12.scope - Session 12 of User core. Nov 12 20:56:54.613567 sshd[4060]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:54.627995 systemd[1]: sshd@11-10.0.0.145:22-10.0.0.1:39650.service: Deactivated successfully. Nov 12 20:56:54.630735 systemd[1]: session-12.scope: Deactivated successfully. Nov 12 20:56:54.632881 systemd-logind[1433]: Session 12 logged out. Waiting for processes to exit. Nov 12 20:56:54.644454 systemd[1]: Started sshd@12-10.0.0.145:22-10.0.0.1:39654.service - OpenSSH per-connection server daemon (10.0.0.1:39654). Nov 12 20:56:54.646969 systemd-logind[1433]: Removed session 12. Nov 12 20:56:54.682429 sshd[4075]: Accepted publickey for core from 10.0.0.1 port 39654 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:54.684425 sshd[4075]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:54.689463 systemd-logind[1433]: New session 13 of user core. Nov 12 20:56:54.702125 systemd[1]: Started session-13.scope - Session 13 of User core. Nov 12 20:56:54.962022 sshd[4075]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:54.972449 systemd[1]: sshd@12-10.0.0.145:22-10.0.0.1:39654.service: Deactivated successfully. Nov 12 20:56:54.974355 systemd[1]: session-13.scope: Deactivated successfully. Nov 12 20:56:54.975815 systemd-logind[1433]: Session 13 logged out. Waiting for processes to exit. Nov 12 20:56:54.982205 systemd[1]: Started sshd@13-10.0.0.145:22-10.0.0.1:39658.service - OpenSSH per-connection server daemon (10.0.0.1:39658). Nov 12 20:56:54.983329 systemd-logind[1433]: Removed session 13. Nov 12 20:56:55.018479 sshd[4088]: Accepted publickey for core from 10.0.0.1 port 39658 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:56:55.020481 sshd[4088]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:56:55.024747 systemd-logind[1433]: New session 14 of user core. Nov 12 20:56:55.033079 systemd[1]: Started session-14.scope - Session 14 of User core. Nov 12 20:56:55.214862 sshd[4088]: pam_unix(sshd:session): session closed for user core Nov 12 20:56:55.219713 systemd[1]: sshd@13-10.0.0.145:22-10.0.0.1:39658.service: Deactivated successfully. Nov 12 20:56:55.222215 systemd[1]: session-14.scope: Deactivated successfully. Nov 12 20:56:55.223144 systemd-logind[1433]: Session 14 logged out. Waiting for processes to exit. Nov 12 20:56:55.224432 systemd-logind[1433]: Removed session 14. Nov 12 20:57:00.230153 systemd[1]: Started sshd@14-10.0.0.145:22-10.0.0.1:57356.service - OpenSSH per-connection server daemon (10.0.0.1:57356). Nov 12 20:57:00.268287 sshd[4102]: Accepted publickey for core from 10.0.0.1 port 57356 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:00.269881 sshd[4102]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:00.273826 systemd-logind[1433]: New session 15 of user core. Nov 12 20:57:00.285052 systemd[1]: Started session-15.scope - Session 15 of User core. Nov 12 20:57:00.401605 sshd[4102]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:00.406073 systemd[1]: sshd@14-10.0.0.145:22-10.0.0.1:57356.service: Deactivated successfully. Nov 12 20:57:00.408322 systemd[1]: session-15.scope: Deactivated successfully. Nov 12 20:57:00.408916 systemd-logind[1433]: Session 15 logged out. Waiting for processes to exit. Nov 12 20:57:00.409897 systemd-logind[1433]: Removed session 15. Nov 12 20:57:05.415850 systemd[1]: Started sshd@15-10.0.0.145:22-10.0.0.1:57372.service - OpenSSH per-connection server daemon (10.0.0.1:57372). Nov 12 20:57:05.454072 sshd[4118]: Accepted publickey for core from 10.0.0.1 port 57372 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:05.455629 sshd[4118]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:05.459531 systemd-logind[1433]: New session 16 of user core. Nov 12 20:57:05.469108 systemd[1]: Started session-16.scope - Session 16 of User core. Nov 12 20:57:05.575855 sshd[4118]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:05.579948 systemd[1]: sshd@15-10.0.0.145:22-10.0.0.1:57372.service: Deactivated successfully. Nov 12 20:57:05.582266 systemd[1]: session-16.scope: Deactivated successfully. Nov 12 20:57:05.582878 systemd-logind[1433]: Session 16 logged out. Waiting for processes to exit. Nov 12 20:57:05.583835 systemd-logind[1433]: Removed session 16. Nov 12 20:57:10.588730 systemd[1]: Started sshd@16-10.0.0.145:22-10.0.0.1:54344.service - OpenSSH per-connection server daemon (10.0.0.1:54344). Nov 12 20:57:10.628685 sshd[4132]: Accepted publickey for core from 10.0.0.1 port 54344 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:10.630483 sshd[4132]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:10.634496 systemd-logind[1433]: New session 17 of user core. Nov 12 20:57:10.645125 systemd[1]: Started session-17.scope - Session 17 of User core. Nov 12 20:57:10.766488 sshd[4132]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:10.783198 systemd[1]: sshd@16-10.0.0.145:22-10.0.0.1:54344.service: Deactivated successfully. Nov 12 20:57:10.785231 systemd[1]: session-17.scope: Deactivated successfully. Nov 12 20:57:10.787316 systemd-logind[1433]: Session 17 logged out. Waiting for processes to exit. Nov 12 20:57:10.796243 systemd[1]: Started sshd@17-10.0.0.145:22-10.0.0.1:54346.service - OpenSSH per-connection server daemon (10.0.0.1:54346). Nov 12 20:57:10.797060 systemd-logind[1433]: Removed session 17. Nov 12 20:57:10.830528 sshd[4146]: Accepted publickey for core from 10.0.0.1 port 54346 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:10.832275 sshd[4146]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:10.836120 systemd-logind[1433]: New session 18 of user core. Nov 12 20:57:10.847029 systemd[1]: Started session-18.scope - Session 18 of User core. Nov 12 20:57:11.161216 sshd[4146]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:11.172269 systemd[1]: sshd@17-10.0.0.145:22-10.0.0.1:54346.service: Deactivated successfully. Nov 12 20:57:11.174421 systemd[1]: session-18.scope: Deactivated successfully. Nov 12 20:57:11.176218 systemd-logind[1433]: Session 18 logged out. Waiting for processes to exit. Nov 12 20:57:11.180482 systemd[1]: Started sshd@18-10.0.0.145:22-10.0.0.1:54348.service - OpenSSH per-connection server daemon (10.0.0.1:54348). Nov 12 20:57:11.181512 systemd-logind[1433]: Removed session 18. Nov 12 20:57:11.221456 sshd[4158]: Accepted publickey for core from 10.0.0.1 port 54348 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:11.223106 sshd[4158]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:11.227451 systemd-logind[1433]: New session 19 of user core. Nov 12 20:57:11.239053 systemd[1]: Started session-19.scope - Session 19 of User core. Nov 12 20:57:12.692099 sshd[4158]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:12.701199 systemd[1]: sshd@18-10.0.0.145:22-10.0.0.1:54348.service: Deactivated successfully. Nov 12 20:57:12.703609 systemd[1]: session-19.scope: Deactivated successfully. Nov 12 20:57:12.706132 systemd-logind[1433]: Session 19 logged out. Waiting for processes to exit. Nov 12 20:57:12.713314 systemd[1]: Started sshd@19-10.0.0.145:22-10.0.0.1:54362.service - OpenSSH per-connection server daemon (10.0.0.1:54362). Nov 12 20:57:12.714680 systemd-logind[1433]: Removed session 19. Nov 12 20:57:12.748034 sshd[4177]: Accepted publickey for core from 10.0.0.1 port 54362 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:12.749972 sshd[4177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:12.754101 systemd-logind[1433]: New session 20 of user core. Nov 12 20:57:12.764028 systemd[1]: Started session-20.scope - Session 20 of User core. Nov 12 20:57:12.989285 sshd[4177]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:12.999291 systemd[1]: sshd@19-10.0.0.145:22-10.0.0.1:54362.service: Deactivated successfully. Nov 12 20:57:13.001701 systemd[1]: session-20.scope: Deactivated successfully. Nov 12 20:57:13.003373 systemd-logind[1433]: Session 20 logged out. Waiting for processes to exit. Nov 12 20:57:13.011455 systemd[1]: Started sshd@20-10.0.0.145:22-10.0.0.1:54378.service - OpenSSH per-connection server daemon (10.0.0.1:54378). Nov 12 20:57:13.012692 systemd-logind[1433]: Removed session 20. Nov 12 20:57:13.044790 sshd[4190]: Accepted publickey for core from 10.0.0.1 port 54378 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:13.046385 sshd[4190]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:13.050309 systemd-logind[1433]: New session 21 of user core. Nov 12 20:57:13.059135 systemd[1]: Started session-21.scope - Session 21 of User core. Nov 12 20:57:13.172476 sshd[4190]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:13.176684 systemd[1]: sshd@20-10.0.0.145:22-10.0.0.1:54378.service: Deactivated successfully. Nov 12 20:57:13.179278 systemd[1]: session-21.scope: Deactivated successfully. Nov 12 20:57:13.180062 systemd-logind[1433]: Session 21 logged out. Waiting for processes to exit. Nov 12 20:57:13.181121 systemd-logind[1433]: Removed session 21. Nov 12 20:57:15.031827 kubelet[2610]: E1112 20:57:15.031788 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:18.185176 systemd[1]: Started sshd@21-10.0.0.145:22-10.0.0.1:33540.service - OpenSSH per-connection server daemon (10.0.0.1:33540). Nov 12 20:57:18.223358 sshd[4207]: Accepted publickey for core from 10.0.0.1 port 33540 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:18.224887 sshd[4207]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:18.229359 systemd-logind[1433]: New session 22 of user core. Nov 12 20:57:18.239067 systemd[1]: Started session-22.scope - Session 22 of User core. Nov 12 20:57:18.358144 sshd[4207]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:18.363417 systemd[1]: sshd@21-10.0.0.145:22-10.0.0.1:33540.service: Deactivated successfully. Nov 12 20:57:18.366223 systemd[1]: session-22.scope: Deactivated successfully. Nov 12 20:57:18.367074 systemd-logind[1433]: Session 22 logged out. Waiting for processes to exit. Nov 12 20:57:18.368003 systemd-logind[1433]: Removed session 22. Nov 12 20:57:20.031630 kubelet[2610]: E1112 20:57:20.031580 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:23.369481 systemd[1]: Started sshd@22-10.0.0.145:22-10.0.0.1:33544.service - OpenSSH per-connection server daemon (10.0.0.1:33544). Nov 12 20:57:23.406722 sshd[4224]: Accepted publickey for core from 10.0.0.1 port 33544 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:23.408178 sshd[4224]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:23.412209 systemd-logind[1433]: New session 23 of user core. Nov 12 20:57:23.428115 systemd[1]: Started session-23.scope - Session 23 of User core. Nov 12 20:57:23.533379 sshd[4224]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:23.537133 systemd[1]: sshd@22-10.0.0.145:22-10.0.0.1:33544.service: Deactivated successfully. Nov 12 20:57:23.539210 systemd[1]: session-23.scope: Deactivated successfully. Nov 12 20:57:23.539820 systemd-logind[1433]: Session 23 logged out. Waiting for processes to exit. Nov 12 20:57:23.540689 systemd-logind[1433]: Removed session 23. Nov 12 20:57:25.031703 kubelet[2610]: E1112 20:57:25.031662 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:28.032128 kubelet[2610]: E1112 20:57:28.032077 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:28.545982 systemd[1]: Started sshd@23-10.0.0.145:22-10.0.0.1:49562.service - OpenSSH per-connection server daemon (10.0.0.1:49562). Nov 12 20:57:28.588985 sshd[4238]: Accepted publickey for core from 10.0.0.1 port 49562 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:28.591070 sshd[4238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:28.595172 systemd-logind[1433]: New session 24 of user core. Nov 12 20:57:28.611164 systemd[1]: Started session-24.scope - Session 24 of User core. Nov 12 20:57:28.726059 sshd[4238]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:28.730256 systemd[1]: sshd@23-10.0.0.145:22-10.0.0.1:49562.service: Deactivated successfully. Nov 12 20:57:28.732236 systemd[1]: session-24.scope: Deactivated successfully. Nov 12 20:57:28.732920 systemd-logind[1433]: Session 24 logged out. Waiting for processes to exit. Nov 12 20:57:28.733902 systemd-logind[1433]: Removed session 24. Nov 12 20:57:33.738768 systemd[1]: Started sshd@24-10.0.0.145:22-10.0.0.1:49644.service - OpenSSH per-connection server daemon (10.0.0.1:49644). Nov 12 20:57:33.777766 sshd[4252]: Accepted publickey for core from 10.0.0.1 port 49644 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:33.779465 sshd[4252]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:33.783806 systemd-logind[1433]: New session 25 of user core. Nov 12 20:57:33.793107 systemd[1]: Started session-25.scope - Session 25 of User core. Nov 12 20:57:33.909708 sshd[4252]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:33.918600 systemd[1]: sshd@24-10.0.0.145:22-10.0.0.1:49644.service: Deactivated successfully. Nov 12 20:57:33.920337 systemd[1]: session-25.scope: Deactivated successfully. Nov 12 20:57:33.921993 systemd-logind[1433]: Session 25 logged out. Waiting for processes to exit. Nov 12 20:57:33.931387 systemd[1]: Started sshd@25-10.0.0.145:22-10.0.0.1:49658.service - OpenSSH per-connection server daemon (10.0.0.1:49658). Nov 12 20:57:33.932529 systemd-logind[1433]: Removed session 25. Nov 12 20:57:33.969140 sshd[4266]: Accepted publickey for core from 10.0.0.1 port 49658 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:33.970968 sshd[4266]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:33.975267 systemd-logind[1433]: New session 26 of user core. Nov 12 20:57:33.982054 systemd[1]: Started session-26.scope - Session 26 of User core. Nov 12 20:57:35.693874 kubelet[2610]: I1112 20:57:35.693803 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-4djqm" podStartSLOduration=80.693778671 podStartE2EDuration="1m20.693778671s" podCreationTimestamp="2024-11-12 20:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:56:49.495352462 +0000 UTC m=+47.571029286" watchObservedRunningTime="2024-11-12 20:57:35.693778671 +0000 UTC m=+93.769455475" Nov 12 20:57:35.745358 containerd[1444]: time="2024-11-12T20:57:35.745106978Z" level=info msg="StopContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" with timeout 30 (s)" Nov 12 20:57:35.746766 containerd[1444]: time="2024-11-12T20:57:35.745554358Z" level=info msg="Stop container \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" with signal terminated" Nov 12 20:57:35.753615 containerd[1444]: time="2024-11-12T20:57:35.753552603Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Nov 12 20:57:35.756559 containerd[1444]: time="2024-11-12T20:57:35.756511276Z" level=info msg="StopContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" with timeout 2 (s)" Nov 12 20:57:35.756792 containerd[1444]: time="2024-11-12T20:57:35.756763491Z" level=info msg="Stop container \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" with signal terminated" Nov 12 20:57:35.759354 systemd[1]: cri-containerd-a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f.scope: Deactivated successfully. Nov 12 20:57:35.765151 systemd-networkd[1387]: lxc_health: Link DOWN Nov 12 20:57:35.765162 systemd-networkd[1387]: lxc_health: Lost carrier Nov 12 20:57:35.787267 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f-rootfs.mount: Deactivated successfully. Nov 12 20:57:35.793979 systemd[1]: cri-containerd-4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de.scope: Deactivated successfully. Nov 12 20:57:35.794334 systemd[1]: cri-containerd-4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de.scope: Consumed 7.347s CPU time. Nov 12 20:57:35.796735 containerd[1444]: time="2024-11-12T20:57:35.796638935Z" level=info msg="shim disconnected" id=a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f namespace=k8s.io Nov 12 20:57:35.796735 containerd[1444]: time="2024-11-12T20:57:35.796698370Z" level=warning msg="cleaning up after shim disconnected" id=a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f namespace=k8s.io Nov 12 20:57:35.796735 containerd[1444]: time="2024-11-12T20:57:35.796707278Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:35.812876 containerd[1444]: time="2024-11-12T20:57:35.812733110Z" level=warning msg="cleanup warnings time=\"2024-11-12T20:57:35Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Nov 12 20:57:35.818601 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de-rootfs.mount: Deactivated successfully. Nov 12 20:57:35.821881 containerd[1444]: time="2024-11-12T20:57:35.821816784Z" level=info msg="StopContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" returns successfully" Nov 12 20:57:35.822724 containerd[1444]: time="2024-11-12T20:57:35.822690084Z" level=info msg="StopPodSandbox for \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\"" Nov 12 20:57:35.822868 containerd[1444]: time="2024-11-12T20:57:35.822750841Z" level=info msg="Container to stop \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.825690 containerd[1444]: time="2024-11-12T20:57:35.825624351Z" level=info msg="shim disconnected" id=4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de namespace=k8s.io Nov 12 20:57:35.825690 containerd[1444]: time="2024-11-12T20:57:35.825680600Z" level=warning msg="cleaning up after shim disconnected" id=4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de namespace=k8s.io Nov 12 20:57:35.825690 containerd[1444]: time="2024-11-12T20:57:35.825693474Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:35.825796 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb-shm.mount: Deactivated successfully. Nov 12 20:57:35.830729 systemd[1]: cri-containerd-6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb.scope: Deactivated successfully. Nov 12 20:57:35.846563 containerd[1444]: time="2024-11-12T20:57:35.846488616Z" level=info msg="StopContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" returns successfully" Nov 12 20:57:35.847327 containerd[1444]: time="2024-11-12T20:57:35.847294407Z" level=info msg="StopPodSandbox for \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\"" Nov 12 20:57:35.847384 containerd[1444]: time="2024-11-12T20:57:35.847329766Z" level=info msg="Container to stop \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.847384 containerd[1444]: time="2024-11-12T20:57:35.847344584Z" level=info msg="Container to stop \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.847384 containerd[1444]: time="2024-11-12T20:57:35.847355465Z" level=info msg="Container to stop \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.847384 containerd[1444]: time="2024-11-12T20:57:35.847366045Z" level=info msg="Container to stop \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.847384 containerd[1444]: time="2024-11-12T20:57:35.847377597Z" level=info msg="Container to stop \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Nov 12 20:57:35.849734 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4-shm.mount: Deactivated successfully. Nov 12 20:57:35.855053 systemd[1]: cri-containerd-d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4.scope: Deactivated successfully. Nov 12 20:57:35.876727 containerd[1444]: time="2024-11-12T20:57:35.876316553Z" level=info msg="shim disconnected" id=6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb namespace=k8s.io Nov 12 20:57:35.876727 containerd[1444]: time="2024-11-12T20:57:35.876452155Z" level=warning msg="cleaning up after shim disconnected" id=6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb namespace=k8s.io Nov 12 20:57:35.876727 containerd[1444]: time="2024-11-12T20:57:35.876471071Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:35.885198 containerd[1444]: time="2024-11-12T20:57:35.885110317Z" level=info msg="shim disconnected" id=d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4 namespace=k8s.io Nov 12 20:57:35.885566 containerd[1444]: time="2024-11-12T20:57:35.885286557Z" level=warning msg="cleaning up after shim disconnected" id=d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4 namespace=k8s.io Nov 12 20:57:35.885566 containerd[1444]: time="2024-11-12T20:57:35.885298369Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:35.903481 containerd[1444]: time="2024-11-12T20:57:35.903428022Z" level=info msg="TearDown network for sandbox \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" successfully" Nov 12 20:57:35.903481 containerd[1444]: time="2024-11-12T20:57:35.903471085Z" level=info msg="StopPodSandbox for \"d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4\" returns successfully" Nov 12 20:57:35.905159 containerd[1444]: time="2024-11-12T20:57:35.905122924Z" level=info msg="TearDown network for sandbox \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\" successfully" Nov 12 20:57:35.905159 containerd[1444]: time="2024-11-12T20:57:35.905152711Z" level=info msg="StopPodSandbox for \"6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb\" returns successfully" Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.062934 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-net\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.063005 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/224c5d84-0aec-401b-bfd3-b3f282e69424-clustermesh-secrets\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.063028 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-hostproc\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.063053 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wmld\" (UniqueName: \"kubernetes.io/projected/69dc6592-b201-4942-a38d-29a1e86e8cc9-kube-api-access-6wmld\") pod \"69dc6592-b201-4942-a38d-29a1e86e8cc9\" (UID: \"69dc6592-b201-4942-a38d-29a1e86e8cc9\") " Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.063077 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-hubble-tls\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.063985 kubelet[2610]: I1112 20:57:36.063108 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-etc-cni-netd\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064488 kubelet[2610]: I1112 20:57:36.063125 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-cgroup\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064488 kubelet[2610]: I1112 20:57:36.063144 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-run\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064488 kubelet[2610]: I1112 20:57:36.063133 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.064488 kubelet[2610]: I1112 20:57:36.063188 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.064488 kubelet[2610]: I1112 20:57:36.063161 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-xtables-lock\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063217 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-hostproc" (OuterVolumeSpecName: "hostproc") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063266 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69dc6592-b201-4942-a38d-29a1e86e8cc9-cilium-config-path\") pod \"69dc6592-b201-4942-a38d-29a1e86e8cc9\" (UID: \"69dc6592-b201-4942-a38d-29a1e86e8cc9\") " Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063288 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cni-path\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063309 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-kernel\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063332 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dchjp\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-kube-api-access-dchjp\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.064826 kubelet[2610]: I1112 20:57:36.063348 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-lib-modules\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063365 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-bpf-maps\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063382 2610 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-config-path\") pod \"224c5d84-0aec-401b-bfd3-b3f282e69424\" (UID: \"224c5d84-0aec-401b-bfd3-b3f282e69424\") " Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063440 2610 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063458 2610 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-hostproc\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063468 2610 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-xtables-lock\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.065012 kubelet[2610]: I1112 20:57:36.063655 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.066962 kubelet[2610]: I1112 20:57:36.066925 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 12 20:57:36.067282 kubelet[2610]: I1112 20:57:36.067112 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.067282 kubelet[2610]: I1112 20:57:36.067167 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.067282 kubelet[2610]: I1112 20:57:36.067189 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cni-path" (OuterVolumeSpecName: "cni-path") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.067282 kubelet[2610]: I1112 20:57:36.067210 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.067282 kubelet[2610]: I1112 20:57:36.067232 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.067653 kubelet[2610]: I1112 20:57:36.067253 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 12 20:57:36.068681 kubelet[2610]: I1112 20:57:36.068615 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69dc6592-b201-4942-a38d-29a1e86e8cc9-kube-api-access-6wmld" (OuterVolumeSpecName: "kube-api-access-6wmld") pod "69dc6592-b201-4942-a38d-29a1e86e8cc9" (UID: "69dc6592-b201-4942-a38d-29a1e86e8cc9"). InnerVolumeSpecName "kube-api-access-6wmld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 12 20:57:36.069659 kubelet[2610]: I1112 20:57:36.069619 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 12 20:57:36.070110 kubelet[2610]: I1112 20:57:36.070056 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/224c5d84-0aec-401b-bfd3-b3f282e69424-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 12 20:57:36.070426 kubelet[2610]: I1112 20:57:36.070390 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69dc6592-b201-4942-a38d-29a1e86e8cc9-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "69dc6592-b201-4942-a38d-29a1e86e8cc9" (UID: "69dc6592-b201-4942-a38d-29a1e86e8cc9"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 12 20:57:36.071113 kubelet[2610]: I1112 20:57:36.071068 2610 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-kube-api-access-dchjp" (OuterVolumeSpecName: "kube-api-access-dchjp") pod "224c5d84-0aec-401b-bfd3-b3f282e69424" (UID: "224c5d84-0aec-401b-bfd3-b3f282e69424"). InnerVolumeSpecName "kube-api-access-dchjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164396 2610 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cni-path\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164440 2610 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69dc6592-b201-4942-a38d-29a1e86e8cc9-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164451 2610 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164461 2610 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-dchjp\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-kube-api-access-dchjp\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164472 2610 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-lib-modules\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164460 kubelet[2610]: I1112 20:57:36.164481 2610 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-bpf-maps\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164491 2610 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164501 2610 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/224c5d84-0aec-401b-bfd3-b3f282e69424-hubble-tls\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164510 2610 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/224c5d84-0aec-401b-bfd3-b3f282e69424-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164520 2610 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-6wmld\" (UniqueName: \"kubernetes.io/projected/69dc6592-b201-4942-a38d-29a1e86e8cc9-kube-api-access-6wmld\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164530 2610 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164539 2610 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.164771 kubelet[2610]: I1112 20:57:36.164548 2610 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/224c5d84-0aec-401b-bfd3-b3f282e69424-cilium-run\") on node \"localhost\" DevicePath \"\"" Nov 12 20:57:36.559863 kubelet[2610]: I1112 20:57:36.559816 2610 scope.go:117] "RemoveContainer" containerID="4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de" Nov 12 20:57:36.562317 containerd[1444]: time="2024-11-12T20:57:36.561961210Z" level=info msg="RemoveContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\"" Nov 12 20:57:36.566954 containerd[1444]: time="2024-11-12T20:57:36.566667342Z" level=info msg="RemoveContainer for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" returns successfully" Nov 12 20:57:36.567068 systemd[1]: Removed slice kubepods-burstable-pod224c5d84_0aec_401b_bfd3_b3f282e69424.slice - libcontainer container kubepods-burstable-pod224c5d84_0aec_401b_bfd3_b3f282e69424.slice. Nov 12 20:57:36.567533 kubelet[2610]: I1112 20:57:36.567211 2610 scope.go:117] "RemoveContainer" containerID="e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6" Nov 12 20:57:36.567381 systemd[1]: kubepods-burstable-pod224c5d84_0aec_401b_bfd3_b3f282e69424.slice: Consumed 7.461s CPU time. Nov 12 20:57:36.570422 containerd[1444]: time="2024-11-12T20:57:36.569316884Z" level=info msg="RemoveContainer for \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\"" Nov 12 20:57:36.571800 systemd[1]: Removed slice kubepods-besteffort-pod69dc6592_b201_4942_a38d_29a1e86e8cc9.slice - libcontainer container kubepods-besteffort-pod69dc6592_b201_4942_a38d_29a1e86e8cc9.slice. Nov 12 20:57:36.573947 containerd[1444]: time="2024-11-12T20:57:36.573842688Z" level=info msg="RemoveContainer for \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\" returns successfully" Nov 12 20:57:36.574151 kubelet[2610]: I1112 20:57:36.574119 2610 scope.go:117] "RemoveContainer" containerID="c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0" Nov 12 20:57:36.575515 containerd[1444]: time="2024-11-12T20:57:36.575465454Z" level=info msg="RemoveContainer for \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\"" Nov 12 20:57:36.580202 containerd[1444]: time="2024-11-12T20:57:36.580144012Z" level=info msg="RemoveContainer for \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\" returns successfully" Nov 12 20:57:36.580650 kubelet[2610]: I1112 20:57:36.580605 2610 scope.go:117] "RemoveContainer" containerID="acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180" Nov 12 20:57:36.582669 containerd[1444]: time="2024-11-12T20:57:36.582603629Z" level=info msg="RemoveContainer for \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\"" Nov 12 20:57:36.587358 containerd[1444]: time="2024-11-12T20:57:36.587307316Z" level=info msg="RemoveContainer for \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\" returns successfully" Nov 12 20:57:36.587610 kubelet[2610]: I1112 20:57:36.587578 2610 scope.go:117] "RemoveContainer" containerID="a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98" Nov 12 20:57:36.589533 containerd[1444]: time="2024-11-12T20:57:36.589494707Z" level=info msg="RemoveContainer for \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\"" Nov 12 20:57:36.593904 containerd[1444]: time="2024-11-12T20:57:36.593846607Z" level=info msg="RemoveContainer for \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\" returns successfully" Nov 12 20:57:36.594424 kubelet[2610]: I1112 20:57:36.594382 2610 scope.go:117] "RemoveContainer" containerID="4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de" Nov 12 20:57:36.598309 containerd[1444]: time="2024-11-12T20:57:36.598230679Z" level=error msg="ContainerStatus for \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\": not found" Nov 12 20:57:36.598497 kubelet[2610]: E1112 20:57:36.598459 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\": not found" containerID="4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de" Nov 12 20:57:36.598650 kubelet[2610]: I1112 20:57:36.598495 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de"} err="failed to get container status \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\": rpc error: code = NotFound desc = an error occurred when try to find container \"4f771995fa8e6b304e8221c684d031adbae05191b3f58f3ebb28985d03a1a7de\": not found" Nov 12 20:57:36.598650 kubelet[2610]: I1112 20:57:36.598582 2610 scope.go:117] "RemoveContainer" containerID="e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6" Nov 12 20:57:36.598986 containerd[1444]: time="2024-11-12T20:57:36.598904477Z" level=error msg="ContainerStatus for \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\": not found" Nov 12 20:57:36.599173 kubelet[2610]: E1112 20:57:36.599132 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\": not found" containerID="e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6" Nov 12 20:57:36.599212 kubelet[2610]: I1112 20:57:36.599177 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6"} err="failed to get container status \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\": rpc error: code = NotFound desc = an error occurred when try to find container \"e938a0462fbb31906400bdf9f62c79978922486e9f0471c3ca96ab51cf6804c6\": not found" Nov 12 20:57:36.599244 kubelet[2610]: I1112 20:57:36.599210 2610 scope.go:117] "RemoveContainer" containerID="c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0" Nov 12 20:57:36.599491 containerd[1444]: time="2024-11-12T20:57:36.599451721Z" level=error msg="ContainerStatus for \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\": not found" Nov 12 20:57:36.599630 kubelet[2610]: E1112 20:57:36.599586 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\": not found" containerID="c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0" Nov 12 20:57:36.599630 kubelet[2610]: I1112 20:57:36.599611 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0"} err="failed to get container status \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\": rpc error: code = NotFound desc = an error occurred when try to find container \"c01830b0a5d2bcb3a76dad237d7461ce5c5721c3bdbee585b2e27ca5ff8cc1f0\": not found" Nov 12 20:57:36.599630 kubelet[2610]: I1112 20:57:36.599630 2610 scope.go:117] "RemoveContainer" containerID="acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180" Nov 12 20:57:36.599831 containerd[1444]: time="2024-11-12T20:57:36.599794511Z" level=error msg="ContainerStatus for \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\": not found" Nov 12 20:57:36.599959 kubelet[2610]: E1112 20:57:36.599936 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\": not found" containerID="acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180" Nov 12 20:57:36.600010 kubelet[2610]: I1112 20:57:36.599963 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180"} err="failed to get container status \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\": rpc error: code = NotFound desc = an error occurred when try to find container \"acc44f17a50ef785325e226e78967dc203c5426adf5acf33939262ef2b5b5180\": not found" Nov 12 20:57:36.600010 kubelet[2610]: I1112 20:57:36.599982 2610 scope.go:117] "RemoveContainer" containerID="a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98" Nov 12 20:57:36.600232 containerd[1444]: time="2024-11-12T20:57:36.600186165Z" level=error msg="ContainerStatus for \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\": not found" Nov 12 20:57:36.600363 kubelet[2610]: E1112 20:57:36.600339 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\": not found" containerID="a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98" Nov 12 20:57:36.600407 kubelet[2610]: I1112 20:57:36.600376 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98"} err="failed to get container status \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\": rpc error: code = NotFound desc = an error occurred when try to find container \"a1840cb5cbd788246fd275596d038e0f7046c51e4bc3c54987e7c5faea460e98\": not found" Nov 12 20:57:36.600407 kubelet[2610]: I1112 20:57:36.600389 2610 scope.go:117] "RemoveContainer" containerID="a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f" Nov 12 20:57:36.601491 containerd[1444]: time="2024-11-12T20:57:36.601458194Z" level=info msg="RemoveContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\"" Nov 12 20:57:36.604945 containerd[1444]: time="2024-11-12T20:57:36.604897557Z" level=info msg="RemoveContainer for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" returns successfully" Nov 12 20:57:36.605228 kubelet[2610]: I1112 20:57:36.605095 2610 scope.go:117] "RemoveContainer" containerID="a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f" Nov 12 20:57:36.605317 containerd[1444]: time="2024-11-12T20:57:36.605286876Z" level=error msg="ContainerStatus for \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\": not found" Nov 12 20:57:36.605563 kubelet[2610]: E1112 20:57:36.605534 2610 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\": not found" containerID="a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f" Nov 12 20:57:36.605626 kubelet[2610]: I1112 20:57:36.605566 2610 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f"} err="failed to get container status \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\": rpc error: code = NotFound desc = an error occurred when try to find container \"a86ee6cc16692cc9d87888a03450c5fc23e29740c734a94b60db9b8ce1b3066f\": not found" Nov 12 20:57:36.732284 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6822e32f059396bcbc8ac449adc53c2e339f415d79f6e01d4aaec29f77e212cb-rootfs.mount: Deactivated successfully. Nov 12 20:57:36.732448 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d5607228f7b75255cb51385687e068446f5f7b364402b1b00a443d584c29e4b4-rootfs.mount: Deactivated successfully. Nov 12 20:57:36.732557 systemd[1]: var-lib-kubelet-pods-69dc6592\x2db201\x2d4942\x2da38d\x2d29a1e86e8cc9-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6wmld.mount: Deactivated successfully. Nov 12 20:57:36.732790 systemd[1]: var-lib-kubelet-pods-224c5d84\x2d0aec\x2d401b\x2dbfd3\x2db3f282e69424-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ddchjp.mount: Deactivated successfully. Nov 12 20:57:36.732900 systemd[1]: var-lib-kubelet-pods-224c5d84\x2d0aec\x2d401b\x2dbfd3\x2db3f282e69424-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Nov 12 20:57:36.733022 systemd[1]: var-lib-kubelet-pods-224c5d84\x2d0aec\x2d401b\x2dbfd3\x2db3f282e69424-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Nov 12 20:57:37.093993 kubelet[2610]: E1112 20:57:37.093941 2610 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Nov 12 20:57:37.619827 sshd[4266]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:37.634437 systemd[1]: sshd@25-10.0.0.145:22-10.0.0.1:49658.service: Deactivated successfully. Nov 12 20:57:37.637407 systemd[1]: session-26.scope: Deactivated successfully. Nov 12 20:57:37.637669 systemd[1]: session-26.scope: Consumed 1.053s CPU time. Nov 12 20:57:37.640298 systemd-logind[1433]: Session 26 logged out. Waiting for processes to exit. Nov 12 20:57:37.648299 systemd[1]: Started sshd@26-10.0.0.145:22-10.0.0.1:54616.service - OpenSSH per-connection server daemon (10.0.0.1:54616). Nov 12 20:57:37.649703 systemd-logind[1433]: Removed session 26. Nov 12 20:57:37.689138 sshd[4429]: Accepted publickey for core from 10.0.0.1 port 54616 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:37.691212 sshd[4429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:37.696644 systemd-logind[1433]: New session 27 of user core. Nov 12 20:57:37.711213 systemd[1]: Started session-27.scope - Session 27 of User core. Nov 12 20:57:38.034325 kubelet[2610]: I1112 20:57:38.034195 2610 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" path="/var/lib/kubelet/pods/224c5d84-0aec-401b-bfd3-b3f282e69424/volumes" Nov 12 20:57:38.035263 kubelet[2610]: I1112 20:57:38.035226 2610 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69dc6592-b201-4942-a38d-29a1e86e8cc9" path="/var/lib/kubelet/pods/69dc6592-b201-4942-a38d-29a1e86e8cc9/volumes" Nov 12 20:57:38.270705 sshd[4429]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:38.285084 systemd[1]: sshd@26-10.0.0.145:22-10.0.0.1:54616.service: Deactivated successfully. Nov 12 20:57:38.286341 kubelet[2610]: I1112 20:57:38.285551 2610 topology_manager.go:215] "Topology Admit Handler" podUID="8a5ae42a-1570-4b84-bca3-baa54b8a83a3" podNamespace="kube-system" podName="cilium-6rgwm" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286635 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="apply-sysctl-overwrites" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286658 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="mount-bpf-fs" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286669 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69dc6592-b201-4942-a38d-29a1e86e8cc9" containerName="cilium-operator" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286677 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="cilium-agent" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286689 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="mount-cgroup" Nov 12 20:57:38.286694 kubelet[2610]: E1112 20:57:38.286697 2610 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="clean-cilium-state" Nov 12 20:57:38.286895 kubelet[2610]: I1112 20:57:38.286730 2610 memory_manager.go:354] "RemoveStaleState removing state" podUID="224c5d84-0aec-401b-bfd3-b3f282e69424" containerName="cilium-agent" Nov 12 20:57:38.286895 kubelet[2610]: I1112 20:57:38.286740 2610 memory_manager.go:354] "RemoveStaleState removing state" podUID="69dc6592-b201-4942-a38d-29a1e86e8cc9" containerName="cilium-operator" Nov 12 20:57:38.289652 systemd[1]: session-27.scope: Deactivated successfully. Nov 12 20:57:38.295078 systemd-logind[1433]: Session 27 logged out. Waiting for processes to exit. Nov 12 20:57:38.305591 systemd[1]: Started sshd@27-10.0.0.145:22-10.0.0.1:54632.service - OpenSSH per-connection server daemon (10.0.0.1:54632). Nov 12 20:57:38.311584 systemd-logind[1433]: Removed session 27. Nov 12 20:57:38.321027 systemd[1]: Created slice kubepods-burstable-pod8a5ae42a_1570_4b84_bca3_baa54b8a83a3.slice - libcontainer container kubepods-burstable-pod8a5ae42a_1570_4b84_bca3_baa54b8a83a3.slice. Nov 12 20:57:38.358503 sshd[4442]: Accepted publickey for core from 10.0.0.1 port 54632 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:38.360353 sshd[4442]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:38.365512 systemd-logind[1433]: New session 28 of user core. Nov 12 20:57:38.377108 systemd[1]: Started session-28.scope - Session 28 of User core. Nov 12 20:57:38.430208 sshd[4442]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:38.441469 systemd[1]: sshd@27-10.0.0.145:22-10.0.0.1:54632.service: Deactivated successfully. Nov 12 20:57:38.443593 systemd[1]: session-28.scope: Deactivated successfully. Nov 12 20:57:38.445714 systemd-logind[1433]: Session 28 logged out. Waiting for processes to exit. Nov 12 20:57:38.464537 systemd[1]: Started sshd@28-10.0.0.145:22-10.0.0.1:54640.service - OpenSSH per-connection server daemon (10.0.0.1:54640). Nov 12 20:57:38.465725 systemd-logind[1433]: Removed session 28. Nov 12 20:57:38.477537 kubelet[2610]: I1112 20:57:38.477473 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-cilium-run\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477537 kubelet[2610]: I1112 20:57:38.477531 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-cilium-cgroup\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477560 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-cni-path\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477581 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-cilium-config-path\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477605 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-host-proc-sys-kernel\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477662 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbnl\" (UniqueName: \"kubernetes.io/projected/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-kube-api-access-8xbnl\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477707 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-bpf-maps\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477769 kubelet[2610]: I1112 20:57:38.477729 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-etc-cni-netd\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477748 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-lib-modules\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477764 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-host-proc-sys-net\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477786 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-clustermesh-secrets\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477805 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-hubble-tls\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477830 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-cilium-ipsec-secrets\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.477978 kubelet[2610]: I1112 20:57:38.477850 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-hostproc\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.478168 kubelet[2610]: I1112 20:57:38.477870 2610 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8a5ae42a-1570-4b84-bca3-baa54b8a83a3-xtables-lock\") pod \"cilium-6rgwm\" (UID: \"8a5ae42a-1570-4b84-bca3-baa54b8a83a3\") " pod="kube-system/cilium-6rgwm" Nov 12 20:57:38.501766 sshd[4450]: Accepted publickey for core from 10.0.0.1 port 54640 ssh2: RSA SHA256:ff+1E3IxvymPzLNMRy6nd5oJGXfM6IAzu8KdPl3+w6U Nov 12 20:57:38.504018 sshd[4450]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Nov 12 20:57:38.508565 systemd-logind[1433]: New session 29 of user core. Nov 12 20:57:38.518334 systemd[1]: Started session-29.scope - Session 29 of User core. Nov 12 20:57:38.625712 kubelet[2610]: E1112 20:57:38.625660 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:38.627180 containerd[1444]: time="2024-11-12T20:57:38.626479585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6rgwm,Uid:8a5ae42a-1570-4b84-bca3-baa54b8a83a3,Namespace:kube-system,Attempt:0,}" Nov 12 20:57:38.660377 containerd[1444]: time="2024-11-12T20:57:38.660291415Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Nov 12 20:57:38.660752 containerd[1444]: time="2024-11-12T20:57:38.660725352Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Nov 12 20:57:38.660857 containerd[1444]: time="2024-11-12T20:57:38.660836296Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:57:38.661054 containerd[1444]: time="2024-11-12T20:57:38.661021813Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Nov 12 20:57:38.697153 systemd[1]: Started cri-containerd-2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f.scope - libcontainer container 2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f. Nov 12 20:57:38.721403 containerd[1444]: time="2024-11-12T20:57:38.721350693Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6rgwm,Uid:8a5ae42a-1570-4b84-bca3-baa54b8a83a3,Namespace:kube-system,Attempt:0,} returns sandbox id \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\"" Nov 12 20:57:38.722064 kubelet[2610]: E1112 20:57:38.722028 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:38.725145 containerd[1444]: time="2024-11-12T20:57:38.725055153Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Nov 12 20:57:39.430484 containerd[1444]: time="2024-11-12T20:57:39.430425825Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44\"" Nov 12 20:57:39.431173 containerd[1444]: time="2024-11-12T20:57:39.431121076Z" level=info msg="StartContainer for \"119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44\"" Nov 12 20:57:39.461146 systemd[1]: Started cri-containerd-119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44.scope - libcontainer container 119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44. Nov 12 20:57:39.528883 systemd[1]: cri-containerd-119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44.scope: Deactivated successfully. Nov 12 20:57:39.530709 containerd[1444]: time="2024-11-12T20:57:39.530648748Z" level=info msg="StartContainer for \"119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44\" returns successfully" Nov 12 20:57:39.573199 kubelet[2610]: E1112 20:57:39.573139 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:39.943699 containerd[1444]: time="2024-11-12T20:57:39.943581743Z" level=info msg="shim disconnected" id=119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44 namespace=k8s.io Nov 12 20:57:39.943699 containerd[1444]: time="2024-11-12T20:57:39.943660696Z" level=warning msg="cleaning up after shim disconnected" id=119681332b121adef76f5e1e5766262135b892d0039929cdb9bddd8e143f1a44 namespace=k8s.io Nov 12 20:57:39.943699 containerd[1444]: time="2024-11-12T20:57:39.943673000Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:40.576872 kubelet[2610]: E1112 20:57:40.576810 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:40.579230 containerd[1444]: time="2024-11-12T20:57:40.579166447Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Nov 12 20:57:40.967402 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3190654947.mount: Deactivated successfully. Nov 12 20:57:40.981325 containerd[1444]: time="2024-11-12T20:57:40.981254916Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1\"" Nov 12 20:57:40.982260 containerd[1444]: time="2024-11-12T20:57:40.982210019Z" level=info msg="StartContainer for \"72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1\"" Nov 12 20:57:41.016101 systemd[1]: Started cri-containerd-72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1.scope - libcontainer container 72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1. Nov 12 20:57:41.046595 containerd[1444]: time="2024-11-12T20:57:41.046527145Z" level=info msg="StartContainer for \"72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1\" returns successfully" Nov 12 20:57:41.053986 systemd[1]: cri-containerd-72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1.scope: Deactivated successfully. Nov 12 20:57:41.081632 containerd[1444]: time="2024-11-12T20:57:41.081554047Z" level=info msg="shim disconnected" id=72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1 namespace=k8s.io Nov 12 20:57:41.081632 containerd[1444]: time="2024-11-12T20:57:41.081628461Z" level=warning msg="cleaning up after shim disconnected" id=72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1 namespace=k8s.io Nov 12 20:57:41.081632 containerd[1444]: time="2024-11-12T20:57:41.081637940Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:41.580305 kubelet[2610]: E1112 20:57:41.580274 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:41.581962 containerd[1444]: time="2024-11-12T20:57:41.581900250Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Nov 12 20:57:41.601275 containerd[1444]: time="2024-11-12T20:57:41.601196230Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964\"" Nov 12 20:57:41.601928 containerd[1444]: time="2024-11-12T20:57:41.601829262Z" level=info msg="StartContainer for \"c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964\"" Nov 12 20:57:41.634061 systemd[1]: Started cri-containerd-c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964.scope - libcontainer container c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964. Nov 12 20:57:41.680426 systemd[1]: cri-containerd-c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964.scope: Deactivated successfully. Nov 12 20:57:41.681117 containerd[1444]: time="2024-11-12T20:57:41.681077442Z" level=info msg="StartContainer for \"c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964\" returns successfully" Nov 12 20:57:41.711411 containerd[1444]: time="2024-11-12T20:57:41.711335139Z" level=info msg="shim disconnected" id=c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964 namespace=k8s.io Nov 12 20:57:41.711411 containerd[1444]: time="2024-11-12T20:57:41.711399273Z" level=warning msg="cleaning up after shim disconnected" id=c1dcfb8e8a8872261e9f495807038f61e34d54f7208a854f739ddff00912f964 namespace=k8s.io Nov 12 20:57:41.711411 containerd[1444]: time="2024-11-12T20:57:41.711410785Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:41.961946 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-72c18507afede2a5494e8347752cbcd24cd5ce1082389924c0c7bb81729361f1-rootfs.mount: Deactivated successfully. Nov 12 20:57:42.096029 kubelet[2610]: E1112 20:57:42.095968 2610 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Nov 12 20:57:42.585168 kubelet[2610]: E1112 20:57:42.585096 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:42.587057 containerd[1444]: time="2024-11-12T20:57:42.587013716Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Nov 12 20:57:42.721558 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount81744167.mount: Deactivated successfully. Nov 12 20:57:42.727168 containerd[1444]: time="2024-11-12T20:57:42.727102862Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559\"" Nov 12 20:57:42.727792 containerd[1444]: time="2024-11-12T20:57:42.727763108Z" level=info msg="StartContainer for \"50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559\"" Nov 12 20:57:42.764252 systemd[1]: Started cri-containerd-50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559.scope - libcontainer container 50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559. Nov 12 20:57:42.795443 systemd[1]: cri-containerd-50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559.scope: Deactivated successfully. Nov 12 20:57:42.798373 containerd[1444]: time="2024-11-12T20:57:42.798319358Z" level=info msg="StartContainer for \"50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559\" returns successfully" Nov 12 20:57:42.904167 containerd[1444]: time="2024-11-12T20:57:42.904060916Z" level=info msg="shim disconnected" id=50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559 namespace=k8s.io Nov 12 20:57:42.904167 containerd[1444]: time="2024-11-12T20:57:42.904129347Z" level=warning msg="cleaning up after shim disconnected" id=50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559 namespace=k8s.io Nov 12 20:57:42.904167 containerd[1444]: time="2024-11-12T20:57:42.904139567Z" level=info msg="cleaning up dead shim" namespace=k8s.io Nov 12 20:57:42.961879 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-50ce0664c69a4b77252fea5f84608830b829b6ec14e4ebe5d59f1bb337049559-rootfs.mount: Deactivated successfully. Nov 12 20:57:43.592901 kubelet[2610]: E1112 20:57:43.592859 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:43.596087 containerd[1444]: time="2024-11-12T20:57:43.595856798Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Nov 12 20:57:43.628864 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2793684298.mount: Deactivated successfully. Nov 12 20:57:43.630404 containerd[1444]: time="2024-11-12T20:57:43.630336569Z" level=info msg="CreateContainer within sandbox \"2cb493674ced3a512b6997c0dfe12a9fda7c11cedc2ab49bef3d04fb1ae7db4f\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"0e717f8cb98440349a3d1b772742e5db368f5b62c636eab179020918b3b2c44a\"" Nov 12 20:57:43.631142 containerd[1444]: time="2024-11-12T20:57:43.631067211Z" level=info msg="StartContainer for \"0e717f8cb98440349a3d1b772742e5db368f5b62c636eab179020918b3b2c44a\"" Nov 12 20:57:43.672248 systemd[1]: Started cri-containerd-0e717f8cb98440349a3d1b772742e5db368f5b62c636eab179020918b3b2c44a.scope - libcontainer container 0e717f8cb98440349a3d1b772742e5db368f5b62c636eab179020918b3b2c44a. Nov 12 20:57:43.708844 containerd[1444]: time="2024-11-12T20:57:43.708788777Z" level=info msg="StartContainer for \"0e717f8cb98440349a3d1b772742e5db368f5b62c636eab179020918b3b2c44a\" returns successfully" Nov 12 20:57:44.179945 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Nov 12 20:57:44.598412 kubelet[2610]: E1112 20:57:44.598144 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:44.616017 kubelet[2610]: I1112 20:57:44.615845 2610 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-6rgwm" podStartSLOduration=6.615825118 podStartE2EDuration="6.615825118s" podCreationTimestamp="2024-11-12 20:57:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-11-12 20:57:44.61549015 +0000 UTC m=+102.691166964" watchObservedRunningTime="2024-11-12 20:57:44.615825118 +0000 UTC m=+102.691501922" Nov 12 20:57:45.425328 kubelet[2610]: I1112 20:57:45.425271 2610 setters.go:580] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-11-12T20:57:45Z","lastTransitionTime":"2024-11-12T20:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Nov 12 20:57:45.601101 kubelet[2610]: E1112 20:57:45.601045 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:47.479323 systemd-networkd[1387]: lxc_health: Link UP Nov 12 20:57:47.490016 systemd-networkd[1387]: lxc_health: Gained carrier Nov 12 20:57:48.628453 kubelet[2610]: E1112 20:57:48.628401 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:49.208021 systemd-networkd[1387]: lxc_health: Gained IPv6LL Nov 12 20:57:49.608632 kubelet[2610]: E1112 20:57:49.608586 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:50.609825 kubelet[2610]: E1112 20:57:50.609788 2610 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Nov 12 20:57:53.641858 sshd[4450]: pam_unix(sshd:session): session closed for user core Nov 12 20:57:53.646335 systemd[1]: sshd@28-10.0.0.145:22-10.0.0.1:54640.service: Deactivated successfully. Nov 12 20:57:53.648980 systemd[1]: session-29.scope: Deactivated successfully. Nov 12 20:57:53.649717 systemd-logind[1433]: Session 29 logged out. Waiting for processes to exit. Nov 12 20:57:53.650869 systemd-logind[1433]: Removed session 29.