Dec 13 08:47:23.085864 kernel: Linux version 6.6.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Dec 12 23:15:00 -00 2024 Dec 13 08:47:23.085907 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 08:47:23.085931 kernel: BIOS-provided physical RAM map: Dec 13 08:47:23.085947 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 13 08:47:23.085961 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 13 08:47:23.085976 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 13 08:47:23.085995 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Dec 13 08:47:23.086011 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Dec 13 08:47:23.086027 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 13 08:47:23.086047 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 13 08:47:23.086068 kernel: NX (Execute Disable) protection: active Dec 13 08:47:23.086084 kernel: APIC: Static calls initialized Dec 13 08:47:23.086100 kernel: SMBIOS 2.8 present. Dec 13 08:47:23.086117 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Dec 13 08:47:23.086138 kernel: Hypervisor detected: KVM Dec 13 08:47:23.086160 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 13 08:47:23.086181 kernel: kvm-clock: using sched offset of 4235813069 cycles Dec 13 08:47:23.086205 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 13 08:47:23.086234 kernel: tsc: Detected 2294.606 MHz processor Dec 13 08:47:23.086278 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 13 08:47:23.086292 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 13 08:47:23.086303 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Dec 13 08:47:23.086315 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Dec 13 08:47:23.086328 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 13 08:47:23.087563 kernel: ACPI: Early table checksum verification disabled Dec 13 08:47:23.087585 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Dec 13 08:47:23.087604 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087623 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087642 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087660 kernel: ACPI: FACS 0x000000007FFE0000 000040 Dec 13 08:47:23.087678 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087697 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087715 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087739 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 08:47:23.087757 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Dec 13 08:47:23.087775 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Dec 13 08:47:23.087794 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Dec 13 08:47:23.087812 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Dec 13 08:47:23.087831 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Dec 13 08:47:23.087850 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Dec 13 08:47:23.087886 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Dec 13 08:47:23.087906 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Dec 13 08:47:23.087925 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Dec 13 08:47:23.087946 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Dec 13 08:47:23.087965 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Dec 13 08:47:23.087986 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Dec 13 08:47:23.088006 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Dec 13 08:47:23.088030 kernel: Zone ranges: Dec 13 08:47:23.088050 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 13 08:47:23.088072 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Dec 13 08:47:23.088096 kernel: Normal empty Dec 13 08:47:23.088120 kernel: Movable zone start for each node Dec 13 08:47:23.088142 kernel: Early memory node ranges Dec 13 08:47:23.088166 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 13 08:47:23.088186 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Dec 13 08:47:23.088206 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Dec 13 08:47:23.088230 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 13 08:47:23.088254 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 13 08:47:23.088274 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Dec 13 08:47:23.088294 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 13 08:47:23.088314 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 13 08:47:23.088357 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 13 08:47:23.088378 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 13 08:47:23.088397 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 13 08:47:23.088417 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 13 08:47:23.088442 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 13 08:47:23.088461 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 13 08:47:23.088481 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 13 08:47:23.088500 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 13 08:47:23.088520 kernel: TSC deadline timer available Dec 13 08:47:23.088540 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Dec 13 08:47:23.088559 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 13 08:47:23.088579 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Dec 13 08:47:23.088602 kernel: Booting paravirtualized kernel on KVM Dec 13 08:47:23.088622 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 13 08:47:23.088649 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Dec 13 08:47:23.088670 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Dec 13 08:47:23.088693 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Dec 13 08:47:23.088712 kernel: pcpu-alloc: [0] 0 1 Dec 13 08:47:23.088735 kernel: kvm-guest: PV spinlocks disabled, no host support Dec 13 08:47:23.088758 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 08:47:23.088784 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 08:47:23.088813 kernel: random: crng init done Dec 13 08:47:23.088838 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 08:47:23.088864 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Dec 13 08:47:23.088890 kernel: Fallback order for Node 0: 0 Dec 13 08:47:23.088916 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Dec 13 08:47:23.088937 kernel: Policy zone: DMA32 Dec 13 08:47:23.088957 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 08:47:23.088978 kernel: Memory: 1971204K/2096612K available (12288K kernel code, 2299K rwdata, 22724K rodata, 42844K init, 2348K bss, 125148K reserved, 0K cma-reserved) Dec 13 08:47:23.088997 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 13 08:47:23.089022 kernel: Kernel/User page tables isolation: enabled Dec 13 08:47:23.089042 kernel: ftrace: allocating 37902 entries in 149 pages Dec 13 08:47:23.089063 kernel: ftrace: allocated 149 pages with 4 groups Dec 13 08:47:23.089082 kernel: Dynamic Preempt: voluntary Dec 13 08:47:23.089102 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 13 08:47:23.089123 kernel: rcu: RCU event tracing is enabled. Dec 13 08:47:23.089143 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 13 08:47:23.089163 kernel: Trampoline variant of Tasks RCU enabled. Dec 13 08:47:23.089183 kernel: Rude variant of Tasks RCU enabled. Dec 13 08:47:23.089206 kernel: Tracing variant of Tasks RCU enabled. Dec 13 08:47:23.089226 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 08:47:23.089246 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 13 08:47:23.089266 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Dec 13 08:47:23.089290 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 13 08:47:23.089310 kernel: Console: colour VGA+ 80x25 Dec 13 08:47:23.090217 kernel: printk: console [tty0] enabled Dec 13 08:47:23.090251 kernel: printk: console [ttyS0] enabled Dec 13 08:47:23.090271 kernel: ACPI: Core revision 20230628 Dec 13 08:47:23.090292 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 13 08:47:23.090320 kernel: APIC: Switch to symmetric I/O mode setup Dec 13 08:47:23.090363 kernel: x2apic enabled Dec 13 08:47:23.090383 kernel: APIC: Switched APIC routing to: physical x2apic Dec 13 08:47:23.090403 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 13 08:47:23.090423 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x21134dbeb26, max_idle_ns: 440795298546 ns Dec 13 08:47:23.090442 kernel: Calibrating delay loop (skipped) preset value.. 4589.21 BogoMIPS (lpj=2294606) Dec 13 08:47:23.090462 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Dec 13 08:47:23.090487 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Dec 13 08:47:23.090531 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 13 08:47:23.090556 kernel: Spectre V2 : Mitigation: Retpolines Dec 13 08:47:23.090578 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Dec 13 08:47:23.090603 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Dec 13 08:47:23.090624 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Dec 13 08:47:23.090645 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 13 08:47:23.090667 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 13 08:47:23.090688 kernel: MDS: Mitigation: Clear CPU buffers Dec 13 08:47:23.090709 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Dec 13 08:47:23.090742 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 13 08:47:23.090764 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 13 08:47:23.090785 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 13 08:47:23.090806 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 13 08:47:23.090827 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Dec 13 08:47:23.090848 kernel: Freeing SMP alternatives memory: 32K Dec 13 08:47:23.090869 kernel: pid_max: default: 32768 minimum: 301 Dec 13 08:47:23.090890 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Dec 13 08:47:23.090916 kernel: landlock: Up and running. Dec 13 08:47:23.090937 kernel: SELinux: Initializing. Dec 13 08:47:23.090958 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 13 08:47:23.090980 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 13 08:47:23.091001 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Dec 13 08:47:23.091022 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 08:47:23.091043 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 08:47:23.091064 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 13 08:47:23.091090 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Dec 13 08:47:23.091111 kernel: signal: max sigframe size: 1776 Dec 13 08:47:23.091132 kernel: rcu: Hierarchical SRCU implementation. Dec 13 08:47:23.091154 kernel: rcu: Max phase no-delay instances is 400. Dec 13 08:47:23.091176 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Dec 13 08:47:23.091197 kernel: smp: Bringing up secondary CPUs ... Dec 13 08:47:23.091218 kernel: smpboot: x86: Booting SMP configuration: Dec 13 08:47:23.091242 kernel: .... node #0, CPUs: #1 Dec 13 08:47:23.091263 kernel: smp: Brought up 1 node, 2 CPUs Dec 13 08:47:23.091284 kernel: smpboot: Max logical packages: 1 Dec 13 08:47:23.091309 kernel: smpboot: Total of 2 processors activated (9178.42 BogoMIPS) Dec 13 08:47:23.091331 kernel: devtmpfs: initialized Dec 13 08:47:23.091366 kernel: x86/mm: Memory block size: 128MB Dec 13 08:47:23.091392 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 08:47:23.091419 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 13 08:47:23.091445 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 08:47:23.091472 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 08:47:23.091499 kernel: audit: initializing netlink subsys (disabled) Dec 13 08:47:23.091526 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 08:47:23.091555 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 13 08:47:23.091577 kernel: audit: type=2000 audit(1734079641.779:1): state=initialized audit_enabled=0 res=1 Dec 13 08:47:23.091595 kernel: cpuidle: using governor menu Dec 13 08:47:23.091611 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 08:47:23.091625 kernel: dca service started, version 1.12.1 Dec 13 08:47:23.091642 kernel: PCI: Using configuration type 1 for base access Dec 13 08:47:23.091664 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 13 08:47:23.091685 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 08:47:23.091707 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 13 08:47:23.091737 kernel: ACPI: Added _OSI(Module Device) Dec 13 08:47:23.091761 kernel: ACPI: Added _OSI(Processor Device) Dec 13 08:47:23.091817 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 08:47:23.091842 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 08:47:23.091863 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 08:47:23.091887 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Dec 13 08:47:23.091911 kernel: ACPI: Interpreter enabled Dec 13 08:47:23.091934 kernel: ACPI: PM: (supports S0 S5) Dec 13 08:47:23.091960 kernel: ACPI: Using IOAPIC for interrupt routing Dec 13 08:47:23.091985 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 13 08:47:23.092007 kernel: PCI: Using E820 reservations for host bridge windows Dec 13 08:47:23.092028 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Dec 13 08:47:23.092052 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 08:47:23.092515 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Dec 13 08:47:23.092688 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Dec 13 08:47:23.092832 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Dec 13 08:47:23.092865 kernel: acpiphp: Slot [3] registered Dec 13 08:47:23.092887 kernel: acpiphp: Slot [4] registered Dec 13 08:47:23.092908 kernel: acpiphp: Slot [5] registered Dec 13 08:47:23.092929 kernel: acpiphp: Slot [6] registered Dec 13 08:47:23.092951 kernel: acpiphp: Slot [7] registered Dec 13 08:47:23.092972 kernel: acpiphp: Slot [8] registered Dec 13 08:47:23.092993 kernel: acpiphp: Slot [9] registered Dec 13 08:47:23.093014 kernel: acpiphp: Slot [10] registered Dec 13 08:47:23.093035 kernel: acpiphp: Slot [11] registered Dec 13 08:47:23.093061 kernel: acpiphp: Slot [12] registered Dec 13 08:47:23.093083 kernel: acpiphp: Slot [13] registered Dec 13 08:47:23.093109 kernel: acpiphp: Slot [14] registered Dec 13 08:47:23.093130 kernel: acpiphp: Slot [15] registered Dec 13 08:47:23.093151 kernel: acpiphp: Slot [16] registered Dec 13 08:47:23.093171 kernel: acpiphp: Slot [17] registered Dec 13 08:47:23.093192 kernel: acpiphp: Slot [18] registered Dec 13 08:47:23.093224 kernel: acpiphp: Slot [19] registered Dec 13 08:47:23.093239 kernel: acpiphp: Slot [20] registered Dec 13 08:47:23.093253 kernel: acpiphp: Slot [21] registered Dec 13 08:47:23.093275 kernel: acpiphp: Slot [22] registered Dec 13 08:47:23.093289 kernel: acpiphp: Slot [23] registered Dec 13 08:47:23.093302 kernel: acpiphp: Slot [24] registered Dec 13 08:47:23.093316 kernel: acpiphp: Slot [25] registered Dec 13 08:47:23.093331 kernel: acpiphp: Slot [26] registered Dec 13 08:47:23.093372 kernel: acpiphp: Slot [27] registered Dec 13 08:47:23.093390 kernel: acpiphp: Slot [28] registered Dec 13 08:47:23.093414 kernel: acpiphp: Slot [29] registered Dec 13 08:47:23.093429 kernel: acpiphp: Slot [30] registered Dec 13 08:47:23.093487 kernel: acpiphp: Slot [31] registered Dec 13 08:47:23.093514 kernel: PCI host bridge to bus 0000:00 Dec 13 08:47:23.093722 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 13 08:47:23.093856 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 13 08:47:23.093987 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 13 08:47:23.094116 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Dec 13 08:47:23.094244 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Dec 13 08:47:23.094391 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 08:47:23.094581 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Dec 13 08:47:23.094779 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Dec 13 08:47:23.094962 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Dec 13 08:47:23.095115 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Dec 13 08:47:23.095270 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Dec 13 08:47:23.095475 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Dec 13 08:47:23.095670 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Dec 13 08:47:23.095821 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Dec 13 08:47:23.096016 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Dec 13 08:47:23.096196 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Dec 13 08:47:23.097539 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Dec 13 08:47:23.097750 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Dec 13 08:47:23.097908 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Dec 13 08:47:23.098083 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Dec 13 08:47:23.098248 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Dec 13 08:47:23.098410 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Dec 13 08:47:23.098556 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Dec 13 08:47:23.098703 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Dec 13 08:47:23.098874 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 13 08:47:23.101370 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Dec 13 08:47:23.101663 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Dec 13 08:47:23.101843 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Dec 13 08:47:23.102001 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Dec 13 08:47:23.102176 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Dec 13 08:47:23.102359 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Dec 13 08:47:23.102566 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Dec 13 08:47:23.102720 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Dec 13 08:47:23.102886 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Dec 13 08:47:23.103045 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Dec 13 08:47:23.103217 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Dec 13 08:47:23.104534 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Dec 13 08:47:23.104731 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Dec 13 08:47:23.104890 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Dec 13 08:47:23.105037 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Dec 13 08:47:23.105181 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Dec 13 08:47:23.105352 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Dec 13 08:47:23.106682 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Dec 13 08:47:23.106841 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Dec 13 08:47:23.107002 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Dec 13 08:47:23.107173 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Dec 13 08:47:23.108378 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Dec 13 08:47:23.108591 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Dec 13 08:47:23.108619 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 13 08:47:23.108641 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 13 08:47:23.108663 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 13 08:47:23.108684 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 13 08:47:23.108713 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Dec 13 08:47:23.108735 kernel: iommu: Default domain type: Translated Dec 13 08:47:23.108757 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 13 08:47:23.108778 kernel: PCI: Using ACPI for IRQ routing Dec 13 08:47:23.108799 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 13 08:47:23.108821 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 13 08:47:23.108842 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Dec 13 08:47:23.108994 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Dec 13 08:47:23.109144 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Dec 13 08:47:23.109296 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 13 08:47:23.109322 kernel: vgaarb: loaded Dec 13 08:47:23.110277 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 13 08:47:23.110292 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 13 08:47:23.110303 kernel: clocksource: Switched to clocksource kvm-clock Dec 13 08:47:23.110312 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 08:47:23.110323 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 08:47:23.110345 kernel: pnp: PnP ACPI init Dec 13 08:47:23.110355 kernel: pnp: PnP ACPI: found 4 devices Dec 13 08:47:23.110373 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 13 08:47:23.110382 kernel: NET: Registered PF_INET protocol family Dec 13 08:47:23.110392 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 08:47:23.110402 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Dec 13 08:47:23.110412 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 08:47:23.110422 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Dec 13 08:47:23.110431 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Dec 13 08:47:23.110441 kernel: TCP: Hash tables configured (established 16384 bind 16384) Dec 13 08:47:23.110451 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 13 08:47:23.110465 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 13 08:47:23.110474 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 08:47:23.110484 kernel: NET: Registered PF_XDP protocol family Dec 13 08:47:23.110621 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 13 08:47:23.110713 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 13 08:47:23.110804 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 13 08:47:23.110966 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Dec 13 08:47:23.111119 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Dec 13 08:47:23.114663 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Dec 13 08:47:23.114883 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Dec 13 08:47:23.114902 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Dec 13 08:47:23.115009 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7a0 took 49760 usecs Dec 13 08:47:23.115023 kernel: PCI: CLS 0 bytes, default 64 Dec 13 08:47:23.115034 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Dec 13 08:47:23.115044 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x21134dbeb26, max_idle_ns: 440795298546 ns Dec 13 08:47:23.115054 kernel: Initialise system trusted keyrings Dec 13 08:47:23.115073 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Dec 13 08:47:23.115083 kernel: Key type asymmetric registered Dec 13 08:47:23.115092 kernel: Asymmetric key parser 'x509' registered Dec 13 08:47:23.115102 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Dec 13 08:47:23.115112 kernel: io scheduler mq-deadline registered Dec 13 08:47:23.115121 kernel: io scheduler kyber registered Dec 13 08:47:23.115131 kernel: io scheduler bfq registered Dec 13 08:47:23.115141 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 13 08:47:23.115151 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Dec 13 08:47:23.115165 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Dec 13 08:47:23.115175 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Dec 13 08:47:23.115184 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 08:47:23.115194 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 13 08:47:23.115204 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 13 08:47:23.115214 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 13 08:47:23.115223 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 13 08:47:23.115413 kernel: rtc_cmos 00:03: RTC can wake from S4 Dec 13 08:47:23.115429 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 13 08:47:23.115665 kernel: rtc_cmos 00:03: registered as rtc0 Dec 13 08:47:23.115775 kernel: rtc_cmos 00:03: setting system clock to 2024-12-13T08:47:22 UTC (1734079642) Dec 13 08:47:23.115867 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Dec 13 08:47:23.115880 kernel: intel_pstate: CPU model not supported Dec 13 08:47:23.115890 kernel: NET: Registered PF_INET6 protocol family Dec 13 08:47:23.115899 kernel: Segment Routing with IPv6 Dec 13 08:47:23.115909 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 08:47:23.115919 kernel: NET: Registered PF_PACKET protocol family Dec 13 08:47:23.115936 kernel: Key type dns_resolver registered Dec 13 08:47:23.115946 kernel: IPI shorthand broadcast: enabled Dec 13 08:47:23.115955 kernel: sched_clock: Marking stable (1247005492, 189063643)->(1488168211, -52099076) Dec 13 08:47:23.115965 kernel: registered taskstats version 1 Dec 13 08:47:23.115975 kernel: Loading compiled-in X.509 certificates Dec 13 08:47:23.115985 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.65-flatcar: c82d546f528d79a5758dcebbc47fb6daf92836a0' Dec 13 08:47:23.115995 kernel: Key type .fscrypt registered Dec 13 08:47:23.116004 kernel: Key type fscrypt-provisioning registered Dec 13 08:47:23.116014 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 08:47:23.116027 kernel: ima: Allocated hash algorithm: sha1 Dec 13 08:47:23.116036 kernel: ima: No architecture policies found Dec 13 08:47:23.116046 kernel: clk: Disabling unused clocks Dec 13 08:47:23.116056 kernel: Freeing unused kernel image (initmem) memory: 42844K Dec 13 08:47:23.116065 kernel: Write protecting the kernel read-only data: 36864k Dec 13 08:47:23.116103 kernel: Freeing unused kernel image (rodata/data gap) memory: 1852K Dec 13 08:47:23.116117 kernel: Run /init as init process Dec 13 08:47:23.116127 kernel: with arguments: Dec 13 08:47:23.116138 kernel: /init Dec 13 08:47:23.116151 kernel: with environment: Dec 13 08:47:23.116161 kernel: HOME=/ Dec 13 08:47:23.116193 kernel: TERM=linux Dec 13 08:47:23.116203 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 08:47:23.116217 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 08:47:23.116230 systemd[1]: Detected virtualization kvm. Dec 13 08:47:23.116241 systemd[1]: Detected architecture x86-64. Dec 13 08:47:23.116255 systemd[1]: Running in initrd. Dec 13 08:47:23.116265 systemd[1]: No hostname configured, using default hostname. Dec 13 08:47:23.116276 systemd[1]: Hostname set to . Dec 13 08:47:23.116287 systemd[1]: Initializing machine ID from VM UUID. Dec 13 08:47:23.116297 systemd[1]: Queued start job for default target initrd.target. Dec 13 08:47:23.116308 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 08:47:23.116319 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 08:47:23.116331 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 13 08:47:23.116464 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 08:47:23.116474 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 13 08:47:23.116485 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 13 08:47:23.116498 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Dec 13 08:47:23.116509 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Dec 13 08:47:23.116519 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 08:47:23.116530 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 08:47:23.116544 systemd[1]: Reached target paths.target - Path Units. Dec 13 08:47:23.116555 systemd[1]: Reached target slices.target - Slice Units. Dec 13 08:47:23.116566 systemd[1]: Reached target swap.target - Swaps. Dec 13 08:47:23.116580 systemd[1]: Reached target timers.target - Timer Units. Dec 13 08:47:23.116591 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 08:47:23.116601 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 08:47:23.116621 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 13 08:47:23.116653 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Dec 13 08:47:23.116669 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 08:47:23.116686 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 08:47:23.116703 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 08:47:23.116718 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 08:47:23.116733 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 13 08:47:23.116751 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 08:47:23.116783 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 13 08:47:23.116806 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 08:47:23.116830 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 08:47:23.116853 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 08:47:23.116876 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:23.116939 systemd-journald[183]: Collecting audit messages is disabled. Dec 13 08:47:23.116997 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 13 08:47:23.117028 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 08:47:23.117044 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 08:47:23.117062 systemd-journald[183]: Journal started Dec 13 08:47:23.117103 systemd-journald[183]: Runtime Journal (/run/log/journal/af4b3ef6c7c7432d847dab2a3c164ea7) is 4.9M, max 39.3M, 34.4M free. Dec 13 08:47:23.117921 systemd-modules-load[184]: Inserted module 'overlay' Dec 13 08:47:23.127492 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 08:47:23.131367 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 08:47:23.146686 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 08:47:23.200293 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 08:47:23.200384 kernel: Bridge firewalling registered Dec 13 08:47:23.164936 systemd-modules-load[184]: Inserted module 'br_netfilter' Dec 13 08:47:23.207960 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 08:47:23.208965 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:23.209873 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 08:47:23.219761 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 08:47:23.226556 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 08:47:23.230940 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 08:47:23.244848 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 08:47:23.255583 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 08:47:23.265664 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 13 08:47:23.266832 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 08:47:23.269049 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 08:47:23.281639 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 08:47:23.296965 dracut-cmdline[214]: dracut-dracut-053 Dec 13 08:47:23.303179 dracut-cmdline[214]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=2fdbba50b59d8c8a9877a81151806ddc16f473fe99b9ba0d8825997d654583ff Dec 13 08:47:23.341357 systemd-resolved[219]: Positive Trust Anchors: Dec 13 08:47:23.341378 systemd-resolved[219]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 08:47:23.341495 systemd-resolved[219]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 08:47:23.346658 systemd-resolved[219]: Defaulting to hostname 'linux'. Dec 13 08:47:23.348678 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 08:47:23.350444 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 08:47:23.439411 kernel: SCSI subsystem initialized Dec 13 08:47:23.452443 kernel: Loading iSCSI transport class v2.0-870. Dec 13 08:47:23.466374 kernel: iscsi: registered transport (tcp) Dec 13 08:47:23.496397 kernel: iscsi: registered transport (qla4xxx) Dec 13 08:47:23.496553 kernel: QLogic iSCSI HBA Driver Dec 13 08:47:23.568362 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 13 08:47:23.575673 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 13 08:47:23.619197 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 08:47:23.619326 kernel: device-mapper: uevent: version 1.0.3 Dec 13 08:47:23.619378 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Dec 13 08:47:23.677414 kernel: raid6: avx2x4 gen() 15213 MB/s Dec 13 08:47:23.695393 kernel: raid6: avx2x2 gen() 15253 MB/s Dec 13 08:47:23.714094 kernel: raid6: avx2x1 gen() 12900 MB/s Dec 13 08:47:23.714196 kernel: raid6: using algorithm avx2x2 gen() 15253 MB/s Dec 13 08:47:23.733102 kernel: raid6: .... xor() 11701 MB/s, rmw enabled Dec 13 08:47:23.733198 kernel: raid6: using avx2x2 recovery algorithm Dec 13 08:47:23.765395 kernel: xor: automatically using best checksumming function avx Dec 13 08:47:23.996404 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 13 08:47:24.015513 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 13 08:47:24.023749 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 08:47:24.061523 systemd-udevd[402]: Using default interface naming scheme 'v255'. Dec 13 08:47:24.072035 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 08:47:24.079138 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 13 08:47:24.113102 dracut-pre-trigger[407]: rd.md=0: removing MD RAID activation Dec 13 08:47:24.163487 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 08:47:24.170737 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 08:47:24.251920 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 08:47:24.263976 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 13 08:47:24.301654 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 13 08:47:24.304382 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 08:47:24.305684 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 08:47:24.307744 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 08:47:24.318672 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 13 08:47:24.354048 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 13 08:47:24.359371 kernel: scsi host0: Virtio SCSI HBA Dec 13 08:47:24.380376 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Dec 13 08:47:24.473971 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Dec 13 08:47:24.474202 kernel: cryptd: max_cpu_qlen set to 1000 Dec 13 08:47:24.474226 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 08:47:24.474246 kernel: GPT:9289727 != 125829119 Dec 13 08:47:24.474264 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 08:47:24.474282 kernel: GPT:9289727 != 125829119 Dec 13 08:47:24.474317 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 08:47:24.474377 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 08:47:24.474411 kernel: AVX2 version of gcm_enc/dec engaged. Dec 13 08:47:24.474430 kernel: AES CTR mode by8 optimization enabled Dec 13 08:47:24.474447 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Dec 13 08:47:24.506743 kernel: virtio_blk virtio5: [vdb] 952 512-byte logical blocks (487 kB/476 KiB) Dec 13 08:47:24.507217 kernel: libata version 3.00 loaded. Dec 13 08:47:24.507243 kernel: ata_piix 0000:00:01.1: version 2.13 Dec 13 08:47:24.522888 kernel: scsi host1: ata_piix Dec 13 08:47:24.523173 kernel: scsi host2: ata_piix Dec 13 08:47:24.523397 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Dec 13 08:47:24.523420 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Dec 13 08:47:24.523438 kernel: ACPI: bus type USB registered Dec 13 08:47:24.496273 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 08:47:24.532044 kernel: usbcore: registered new interface driver usbfs Dec 13 08:47:24.532093 kernel: usbcore: registered new interface driver hub Dec 13 08:47:24.532121 kernel: usbcore: registered new device driver usb Dec 13 08:47:24.496563 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 08:47:24.498518 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 08:47:24.499250 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 08:47:24.499617 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:24.501502 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:24.509433 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:24.596364 kernel: BTRFS: device fsid c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be devid 1 transid 41 /dev/vda3 scanned by (udev-worker) (446) Dec 13 08:47:24.614370 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (444) Dec 13 08:47:24.615815 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Dec 13 08:47:24.648190 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:24.666568 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Dec 13 08:47:24.681852 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 08:47:24.697800 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Dec 13 08:47:24.698633 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Dec 13 08:47:24.709662 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 13 08:47:24.717152 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 13 08:47:24.733612 disk-uuid[539]: Primary Header is updated. Dec 13 08:47:24.733612 disk-uuid[539]: Secondary Entries is updated. Dec 13 08:47:24.733612 disk-uuid[539]: Secondary Header is updated. Dec 13 08:47:24.743900 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Dec 13 08:47:24.752240 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Dec 13 08:47:24.752549 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Dec 13 08:47:24.752775 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Dec 13 08:47:24.753005 kernel: hub 1-0:1.0: USB hub found Dec 13 08:47:24.753263 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 08:47:24.753290 kernel: hub 1-0:1.0: 2 ports detected Dec 13 08:47:24.752307 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 08:47:24.758378 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 08:47:24.777371 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 08:47:25.772433 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 08:47:25.773954 disk-uuid[544]: The operation has completed successfully. Dec 13 08:47:25.859357 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 08:47:25.859574 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 13 08:47:25.885773 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Dec 13 08:47:25.900579 sh[561]: Success Dec 13 08:47:25.923495 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Dec 13 08:47:26.041688 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Dec 13 08:47:26.049679 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Dec 13 08:47:26.055912 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Dec 13 08:47:26.104784 kernel: BTRFS info (device dm-0): first mount of filesystem c3b72f8a-27ca-4d37-9d0e-1ec3c4bdc3be Dec 13 08:47:26.104907 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 13 08:47:26.106936 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Dec 13 08:47:26.110165 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 13 08:47:26.110289 kernel: BTRFS info (device dm-0): using free space tree Dec 13 08:47:26.127372 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Dec 13 08:47:26.129283 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 13 08:47:26.135704 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 13 08:47:26.144615 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 13 08:47:26.172287 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 08:47:26.172433 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 08:47:26.172462 kernel: BTRFS info (device vda6): using free space tree Dec 13 08:47:26.180506 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 08:47:26.200848 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 08:47:26.202722 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 08:47:26.214714 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 13 08:47:26.222708 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 13 08:47:26.362223 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 08:47:26.389706 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 08:47:26.432145 ignition[673]: Ignition 2.19.0 Dec 13 08:47:26.432167 ignition[673]: Stage: fetch-offline Dec 13 08:47:26.432240 ignition[673]: no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:26.432256 ignition[673]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:26.432451 ignition[673]: parsed url from cmdline: "" Dec 13 08:47:26.432456 ignition[673]: no config URL provided Dec 13 08:47:26.432464 ignition[673]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 08:47:26.432480 ignition[673]: no config at "/usr/lib/ignition/user.ign" Dec 13 08:47:26.432487 ignition[673]: failed to fetch config: resource requires networking Dec 13 08:47:26.437822 ignition[673]: Ignition finished successfully Dec 13 08:47:26.441035 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 08:47:26.454921 systemd-networkd[749]: lo: Link UP Dec 13 08:47:26.454940 systemd-networkd[749]: lo: Gained carrier Dec 13 08:47:26.458327 systemd-networkd[749]: Enumeration completed Dec 13 08:47:26.458608 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 08:47:26.458989 systemd-networkd[749]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Dec 13 08:47:26.458995 systemd-networkd[749]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Dec 13 08:47:26.459981 systemd[1]: Reached target network.target - Network. Dec 13 08:47:26.460104 systemd-networkd[749]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 08:47:26.460110 systemd-networkd[749]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 08:47:26.461123 systemd-networkd[749]: eth0: Link UP Dec 13 08:47:26.461130 systemd-networkd[749]: eth0: Gained carrier Dec 13 08:47:26.461143 systemd-networkd[749]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Dec 13 08:47:26.466498 systemd-networkd[749]: eth1: Link UP Dec 13 08:47:26.466504 systemd-networkd[749]: eth1: Gained carrier Dec 13 08:47:26.466524 systemd-networkd[749]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Dec 13 08:47:26.471707 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 13 08:47:26.483469 systemd-networkd[749]: eth0: DHCPv4 address 64.23.218.120/19, gateway 64.23.192.1 acquired from 169.254.169.253 Dec 13 08:47:26.487469 systemd-networkd[749]: eth1: DHCPv4 address 10.124.0.16/20, gateway 10.124.0.1 acquired from 169.254.169.253 Dec 13 08:47:26.508354 ignition[753]: Ignition 2.19.0 Dec 13 08:47:26.508368 ignition[753]: Stage: fetch Dec 13 08:47:26.508725 ignition[753]: no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:26.508746 ignition[753]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:26.508932 ignition[753]: parsed url from cmdline: "" Dec 13 08:47:26.508938 ignition[753]: no config URL provided Dec 13 08:47:26.508946 ignition[753]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 08:47:26.508959 ignition[753]: no config at "/usr/lib/ignition/user.ign" Dec 13 08:47:26.508990 ignition[753]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Dec 13 08:47:26.525780 ignition[753]: GET result: OK Dec 13 08:47:26.525946 ignition[753]: parsing config with SHA512: aa5787e7faf0e8b6031ab46a67afa13eb3c0a29f9c3d88417c103bc6b95daf8db723349b7d292b7f07e24e6b86e1be5bc70dd434cf99b41a4339b5e056b12578 Dec 13 08:47:26.534876 unknown[753]: fetched base config from "system" Dec 13 08:47:26.534892 unknown[753]: fetched base config from "system" Dec 13 08:47:26.535709 ignition[753]: fetch: fetch complete Dec 13 08:47:26.534903 unknown[753]: fetched user config from "digitalocean" Dec 13 08:47:26.535720 ignition[753]: fetch: fetch passed Dec 13 08:47:26.540088 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 13 08:47:26.535813 ignition[753]: Ignition finished successfully Dec 13 08:47:26.546695 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 13 08:47:26.591369 ignition[760]: Ignition 2.19.0 Dec 13 08:47:26.591386 ignition[760]: Stage: kargs Dec 13 08:47:26.591686 ignition[760]: no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:26.591702 ignition[760]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:26.593432 ignition[760]: kargs: kargs passed Dec 13 08:47:26.595201 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 13 08:47:26.593526 ignition[760]: Ignition finished successfully Dec 13 08:47:26.602721 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 13 08:47:26.639400 ignition[766]: Ignition 2.19.0 Dec 13 08:47:26.639419 ignition[766]: Stage: disks Dec 13 08:47:26.639760 ignition[766]: no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:26.639794 ignition[766]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:26.643402 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 13 08:47:26.641437 ignition[766]: disks: disks passed Dec 13 08:47:26.641536 ignition[766]: Ignition finished successfully Dec 13 08:47:26.652383 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 13 08:47:26.654660 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 13 08:47:26.655484 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 08:47:26.657145 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 08:47:26.658850 systemd[1]: Reached target basic.target - Basic System. Dec 13 08:47:26.667685 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 13 08:47:26.714026 systemd-fsck[774]: ROOT: clean, 14/553520 files, 52654/553472 blocks Dec 13 08:47:26.721119 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 13 08:47:26.726563 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 13 08:47:26.895016 kernel: EXT4-fs (vda9): mounted filesystem 390119fa-ab9c-4f50-b046-3b5c76c46193 r/w with ordered data mode. Quota mode: none. Dec 13 08:47:26.895880 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 13 08:47:26.897533 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 13 08:47:26.907626 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 08:47:26.912542 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 13 08:47:26.914326 systemd[1]: Starting flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent... Dec 13 08:47:26.928651 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Dec 13 08:47:26.933410 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 08:47:26.934758 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 08:47:26.948409 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (782) Dec 13 08:47:26.948475 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 08:47:26.948505 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 08:47:26.948535 kernel: BTRFS info (device vda6): using free space tree Dec 13 08:47:26.952120 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 13 08:47:26.959395 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 08:47:26.963728 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 13 08:47:26.972646 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 08:47:27.059559 coreos-metadata[784]: Dec 13 08:47:27.058 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Dec 13 08:47:27.070678 initrd-setup-root[813]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 08:47:27.074186 coreos-metadata[784]: Dec 13 08:47:27.074 INFO Fetch successful Dec 13 08:47:27.093878 initrd-setup-root[820]: cut: /sysroot/etc/group: No such file or directory Dec 13 08:47:27.101083 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Dec 13 08:47:27.107782 systemd[1]: Finished flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent. Dec 13 08:47:27.141487 coreos-metadata[785]: Dec 13 08:47:27.101 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Dec 13 08:47:27.153511 coreos-metadata[785]: Dec 13 08:47:27.144 INFO Fetch successful Dec 13 08:47:27.163611 initrd-setup-root[828]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 08:47:27.214552 coreos-metadata[785]: Dec 13 08:47:27.214 INFO wrote hostname ci-4081.2.1-7-8969f226b5 to /sysroot/etc/hostname Dec 13 08:47:27.226516 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 13 08:47:27.237812 initrd-setup-root[835]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 08:47:27.426520 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 13 08:47:27.434595 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 13 08:47:27.448799 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 13 08:47:27.463423 kernel: BTRFS info (device vda6): last unmount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 08:47:27.461873 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 13 08:47:27.492513 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 13 08:47:27.521478 ignition[904]: INFO : Ignition 2.19.0 Dec 13 08:47:27.523818 ignition[904]: INFO : Stage: mount Dec 13 08:47:27.526613 ignition[904]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:27.526613 ignition[904]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:27.528919 ignition[904]: INFO : mount: mount passed Dec 13 08:47:27.528919 ignition[904]: INFO : Ignition finished successfully Dec 13 08:47:27.532957 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 13 08:47:27.539580 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 13 08:47:27.572770 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 13 08:47:27.590377 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (915) Dec 13 08:47:27.593375 kernel: BTRFS info (device vda6): first mount of filesystem db063747-cac8-4176-8963-c216c1b11dcb Dec 13 08:47:27.597216 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 08:47:27.597386 kernel: BTRFS info (device vda6): using free space tree Dec 13 08:47:27.602455 kernel: BTRFS info (device vda6): auto enabling async discard Dec 13 08:47:27.606413 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 13 08:47:27.649764 ignition[932]: INFO : Ignition 2.19.0 Dec 13 08:47:27.650912 ignition[932]: INFO : Stage: files Dec 13 08:47:27.650912 ignition[932]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:27.650912 ignition[932]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:27.653623 ignition[932]: DEBUG : files: compiled without relabeling support, skipping Dec 13 08:47:27.654854 ignition[932]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 08:47:27.654854 ignition[932]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 08:47:27.661038 ignition[932]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 08:47:27.662462 ignition[932]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 08:47:27.663633 ignition[932]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 08:47:27.663063 unknown[932]: wrote ssh authorized keys file for user: core Dec 13 08:47:27.667174 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Dec 13 08:47:27.668565 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Dec 13 08:47:27.723032 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 13 08:47:27.807183 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 13 08:47:27.808843 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 08:47:27.822167 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 08:47:27.822167 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 13 08:47:27.822167 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 13 08:47:27.822167 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 13 08:47:27.822167 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Dec 13 08:47:27.948744 systemd-networkd[749]: eth0: Gained IPv6LL Dec 13 08:47:28.204711 systemd-networkd[749]: eth1: Gained IPv6LL Dec 13 08:47:28.354695 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Dec 13 08:47:29.305073 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Dec 13 08:47:29.305073 ignition[932]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Dec 13 08:47:29.308841 ignition[932]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 13 08:47:29.310285 ignition[932]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 13 08:47:29.310285 ignition[932]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Dec 13 08:47:29.310285 ignition[932]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Dec 13 08:47:29.310285 ignition[932]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Dec 13 08:47:29.316928 ignition[932]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 08:47:29.316928 ignition[932]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 08:47:29.316928 ignition[932]: INFO : files: files passed Dec 13 08:47:29.316928 ignition[932]: INFO : Ignition finished successfully Dec 13 08:47:29.314365 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 13 08:47:29.325751 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 13 08:47:29.331776 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 13 08:47:29.333777 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 08:47:29.333922 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 13 08:47:29.365975 initrd-setup-root-after-ignition[961]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 08:47:29.365975 initrd-setup-root-after-ignition[961]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 13 08:47:29.370059 initrd-setup-root-after-ignition[965]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 08:47:29.373321 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 08:47:29.375090 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 13 08:47:29.381719 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 13 08:47:29.452355 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 08:47:29.452549 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 13 08:47:29.454184 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 13 08:47:29.455672 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 13 08:47:29.457492 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 13 08:47:29.466701 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 13 08:47:29.495566 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 08:47:29.503653 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 13 08:47:29.529116 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 13 08:47:29.530188 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 08:47:29.531778 systemd[1]: Stopped target timers.target - Timer Units. Dec 13 08:47:29.533319 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 08:47:29.533601 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 13 08:47:29.535184 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 13 08:47:29.536235 systemd[1]: Stopped target basic.target - Basic System. Dec 13 08:47:29.537871 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 13 08:47:29.539122 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 13 08:47:29.540415 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 13 08:47:29.546855 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 13 08:47:29.548415 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 13 08:47:29.549951 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 13 08:47:29.552183 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 13 08:47:29.553466 systemd[1]: Stopped target swap.target - Swaps. Dec 13 08:47:29.554498 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 08:47:29.554772 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 13 08:47:29.556442 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 13 08:47:29.558181 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 08:47:29.559718 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 13 08:47:29.561377 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 08:47:29.562412 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 08:47:29.562665 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 13 08:47:29.564530 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 08:47:29.564778 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 13 08:47:29.567989 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 08:47:29.568201 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 13 08:47:29.570133 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Dec 13 08:47:29.570362 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 13 08:47:29.583533 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 13 08:47:29.588448 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 13 08:47:29.590600 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 08:47:29.591488 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 08:47:29.597550 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 08:47:29.597824 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 13 08:47:29.611873 ignition[985]: INFO : Ignition 2.19.0 Dec 13 08:47:29.611873 ignition[985]: INFO : Stage: umount Dec 13 08:47:29.616449 ignition[985]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 08:47:29.616449 ignition[985]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Dec 13 08:47:29.616449 ignition[985]: INFO : umount: umount passed Dec 13 08:47:29.616449 ignition[985]: INFO : Ignition finished successfully Dec 13 08:47:29.613133 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 08:47:29.614271 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 13 08:47:29.616315 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 08:47:29.618682 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 13 08:47:29.624531 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 08:47:29.624646 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 13 08:47:29.625695 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 08:47:29.625776 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 13 08:47:29.626540 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 13 08:47:29.626622 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 13 08:47:29.635522 systemd[1]: Stopped target network.target - Network. Dec 13 08:47:29.636355 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 08:47:29.636488 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 13 08:47:29.637536 systemd[1]: Stopped target paths.target - Path Units. Dec 13 08:47:29.638233 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 08:47:29.645501 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 08:47:29.646400 systemd[1]: Stopped target slices.target - Slice Units. Dec 13 08:47:29.647105 systemd[1]: Stopped target sockets.target - Socket Units. Dec 13 08:47:29.658786 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 08:47:29.658880 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 13 08:47:29.659780 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 08:47:29.659943 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 13 08:47:29.661382 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 08:47:29.661486 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 13 08:47:29.662256 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 13 08:47:29.662369 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 13 08:47:29.663539 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 13 08:47:29.665659 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 13 08:47:29.674434 systemd-networkd[749]: eth0: DHCPv6 lease lost Dec 13 08:47:29.676702 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 08:47:29.679328 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 08:47:29.679454 systemd-networkd[749]: eth1: DHCPv6 lease lost Dec 13 08:47:29.681528 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 13 08:47:29.687426 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 08:47:29.687712 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 13 08:47:29.695244 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 08:47:29.695406 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 13 08:47:29.714651 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 13 08:47:29.724272 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 08:47:29.724374 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 13 08:47:29.728669 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 08:47:29.728869 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 13 08:47:29.733858 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 08:47:29.733966 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 13 08:47:29.735497 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 13 08:47:29.735623 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 08:47:29.738652 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 08:47:29.744855 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 08:47:29.745020 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 13 08:47:29.758915 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 08:47:29.759164 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 08:47:29.763843 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 08:47:29.763964 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 13 08:47:29.764966 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 08:47:29.765022 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 08:47:29.768541 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 08:47:29.768651 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 13 08:47:29.770682 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 08:47:29.770779 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 13 08:47:29.772936 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 08:47:29.773164 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 13 08:47:29.775544 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 08:47:29.775648 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 13 08:47:29.782682 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 13 08:47:29.784931 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 08:47:29.785061 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 08:47:29.788948 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 13 08:47:29.789072 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 08:47:29.791329 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 08:47:29.791462 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 08:47:29.793701 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 08:47:29.793818 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:29.799817 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 08:47:29.799969 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 13 08:47:29.811410 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 08:47:29.811609 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 13 08:47:29.813565 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 13 08:47:29.822678 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 13 08:47:29.837321 systemd[1]: Switching root. Dec 13 08:47:29.907415 systemd-journald[183]: Journal stopped Dec 13 08:47:31.736188 systemd-journald[183]: Received SIGTERM from PID 1 (systemd). Dec 13 08:47:31.736335 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 08:47:31.736451 kernel: SELinux: policy capability open_perms=1 Dec 13 08:47:31.736475 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 08:47:31.736549 kernel: SELinux: policy capability always_check_network=0 Dec 13 08:47:31.736581 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 08:47:31.736620 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 08:47:31.737523 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 08:47:31.737556 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 08:47:31.737590 kernel: audit: type=1403 audit(1734079650.140:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 08:47:31.737619 systemd[1]: Successfully loaded SELinux policy in 62.380ms. Dec 13 08:47:31.737670 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 20.443ms. Dec 13 08:47:31.737717 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Dec 13 08:47:31.737746 systemd[1]: Detected virtualization kvm. Dec 13 08:47:31.737773 systemd[1]: Detected architecture x86-64. Dec 13 08:47:31.737811 systemd[1]: Detected first boot. Dec 13 08:47:31.737839 systemd[1]: Hostname set to . Dec 13 08:47:31.737867 systemd[1]: Initializing machine ID from VM UUID. Dec 13 08:47:31.737891 zram_generator::config[1028]: No configuration found. Dec 13 08:47:31.737916 systemd[1]: Populated /etc with preset unit settings. Dec 13 08:47:31.737954 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 13 08:47:31.737982 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 13 08:47:31.738010 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 13 08:47:31.738256 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 13 08:47:31.738290 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 13 08:47:31.738320 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 13 08:47:31.741481 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 13 08:47:31.741532 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 13 08:47:31.741564 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 13 08:47:31.741587 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 13 08:47:31.741611 systemd[1]: Created slice user.slice - User and Session Slice. Dec 13 08:47:31.741634 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 13 08:47:31.741659 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 13 08:47:31.741683 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 13 08:47:31.741707 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 13 08:47:31.741732 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 13 08:47:31.741758 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 13 08:47:31.741786 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 13 08:47:31.741810 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 13 08:47:31.741832 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 13 08:47:31.741857 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 13 08:47:31.741881 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 13 08:47:31.741905 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 13 08:47:31.741933 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 13 08:47:31.741959 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 13 08:47:31.741993 systemd[1]: Reached target slices.target - Slice Units. Dec 13 08:47:31.742027 systemd[1]: Reached target swap.target - Swaps. Dec 13 08:47:31.742061 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 13 08:47:31.742086 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 13 08:47:31.742110 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 13 08:47:31.742133 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 13 08:47:31.742156 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 13 08:47:31.742179 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 13 08:47:31.742208 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 13 08:47:31.742232 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 13 08:47:31.742256 systemd[1]: Mounting media.mount - External Media Directory... Dec 13 08:47:31.742281 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:31.742305 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 13 08:47:31.742328 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 13 08:47:31.742376 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 13 08:47:31.742402 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 08:47:31.742431 systemd[1]: Reached target machines.target - Containers. Dec 13 08:47:31.742455 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 13 08:47:31.742480 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 08:47:31.742503 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 13 08:47:31.742527 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 13 08:47:31.742550 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 08:47:31.742574 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 08:47:31.742598 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 08:47:31.742622 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 13 08:47:31.742652 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 08:47:31.742677 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 08:47:31.742732 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 13 08:47:31.742756 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 13 08:47:31.742780 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 13 08:47:31.742805 systemd[1]: Stopped systemd-fsck-usr.service. Dec 13 08:47:31.742829 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 13 08:47:31.742854 kernel: loop: module loaded Dec 13 08:47:31.742885 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 13 08:47:31.742909 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 13 08:47:31.742932 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 13 08:47:31.746409 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 13 08:47:31.746467 systemd[1]: verity-setup.service: Deactivated successfully. Dec 13 08:47:31.746492 systemd[1]: Stopped verity-setup.service. Dec 13 08:47:31.746517 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:31.746541 kernel: ACPI: bus type drm_connector registered Dec 13 08:47:31.746566 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 13 08:47:31.746600 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 13 08:47:31.746626 systemd[1]: Mounted media.mount - External Media Directory. Dec 13 08:47:31.746651 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 13 08:47:31.746675 kernel: fuse: init (API version 7.39) Dec 13 08:47:31.746697 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 13 08:47:31.746729 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 13 08:47:31.746760 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 13 08:47:31.746814 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 08:47:31.746840 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 13 08:47:31.746866 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 08:47:31.746894 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 08:47:31.746918 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 08:47:31.746944 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 08:47:31.746968 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 08:47:31.746992 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 08:47:31.747017 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 08:47:31.747094 systemd-journald[1101]: Collecting audit messages is disabled. Dec 13 08:47:31.747142 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 13 08:47:31.747170 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 08:47:31.747195 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 08:47:31.747220 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 13 08:47:31.747245 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 13 08:47:31.747269 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 13 08:47:31.747291 systemd-journald[1101]: Journal started Dec 13 08:47:31.748381 systemd-journald[1101]: Runtime Journal (/run/log/journal/af4b3ef6c7c7432d847dab2a3c164ea7) is 4.9M, max 39.3M, 34.4M free. Dec 13 08:47:31.748494 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 13 08:47:31.193295 systemd[1]: Queued start job for default target multi-user.target. Dec 13 08:47:31.230608 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Dec 13 08:47:31.231383 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 13 08:47:31.763374 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 13 08:47:31.773408 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 13 08:47:31.781841 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 08:47:31.786380 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 13 08:47:31.791371 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Dec 13 08:47:31.809381 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 13 08:47:31.819407 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 13 08:47:31.819498 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 08:47:31.826371 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 13 08:47:31.830611 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 08:47:31.842433 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 13 08:47:31.846386 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 08:47:31.861372 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 13 08:47:31.879594 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 13 08:47:31.896792 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 13 08:47:31.903662 systemd[1]: Started systemd-journald.service - Journal Service. Dec 13 08:47:31.909922 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 13 08:47:31.911146 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 13 08:47:31.912676 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 13 08:47:31.916095 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 13 08:47:31.967291 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 13 08:47:31.989367 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 13 08:47:32.001704 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 13 08:47:32.007731 kernel: loop0: detected capacity change from 0 to 205544 Dec 13 08:47:32.012733 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Dec 13 08:47:32.014928 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 13 08:47:32.028675 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Dec 13 08:47:32.051745 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 13 08:47:32.091413 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 08:47:32.097958 systemd-journald[1101]: Time spent on flushing to /var/log/journal/af4b3ef6c7c7432d847dab2a3c164ea7 is 82.129ms for 997 entries. Dec 13 08:47:32.097958 systemd-journald[1101]: System Journal (/var/log/journal/af4b3ef6c7c7432d847dab2a3c164ea7) is 8.0M, max 195.6M, 187.6M free. Dec 13 08:47:32.210432 systemd-journald[1101]: Received client request to flush runtime journal. Dec 13 08:47:32.210538 kernel: loop1: detected capacity change from 0 to 140768 Dec 13 08:47:32.103088 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 08:47:32.108328 systemd-tmpfiles[1131]: ACLs are not supported, ignoring. Dec 13 08:47:32.108396 systemd-tmpfiles[1131]: ACLs are not supported, ignoring. Dec 13 08:47:32.109647 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Dec 13 08:47:32.131380 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 13 08:47:32.149999 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 13 08:47:32.156574 udevadm[1159]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Dec 13 08:47:32.217241 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 13 08:47:32.230847 kernel: loop2: detected capacity change from 0 to 142488 Dec 13 08:47:32.268596 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 13 08:47:32.281768 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 13 08:47:32.340492 kernel: loop3: detected capacity change from 0 to 8 Dec 13 08:47:32.352866 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Dec 13 08:47:32.352903 systemd-tmpfiles[1171]: ACLs are not supported, ignoring. Dec 13 08:47:32.374413 kernel: loop4: detected capacity change from 0 to 205544 Dec 13 08:47:32.375953 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 13 08:47:32.433557 kernel: loop5: detected capacity change from 0 to 140768 Dec 13 08:47:32.486399 kernel: loop6: detected capacity change from 0 to 142488 Dec 13 08:47:32.533477 kernel: loop7: detected capacity change from 0 to 8 Dec 13 08:47:32.535547 (sd-merge)[1176]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Dec 13 08:47:32.539983 (sd-merge)[1176]: Merged extensions into '/usr'. Dec 13 08:47:32.552266 systemd[1]: Reloading requested from client PID 1130 ('systemd-sysext') (unit systemd-sysext.service)... Dec 13 08:47:32.552723 systemd[1]: Reloading... Dec 13 08:47:32.786426 zram_generator::config[1203]: No configuration found. Dec 13 08:47:33.053375 ldconfig[1126]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 08:47:33.160980 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 08:47:33.277026 systemd[1]: Reloading finished in 723 ms. Dec 13 08:47:33.318379 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 13 08:47:33.321685 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 13 08:47:33.336820 systemd[1]: Starting ensure-sysext.service... Dec 13 08:47:33.347322 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 13 08:47:33.373566 systemd[1]: Reloading requested from client PID 1246 ('systemctl') (unit ensure-sysext.service)... Dec 13 08:47:33.373595 systemd[1]: Reloading... Dec 13 08:47:33.417160 systemd-tmpfiles[1247]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 08:47:33.418684 systemd-tmpfiles[1247]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Dec 13 08:47:33.420724 systemd-tmpfiles[1247]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 08:47:33.421403 systemd-tmpfiles[1247]: ACLs are not supported, ignoring. Dec 13 08:47:33.421627 systemd-tmpfiles[1247]: ACLs are not supported, ignoring. Dec 13 08:47:33.429542 systemd-tmpfiles[1247]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 08:47:33.429815 systemd-tmpfiles[1247]: Skipping /boot Dec 13 08:47:33.469447 systemd-tmpfiles[1247]: Detected autofs mount point /boot during canonicalization of boot. Dec 13 08:47:33.472161 systemd-tmpfiles[1247]: Skipping /boot Dec 13 08:47:33.593430 zram_generator::config[1274]: No configuration found. Dec 13 08:47:33.862867 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 08:47:34.004792 systemd[1]: Reloading finished in 630 ms. Dec 13 08:47:34.026802 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 13 08:47:34.033214 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 13 08:47:34.055738 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Dec 13 08:47:34.060671 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 13 08:47:34.066677 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 13 08:47:34.079750 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 13 08:47:34.086738 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 13 08:47:34.092692 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 13 08:47:34.104195 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.104653 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 08:47:34.112797 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 08:47:34.121946 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 08:47:34.127937 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 08:47:34.129631 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 08:47:34.129946 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.137584 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.137971 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 08:47:34.138401 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 08:47:34.138634 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.146717 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.147117 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 08:47:34.154937 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 13 08:47:34.156630 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 08:47:34.157041 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.160895 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 08:47:34.161152 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 08:47:34.163809 systemd[1]: Finished ensure-sysext.service. Dec 13 08:47:34.188825 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 13 08:47:34.200800 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 13 08:47:34.203272 systemd-udevd[1330]: Using default interface naming scheme 'v255'. Dec 13 08:47:34.244037 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 13 08:47:34.254431 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 13 08:47:34.256053 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 08:47:34.269227 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 08:47:34.271097 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 08:47:34.274086 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 08:47:34.275477 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 08:47:34.275789 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 08:47:34.279212 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 08:47:34.284089 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 08:47:34.284429 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 13 08:47:34.310969 augenrules[1353]: No rules Dec 13 08:47:34.315760 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Dec 13 08:47:34.317477 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 13 08:47:34.328773 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 13 08:47:34.331669 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 13 08:47:34.344669 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 13 08:47:34.345817 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 13 08:47:34.370981 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 13 08:47:34.559376 systemd-networkd[1365]: lo: Link UP Dec 13 08:47:34.559390 systemd-networkd[1365]: lo: Gained carrier Dec 13 08:47:34.565054 systemd-networkd[1365]: Enumeration completed Dec 13 08:47:34.565263 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 13 08:47:34.575803 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 13 08:47:34.604101 systemd-resolved[1326]: Positive Trust Anchors: Dec 13 08:47:34.604132 systemd-resolved[1326]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 08:47:34.604185 systemd-resolved[1326]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 13 08:47:34.612151 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 13 08:47:34.613858 systemd[1]: Reached target time-set.target - System Time Set. Dec 13 08:47:34.615782 systemd-resolved[1326]: Using system hostname 'ci-4081.2.1-7-8969f226b5'. Dec 13 08:47:34.620192 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 13 08:47:34.621486 systemd[1]: Reached target network.target - Network. Dec 13 08:47:34.622837 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 13 08:47:34.663024 systemd-networkd[1365]: eth1: Configuring with /run/systemd/network/10-4a:4f:4f:69:68:53.network. Dec 13 08:47:34.664191 systemd-networkd[1365]: eth1: Link UP Dec 13 08:47:34.664205 systemd-networkd[1365]: eth1: Gained carrier Dec 13 08:47:34.671063 systemd-timesyncd[1340]: Network configuration changed, trying to establish connection. Dec 13 08:47:34.677413 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1362) Dec 13 08:47:34.677789 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 13 08:47:34.715137 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1362) Dec 13 08:47:34.704524 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Dec 13 08:47:34.705416 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.705707 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 13 08:47:34.713680 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 13 08:47:34.725668 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 13 08:47:34.735659 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 13 08:47:34.737477 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 13 08:47:34.737543 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 08:47:34.737578 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 08:47:34.742639 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Dec 13 08:47:34.747380 kernel: ACPI: button: Power Button [PWRF] Dec 13 08:47:34.787692 kernel: ISO 9660 Extensions: RRIP_1991A Dec 13 08:47:34.803719 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1381) Dec 13 08:47:34.804416 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Dec 13 08:47:34.806794 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 08:47:34.807435 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 13 08:47:34.808879 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 08:47:34.811434 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 13 08:47:34.813054 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 08:47:34.813272 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 13 08:47:34.820263 systemd-networkd[1365]: eth0: Configuring with /run/systemd/network/10-f6:59:8c:9a:0c:77.network. Dec 13 08:47:34.821539 systemd-networkd[1365]: eth0: Link UP Dec 13 08:47:34.821551 systemd-networkd[1365]: eth0: Gained carrier Dec 13 08:47:34.837577 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 08:47:34.837658 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 13 08:47:34.870754 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Dec 13 08:47:34.884649 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Dec 13 08:47:34.998375 kernel: mousedev: PS/2 mouse device common for all mice Dec 13 08:47:35.034856 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:35.083074 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Dec 13 08:47:35.127978 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 13 08:47:35.146946 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Dec 13 08:47:35.147090 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Dec 13 08:47:35.155004 kernel: Console: switching to colour dummy device 80x25 Dec 13 08:47:35.157389 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 13 08:47:35.157952 kernel: [drm] features: -context_init Dec 13 08:47:35.158109 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 13 08:47:35.162442 kernel: [drm] number of scanouts: 1 Dec 13 08:47:35.162629 kernel: [drm] number of cap sets: 0 Dec 13 08:47:35.166387 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Dec 13 08:47:35.202503 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Dec 13 08:47:35.202620 kernel: Console: switching to colour frame buffer device 128x48 Dec 13 08:47:35.215812 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 13 08:47:35.216772 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 08:47:35.217529 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:35.236942 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:35.263518 kernel: EDAC MC: Ver: 3.0.0 Dec 13 08:47:35.273697 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 08:47:35.274891 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:35.286755 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 13 08:47:35.300011 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Dec 13 08:47:35.311816 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Dec 13 08:47:35.342509 lvm[1428]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 08:47:35.347468 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 13 08:47:35.391268 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Dec 13 08:47:35.393674 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 13 08:47:35.393881 systemd[1]: Reached target sysinit.target - System Initialization. Dec 13 08:47:35.394204 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 13 08:47:35.394444 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 13 08:47:35.394895 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 13 08:47:35.395198 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 13 08:47:35.395624 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 13 08:47:35.398932 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 08:47:35.399033 systemd[1]: Reached target paths.target - Path Units. Dec 13 08:47:35.399208 systemd[1]: Reached target timers.target - Timer Units. Dec 13 08:47:35.402542 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 13 08:47:35.405081 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 13 08:47:35.417232 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 13 08:47:35.424704 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Dec 13 08:47:35.429461 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 13 08:47:35.431499 systemd[1]: Reached target sockets.target - Socket Units. Dec 13 08:47:35.432360 systemd[1]: Reached target basic.target - Basic System. Dec 13 08:47:35.435199 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 13 08:47:35.435320 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 13 08:47:35.439556 systemd[1]: Starting containerd.service - containerd container runtime... Dec 13 08:47:35.447964 lvm[1436]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 08:47:35.450651 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 13 08:47:35.458702 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 13 08:47:35.465604 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 13 08:47:35.474708 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 13 08:47:35.477183 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 13 08:47:35.487881 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 13 08:47:35.498543 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 13 08:47:35.513659 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 13 08:47:35.523734 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 13 08:47:35.531273 jq[1440]: false Dec 13 08:47:35.539737 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 13 08:47:35.543666 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 08:47:35.546712 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 08:47:35.554866 systemd[1]: Starting update-engine.service - Update Engine... Dec 13 08:47:35.568511 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 13 08:47:35.569384 coreos-metadata[1438]: Dec 13 08:47:35.569 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Dec 13 08:47:35.576415 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Dec 13 08:47:35.585113 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 08:47:35.587815 coreos-metadata[1438]: Dec 13 08:47:35.585 INFO Fetch successful Dec 13 08:47:35.585441 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 13 08:47:35.588719 dbus-daemon[1439]: [system] SELinux support is enabled Dec 13 08:47:35.592800 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 13 08:47:35.639836 (ntainerd)[1457]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Dec 13 08:47:35.648673 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 08:47:35.648735 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 13 08:47:35.656628 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 08:47:35.656793 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Dec 13 08:47:35.656830 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 13 08:47:35.672001 jq[1450]: true Dec 13 08:47:35.672127 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 08:47:35.674151 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 13 08:47:35.706373 update_engine[1449]: I20241213 08:47:35.702728 1449 main.cc:92] Flatcar Update Engine starting Dec 13 08:47:35.714622 tar[1452]: linux-amd64/helm Dec 13 08:47:35.717685 systemd[1]: Started update-engine.service - Update Engine. Dec 13 08:47:35.721146 update_engine[1449]: I20241213 08:47:35.720856 1449 update_check_scheduler.cc:74] Next update check in 6m50s Dec 13 08:47:35.727637 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 13 08:47:35.734503 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 13 08:47:35.739431 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 13 08:47:35.748412 extend-filesystems[1441]: Found loop4 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found loop5 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found loop6 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found loop7 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda1 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda2 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda3 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found usr Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda4 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda6 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda7 Dec 13 08:47:35.748412 extend-filesystems[1441]: Found vda9 Dec 13 08:47:35.748412 extend-filesystems[1441]: Checking size of /dev/vda9 Dec 13 08:47:35.779031 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 08:47:35.779392 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 13 08:47:35.819929 jq[1471]: true Dec 13 08:47:35.845848 extend-filesystems[1441]: Resized partition /dev/vda9 Dec 13 08:47:35.860095 extend-filesystems[1486]: resize2fs 1.47.1 (20-May-2024) Dec 13 08:47:35.876382 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Dec 13 08:47:35.933596 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1362) Dec 13 08:47:36.098331 sshd_keygen[1479]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 08:47:36.117377 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Dec 13 08:47:36.147535 systemd-logind[1448]: New seat seat0. Dec 13 08:47:36.207860 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 13 08:47:36.209664 systemd-logind[1448]: Watching system buttons on /dev/input/event1 (Power Button) Dec 13 08:47:36.209899 systemd-logind[1448]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 13 08:47:36.212700 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 13 08:47:36.213468 systemd[1]: Started systemd-logind.service - User Login Management. Dec 13 08:47:36.214210 extend-filesystems[1486]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 13 08:47:36.214210 extend-filesystems[1486]: old_desc_blocks = 1, new_desc_blocks = 8 Dec 13 08:47:36.214210 extend-filesystems[1486]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Dec 13 08:47:36.230650 extend-filesystems[1441]: Resized filesystem in /dev/vda9 Dec 13 08:47:36.230650 extend-filesystems[1441]: Found vdb Dec 13 08:47:36.248202 bash[1501]: Updated "/home/core/.ssh/authorized_keys" Dec 13 08:47:36.219435 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 08:47:36.219836 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 13 08:47:36.232020 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 13 08:47:36.267794 systemd[1]: Starting sshkeys.service... Dec 13 08:47:36.292623 locksmithd[1476]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 08:47:36.308654 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 08:47:36.309388 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 13 08:47:36.327223 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 13 08:47:36.342974 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 13 08:47:36.357039 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 13 08:47:36.396842 systemd-networkd[1365]: eth0: Gained IPv6LL Dec 13 08:47:36.405791 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 13 08:47:36.411065 systemd[1]: Reached target network-online.target - Network is Online. Dec 13 08:47:36.423912 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:47:36.435991 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 13 08:47:36.444568 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 13 08:47:36.460704 systemd-networkd[1365]: eth1: Gained IPv6LL Dec 13 08:47:36.463838 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 13 08:47:36.473825 coreos-metadata[1527]: Dec 13 08:47:36.473 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Dec 13 08:47:36.480072 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 13 08:47:36.483413 systemd[1]: Reached target getty.target - Login Prompts. Dec 13 08:47:36.506581 coreos-metadata[1527]: Dec 13 08:47:36.498 INFO Fetch successful Dec 13 08:47:36.551289 unknown[1527]: wrote ssh authorized keys file for user: core Dec 13 08:47:36.558983 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 13 08:47:36.596720 containerd[1457]: time="2024-12-13T08:47:36.596124213Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Dec 13 08:47:36.652861 update-ssh-keys[1546]: Updated "/home/core/.ssh/authorized_keys" Dec 13 08:47:36.656620 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 13 08:47:36.672842 systemd[1]: Finished sshkeys.service. Dec 13 08:47:36.707259 containerd[1457]: time="2024-12-13T08:47:36.706467294Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.714876 containerd[1457]: time="2024-12-13T08:47:36.714559159Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.65-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 08:47:36.714876 containerd[1457]: time="2024-12-13T08:47:36.714614894Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 08:47:36.714876 containerd[1457]: time="2024-12-13T08:47:36.714645451Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 08:47:36.714876 containerd[1457]: time="2024-12-13T08:47:36.714866106Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Dec 13 08:47:36.715197 containerd[1457]: time="2024-12-13T08:47:36.714920960Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715197 containerd[1457]: time="2024-12-13T08:47:36.715035510Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715197 containerd[1457]: time="2024-12-13T08:47:36.715062528Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715421 containerd[1457]: time="2024-12-13T08:47:36.715368253Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715421 containerd[1457]: time="2024-12-13T08:47:36.715405458Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715520 containerd[1457]: time="2024-12-13T08:47:36.715428514Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715520 containerd[1457]: time="2024-12-13T08:47:36.715443341Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.715613 containerd[1457]: time="2024-12-13T08:47:36.715591363Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.716442 containerd[1457]: time="2024-12-13T08:47:36.715966374Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 08:47:36.716442 containerd[1457]: time="2024-12-13T08:47:36.716184087Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 08:47:36.716442 containerd[1457]: time="2024-12-13T08:47:36.716217087Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 08:47:36.719642 containerd[1457]: time="2024-12-13T08:47:36.719566155Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 08:47:36.719860 containerd[1457]: time="2024-12-13T08:47:36.719702846Z" level=info msg="metadata content store policy set" policy=shared Dec 13 08:47:36.738752 containerd[1457]: time="2024-12-13T08:47:36.738689584Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 08:47:36.739491 containerd[1457]: time="2024-12-13T08:47:36.738830462Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 08:47:36.739491 containerd[1457]: time="2024-12-13T08:47:36.738984679Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Dec 13 08:47:36.739491 containerd[1457]: time="2024-12-13T08:47:36.739026119Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Dec 13 08:47:36.739491 containerd[1457]: time="2024-12-13T08:47:36.739050270Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 08:47:36.739491 containerd[1457]: time="2024-12-13T08:47:36.739288438Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 08:47:36.741831 containerd[1457]: time="2024-12-13T08:47:36.741766929Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 08:47:36.742042 containerd[1457]: time="2024-12-13T08:47:36.742008829Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Dec 13 08:47:36.742094 containerd[1457]: time="2024-12-13T08:47:36.742061891Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Dec 13 08:47:36.742138 containerd[1457]: time="2024-12-13T08:47:36.742092564Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Dec 13 08:47:36.742138 containerd[1457]: time="2024-12-13T08:47:36.742113101Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742138 containerd[1457]: time="2024-12-13T08:47:36.742132252Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742150011Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742169955Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742190176Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742209663Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742227648Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742280 containerd[1457]: time="2024-12-13T08:47:36.742245876Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742288202Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742310562Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742328214Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742364715Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742384717Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742411996Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742431612Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742465360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742485660Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742506627Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742523277Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.742546 containerd[1457]: time="2024-12-13T08:47:36.742539911Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742557562Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742578631Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742608062Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742624656Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742639626Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742700288Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742723575Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742740350Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742759244Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742783037Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742805499Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742822042Z" level=info msg="NRI interface is disabled by configuration." Dec 13 08:47:36.744714 containerd[1457]: time="2024-12-13T08:47:36.742837111Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 08:47:36.745556 containerd[1457]: time="2024-12-13T08:47:36.743217127Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 08:47:36.746241 containerd[1457]: time="2024-12-13T08:47:36.746073845Z" level=info msg="Connect containerd service" Dec 13 08:47:36.746241 containerd[1457]: time="2024-12-13T08:47:36.746162792Z" level=info msg="using legacy CRI server" Dec 13 08:47:36.746241 containerd[1457]: time="2024-12-13T08:47:36.746175080Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 13 08:47:36.746437 containerd[1457]: time="2024-12-13T08:47:36.746375843Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 08:47:36.748256 containerd[1457]: time="2024-12-13T08:47:36.747381746Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 08:47:36.750808 containerd[1457]: time="2024-12-13T08:47:36.750750197Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 08:47:36.750966 containerd[1457]: time="2024-12-13T08:47:36.750867492Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 08:47:36.753949 containerd[1457]: time="2024-12-13T08:47:36.752191494Z" level=info msg="Start subscribing containerd event" Dec 13 08:47:36.753949 containerd[1457]: time="2024-12-13T08:47:36.752280463Z" level=info msg="Start recovering state" Dec 13 08:47:36.756606 containerd[1457]: time="2024-12-13T08:47:36.756535799Z" level=info msg="Start event monitor" Dec 13 08:47:36.756606 containerd[1457]: time="2024-12-13T08:47:36.756607826Z" level=info msg="Start snapshots syncer" Dec 13 08:47:36.756804 containerd[1457]: time="2024-12-13T08:47:36.756630014Z" level=info msg="Start cni network conf syncer for default" Dec 13 08:47:36.756804 containerd[1457]: time="2024-12-13T08:47:36.756654330Z" level=info msg="Start streaming server" Dec 13 08:47:36.758401 containerd[1457]: time="2024-12-13T08:47:36.758332315Z" level=info msg="containerd successfully booted in 0.169315s" Dec 13 08:47:36.759483 systemd[1]: Started containerd.service - containerd container runtime. Dec 13 08:47:37.140680 tar[1452]: linux-amd64/LICENSE Dec 13 08:47:37.140680 tar[1452]: linux-amd64/README.md Dec 13 08:47:37.158662 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 13 08:47:38.080905 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:47:38.084521 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 13 08:47:38.090730 systemd[1]: Startup finished in 1.424s (kernel) + 7.373s (initrd) + 8.008s (userspace) = 16.807s. Dec 13 08:47:38.098207 (kubelet)[1560]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 08:47:39.069289 kubelet[1560]: E1213 08:47:39.069197 1560 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 08:47:39.072856 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 08:47:39.073094 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 08:47:39.074068 systemd[1]: kubelet.service: Consumed 1.422s CPU time. Dec 13 08:47:41.760938 systemd-timesyncd[1340]: Contacted time server 64.111.99.224:123 (1.flatcar.pool.ntp.org). Dec 13 08:47:41.761021 systemd-timesyncd[1340]: Initial clock synchronization to Fri 2024-12-13 08:47:41.760597 UTC. Dec 13 08:47:41.761118 systemd-resolved[1326]: Clock change detected. Flushing caches. Dec 13 08:47:46.039689 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 13 08:47:46.047424 systemd[1]: Started sshd@0-64.23.218.120:22-147.75.109.163:34766.service - OpenSSH per-connection server daemon (147.75.109.163:34766). Dec 13 08:47:46.160062 sshd[1573]: Accepted publickey for core from 147.75.109.163 port 34766 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:47:46.163191 sshd[1573]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:47:46.182900 systemd-logind[1448]: New session 1 of user core. Dec 13 08:47:46.184486 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 13 08:47:46.191407 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 13 08:47:46.216983 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 13 08:47:46.225702 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 13 08:47:46.245641 (systemd)[1577]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 08:47:46.449604 systemd[1577]: Queued start job for default target default.target. Dec 13 08:47:46.461073 systemd[1577]: Created slice app.slice - User Application Slice. Dec 13 08:47:46.461140 systemd[1577]: Reached target paths.target - Paths. Dec 13 08:47:46.461177 systemd[1577]: Reached target timers.target - Timers. Dec 13 08:47:46.464160 systemd[1577]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 13 08:47:46.488036 systemd[1577]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 13 08:47:46.488831 systemd[1577]: Reached target sockets.target - Sockets. Dec 13 08:47:46.489080 systemd[1577]: Reached target basic.target - Basic System. Dec 13 08:47:46.489294 systemd[1577]: Reached target default.target - Main User Target. Dec 13 08:47:46.489493 systemd[1577]: Startup finished in 232ms. Dec 13 08:47:46.489739 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 13 08:47:46.499697 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 13 08:47:46.592316 systemd[1]: Started sshd@1-64.23.218.120:22-147.75.109.163:34776.service - OpenSSH per-connection server daemon (147.75.109.163:34776). Dec 13 08:47:46.661344 sshd[1588]: Accepted publickey for core from 147.75.109.163 port 34776 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:47:46.663719 sshd[1588]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:47:46.673024 systemd-logind[1448]: New session 2 of user core. Dec 13 08:47:46.681251 systemd[1]: Started session-2.scope - Session 2 of User core. Dec 13 08:47:46.749302 sshd[1588]: pam_unix(sshd:session): session closed for user core Dec 13 08:47:46.762821 systemd[1]: sshd@1-64.23.218.120:22-147.75.109.163:34776.service: Deactivated successfully. Dec 13 08:47:46.766120 systemd[1]: session-2.scope: Deactivated successfully. Dec 13 08:47:46.770097 systemd-logind[1448]: Session 2 logged out. Waiting for processes to exit. Dec 13 08:47:46.777448 systemd[1]: Started sshd@2-64.23.218.120:22-147.75.109.163:34786.service - OpenSSH per-connection server daemon (147.75.109.163:34786). Dec 13 08:47:46.780296 systemd-logind[1448]: Removed session 2. Dec 13 08:47:46.835338 sshd[1595]: Accepted publickey for core from 147.75.109.163 port 34786 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:47:46.836542 sshd[1595]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:47:46.845987 systemd-logind[1448]: New session 3 of user core. Dec 13 08:47:46.854200 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 13 08:47:46.916908 sshd[1595]: pam_unix(sshd:session): session closed for user core Dec 13 08:47:46.931407 systemd[1]: sshd@2-64.23.218.120:22-147.75.109.163:34786.service: Deactivated successfully. Dec 13 08:47:46.934434 systemd[1]: session-3.scope: Deactivated successfully. Dec 13 08:47:46.937383 systemd-logind[1448]: Session 3 logged out. Waiting for processes to exit. Dec 13 08:47:46.944090 systemd[1]: Started sshd@3-64.23.218.120:22-147.75.109.163:34796.service - OpenSSH per-connection server daemon (147.75.109.163:34796). Dec 13 08:47:46.946673 systemd-logind[1448]: Removed session 3. Dec 13 08:47:46.998460 sshd[1602]: Accepted publickey for core from 147.75.109.163 port 34796 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:47:47.001924 sshd[1602]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:47:47.012168 systemd-logind[1448]: New session 4 of user core. Dec 13 08:47:47.025235 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 13 08:47:47.091818 sshd[1602]: pam_unix(sshd:session): session closed for user core Dec 13 08:47:47.105537 systemd[1]: sshd@3-64.23.218.120:22-147.75.109.163:34796.service: Deactivated successfully. Dec 13 08:47:47.108193 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 08:47:47.110946 systemd-logind[1448]: Session 4 logged out. Waiting for processes to exit. Dec 13 08:47:47.117517 systemd[1]: Started sshd@4-64.23.218.120:22-147.75.109.163:34812.service - OpenSSH per-connection server daemon (147.75.109.163:34812). Dec 13 08:47:47.120329 systemd-logind[1448]: Removed session 4. Dec 13 08:47:47.181640 sshd[1609]: Accepted publickey for core from 147.75.109.163 port 34812 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:47:47.183925 sshd[1609]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:47:47.192772 systemd-logind[1448]: New session 5 of user core. Dec 13 08:47:47.202122 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 13 08:47:47.289732 sudo[1612]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 08:47:47.290979 sudo[1612]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 13 08:47:47.895409 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 13 08:47:47.898188 (dockerd)[1628]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 13 08:47:48.531697 dockerd[1628]: time="2024-12-13T08:47:48.531603071Z" level=info msg="Starting up" Dec 13 08:47:48.800980 dockerd[1628]: time="2024-12-13T08:47:48.800444538Z" level=info msg="Loading containers: start." Dec 13 08:47:49.002989 kernel: Initializing XFRM netlink socket Dec 13 08:47:49.147714 systemd-networkd[1365]: docker0: Link UP Dec 13 08:47:49.194367 dockerd[1628]: time="2024-12-13T08:47:49.191753468Z" level=info msg="Loading containers: done." Dec 13 08:47:49.226502 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck917054824-merged.mount: Deactivated successfully. Dec 13 08:47:49.230699 dockerd[1628]: time="2024-12-13T08:47:49.230282776Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 13 08:47:49.231601 dockerd[1628]: time="2024-12-13T08:47:49.231082636Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Dec 13 08:47:49.231601 dockerd[1628]: time="2024-12-13T08:47:49.231303770Z" level=info msg="Daemon has completed initialization" Dec 13 08:47:49.354162 dockerd[1628]: time="2024-12-13T08:47:49.353600947Z" level=info msg="API listen on /run/docker.sock" Dec 13 08:47:49.354749 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 13 08:47:49.777715 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 13 08:47:49.786415 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:47:50.033194 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:47:50.038315 (kubelet)[1780]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 08:47:50.144318 kubelet[1780]: E1213 08:47:50.144250 1780 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 08:47:50.152309 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 08:47:50.152626 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 08:47:50.689246 containerd[1457]: time="2024-12-13T08:47:50.689194786Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\"" Dec 13 08:47:50.724099 systemd-resolved[1326]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Dec 13 08:47:51.568985 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2254308093.mount: Deactivated successfully. Dec 13 08:47:53.756900 containerd[1457]: time="2024-12-13T08:47:53.756102956Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:53.768136 containerd[1457]: time="2024-12-13T08:47:53.767492726Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.4: active requests=0, bytes read=27975483" Dec 13 08:47:53.775147 systemd-resolved[1326]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Dec 13 08:47:53.778440 containerd[1457]: time="2024-12-13T08:47:53.777520283Z" level=info msg="ImageCreate event name:\"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:53.790286 containerd[1457]: time="2024-12-13T08:47:53.790158200Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:53.793475 containerd[1457]: time="2024-12-13T08:47:53.793406528Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.4\" with image id \"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\", size \"27972283\" in 3.104155508s" Dec 13 08:47:53.793826 containerd[1457]: time="2024-12-13T08:47:53.793766320Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\" returns image reference \"sha256:bdc2eadbf366279693097982a31da61cc2f1d90f07ada3f4b3b91251a18f665e\"" Dec 13 08:47:53.797115 containerd[1457]: time="2024-12-13T08:47:53.797059853Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\"" Dec 13 08:47:56.001886 containerd[1457]: time="2024-12-13T08:47:56.001527431Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:56.006383 containerd[1457]: time="2024-12-13T08:47:56.005915574Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.4: active requests=0, bytes read=24702157" Dec 13 08:47:56.011312 containerd[1457]: time="2024-12-13T08:47:56.011180711Z" level=info msg="ImageCreate event name:\"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:56.020246 containerd[1457]: time="2024-12-13T08:47:56.020099653Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:56.022908 containerd[1457]: time="2024-12-13T08:47:56.022508965Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.4\" with image id \"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\", size \"26147269\" in 2.225390273s" Dec 13 08:47:56.022908 containerd[1457]: time="2024-12-13T08:47:56.022638022Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\" returns image reference \"sha256:359b9f2307326a4c66172318ca63ee9792c3146ca57d53329239bd123ea70079\"" Dec 13 08:47:56.024096 containerd[1457]: time="2024-12-13T08:47:56.023769870Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\"" Dec 13 08:47:57.603833 containerd[1457]: time="2024-12-13T08:47:57.602019019Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:57.605103 containerd[1457]: time="2024-12-13T08:47:57.605056294Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.4: active requests=0, bytes read=18652067" Dec 13 08:47:57.606985 containerd[1457]: time="2024-12-13T08:47:57.606921143Z" level=info msg="ImageCreate event name:\"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:57.616833 containerd[1457]: time="2024-12-13T08:47:57.616748484Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:57.619208 containerd[1457]: time="2024-12-13T08:47:57.619153187Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.4\" with image id \"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\", size \"20097197\" in 1.595308232s" Dec 13 08:47:57.619407 containerd[1457]: time="2024-12-13T08:47:57.619387184Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\" returns image reference \"sha256:3a66234066fe10fa299c0a52265f90a107450f0372652867118cd9007940d674\"" Dec 13 08:47:57.621258 containerd[1457]: time="2024-12-13T08:47:57.621215655Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\"" Dec 13 08:47:57.626222 systemd-resolved[1326]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Dec 13 08:47:59.087583 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2739161504.mount: Deactivated successfully. Dec 13 08:47:59.881951 containerd[1457]: time="2024-12-13T08:47:59.881878643Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:59.884841 containerd[1457]: time="2024-12-13T08:47:59.884741152Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.4: active requests=0, bytes read=30230243" Dec 13 08:47:59.887879 containerd[1457]: time="2024-12-13T08:47:59.887757980Z" level=info msg="ImageCreate event name:\"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:59.899673 containerd[1457]: time="2024-12-13T08:47:59.899540530Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:47:59.901287 containerd[1457]: time="2024-12-13T08:47:59.900956174Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.4\" with image id \"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\", repo tag \"registry.k8s.io/kube-proxy:v1.31.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\", size \"30229262\" in 2.279283639s" Dec 13 08:47:59.901287 containerd[1457]: time="2024-12-13T08:47:59.901011667Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\" returns image reference \"sha256:ebf80573666f86f115452db568feb34f6f771c3bdc7bfed14b9577f992cfa300\"" Dec 13 08:47:59.902306 containerd[1457]: time="2024-12-13T08:47:59.902044186Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Dec 13 08:48:00.277053 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 13 08:48:00.285201 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:48:00.461471 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:00.474738 (kubelet)[1867]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 13 08:48:00.556965 kubelet[1867]: E1213 08:48:00.556674 1867 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 08:48:00.560549 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 08:48:00.560776 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 08:48:00.763896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1954894567.mount: Deactivated successfully. Dec 13 08:48:02.469843 containerd[1457]: time="2024-12-13T08:48:02.467920842Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:02.473849 containerd[1457]: time="2024-12-13T08:48:02.473714182Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Dec 13 08:48:02.479921 containerd[1457]: time="2024-12-13T08:48:02.479847610Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:02.491831 containerd[1457]: time="2024-12-13T08:48:02.491740009Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:02.494488 containerd[1457]: time="2024-12-13T08:48:02.494411173Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 2.592315886s" Dec 13 08:48:02.494883 containerd[1457]: time="2024-12-13T08:48:02.494833351Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Dec 13 08:48:02.495881 containerd[1457]: time="2024-12-13T08:48:02.495847112Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Dec 13 08:48:03.180723 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2196891717.mount: Deactivated successfully. Dec 13 08:48:03.204120 containerd[1457]: time="2024-12-13T08:48:03.204003372Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:03.213638 containerd[1457]: time="2024-12-13T08:48:03.213540102Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Dec 13 08:48:03.221865 containerd[1457]: time="2024-12-13T08:48:03.220848438Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:03.229220 containerd[1457]: time="2024-12-13T08:48:03.229100484Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:03.231234 containerd[1457]: time="2024-12-13T08:48:03.230962545Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 734.930429ms" Dec 13 08:48:03.231234 containerd[1457]: time="2024-12-13T08:48:03.231092116Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Dec 13 08:48:03.232155 containerd[1457]: time="2024-12-13T08:48:03.232081707Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Dec 13 08:48:03.921674 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1269567475.mount: Deactivated successfully. Dec 13 08:48:06.580294 containerd[1457]: time="2024-12-13T08:48:06.580211287Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:06.587131 containerd[1457]: time="2024-12-13T08:48:06.587061598Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56779973" Dec 13 08:48:06.591876 containerd[1457]: time="2024-12-13T08:48:06.590606108Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:06.600284 containerd[1457]: time="2024-12-13T08:48:06.600183668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:06.602910 containerd[1457]: time="2024-12-13T08:48:06.602838667Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 3.370677391s" Dec 13 08:48:06.603165 containerd[1457]: time="2024-12-13T08:48:06.603136099Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Dec 13 08:48:09.718467 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:09.732864 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:48:09.788015 systemd[1]: Reloading requested from client PID 2003 ('systemctl') (unit session-5.scope)... Dec 13 08:48:09.788042 systemd[1]: Reloading... Dec 13 08:48:09.982888 zram_generator::config[2042]: No configuration found. Dec 13 08:48:10.191188 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 08:48:10.313763 systemd[1]: Reloading finished in 525 ms. Dec 13 08:48:10.381107 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Dec 13 08:48:10.381248 systemd[1]: kubelet.service: Failed with result 'signal'. Dec 13 08:48:10.381652 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:10.390481 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:48:10.592217 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:10.603428 (kubelet)[2094]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 08:48:10.684817 kubelet[2094]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 08:48:10.684817 kubelet[2094]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 08:48:10.684817 kubelet[2094]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 08:48:10.688535 kubelet[2094]: I1213 08:48:10.688020 2094 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 08:48:11.296598 kubelet[2094]: I1213 08:48:11.296505 2094 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Dec 13 08:48:11.296598 kubelet[2094]: I1213 08:48:11.296566 2094 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 08:48:11.297090 kubelet[2094]: I1213 08:48:11.297047 2094 server.go:929] "Client rotation is on, will bootstrap in background" Dec 13 08:48:11.356850 kubelet[2094]: I1213 08:48:11.355892 2094 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 08:48:11.359448 kubelet[2094]: E1213 08:48:11.359403 2094 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://64.23.218.120:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:11.378427 kubelet[2094]: E1213 08:48:11.378153 2094 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Dec 13 08:48:11.378427 kubelet[2094]: I1213 08:48:11.378212 2094 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Dec 13 08:48:11.385790 kubelet[2094]: I1213 08:48:11.385735 2094 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 08:48:11.388762 kubelet[2094]: I1213 08:48:11.388645 2094 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 13 08:48:11.389183 kubelet[2094]: I1213 08:48:11.389113 2094 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 08:48:11.389562 kubelet[2094]: I1213 08:48:11.389191 2094 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.2.1-7-8969f226b5","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 13 08:48:11.389562 kubelet[2094]: I1213 08:48:11.389560 2094 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 08:48:11.389832 kubelet[2094]: I1213 08:48:11.389577 2094 container_manager_linux.go:300] "Creating device plugin manager" Dec 13 08:48:11.389832 kubelet[2094]: I1213 08:48:11.389776 2094 state_mem.go:36] "Initialized new in-memory state store" Dec 13 08:48:11.395627 kubelet[2094]: I1213 08:48:11.395217 2094 kubelet.go:408] "Attempting to sync node with API server" Dec 13 08:48:11.395627 kubelet[2094]: I1213 08:48:11.395299 2094 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 08:48:11.395627 kubelet[2094]: I1213 08:48:11.395355 2094 kubelet.go:314] "Adding apiserver pod source" Dec 13 08:48:11.395627 kubelet[2094]: I1213 08:48:11.395396 2094 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 08:48:11.406276 kubelet[2094]: W1213 08:48:11.406184 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.218.120:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.2.1-7-8969f226b5&limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:11.406562 kubelet[2094]: E1213 08:48:11.406527 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.218.120:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.2.1-7-8969f226b5&limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:11.407660 kubelet[2094]: W1213 08:48:11.407416 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.218.120:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:11.407660 kubelet[2094]: E1213 08:48:11.407497 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.218.120:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:11.408769 kubelet[2094]: I1213 08:48:11.408452 2094 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 13 08:48:11.411706 kubelet[2094]: I1213 08:48:11.411657 2094 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 08:48:11.413849 kubelet[2094]: W1213 08:48:11.413697 2094 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 08:48:11.415533 kubelet[2094]: I1213 08:48:11.415280 2094 server.go:1269] "Started kubelet" Dec 13 08:48:11.418533 kubelet[2094]: I1213 08:48:11.418393 2094 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 08:48:11.422842 kubelet[2094]: I1213 08:48:11.419935 2094 server.go:460] "Adding debug handlers to kubelet server" Dec 13 08:48:11.423376 kubelet[2094]: I1213 08:48:11.423341 2094 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 08:48:11.425860 kubelet[2094]: I1213 08:48:11.425321 2094 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 08:48:11.425860 kubelet[2094]: I1213 08:48:11.425702 2094 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 08:48:11.439118 kubelet[2094]: E1213 08:48:11.429560 2094 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://64.23.218.120:6443/api/v1/namespaces/default/events\": dial tcp 64.23.218.120:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.2.1-7-8969f226b5.1810b04ab30638ae default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.2.1-7-8969f226b5,UID:ci-4081.2.1-7-8969f226b5,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.2.1-7-8969f226b5,},FirstTimestamp:2024-12-13 08:48:11.41523883 +0000 UTC m=+0.806867182,LastTimestamp:2024-12-13 08:48:11.41523883 +0000 UTC m=+0.806867182,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.2.1-7-8969f226b5,}" Dec 13 08:48:11.439823 kubelet[2094]: I1213 08:48:11.439590 2094 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 13 08:48:11.440125 kubelet[2094]: E1213 08:48:11.439991 2094 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.2.1-7-8969f226b5\" not found" Dec 13 08:48:11.443908 kubelet[2094]: I1213 08:48:11.443753 2094 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 13 08:48:11.444064 kubelet[2094]: I1213 08:48:11.444012 2094 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 13 08:48:11.444137 kubelet[2094]: I1213 08:48:11.444120 2094 reconciler.go:26] "Reconciler: start to sync state" Dec 13 08:48:11.445842 kubelet[2094]: W1213 08:48:11.444860 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.218.120:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:11.445842 kubelet[2094]: E1213 08:48:11.445006 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.218.120:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:11.445842 kubelet[2094]: I1213 08:48:11.445167 2094 factory.go:221] Registration of the systemd container factory successfully Dec 13 08:48:11.445842 kubelet[2094]: I1213 08:48:11.445322 2094 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 08:48:11.445842 kubelet[2094]: E1213 08:48:11.445603 2094 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.218.120:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.2.1-7-8969f226b5?timeout=10s\": dial tcp 64.23.218.120:6443: connect: connection refused" interval="200ms" Dec 13 08:48:11.448342 kubelet[2094]: I1213 08:48:11.448295 2094 factory.go:221] Registration of the containerd container factory successfully Dec 13 08:48:11.450260 kubelet[2094]: E1213 08:48:11.450119 2094 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 08:48:11.478116 kubelet[2094]: I1213 08:48:11.477729 2094 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 08:48:11.478116 kubelet[2094]: I1213 08:48:11.477758 2094 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 08:48:11.478116 kubelet[2094]: I1213 08:48:11.477785 2094 state_mem.go:36] "Initialized new in-memory state store" Dec 13 08:48:11.487024 kubelet[2094]: I1213 08:48:11.486978 2094 policy_none.go:49] "None policy: Start" Dec 13 08:48:11.490767 kubelet[2094]: I1213 08:48:11.490631 2094 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 08:48:11.491116 kubelet[2094]: I1213 08:48:11.491066 2094 state_mem.go:35] "Initializing new in-memory state store" Dec 13 08:48:11.501744 kubelet[2094]: I1213 08:48:11.501602 2094 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 08:48:11.504972 kubelet[2094]: I1213 08:48:11.504691 2094 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 08:48:11.504972 kubelet[2094]: I1213 08:48:11.504780 2094 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 08:48:11.504972 kubelet[2094]: I1213 08:48:11.504836 2094 kubelet.go:2321] "Starting kubelet main sync loop" Dec 13 08:48:11.504972 kubelet[2094]: E1213 08:48:11.504930 2094 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 08:48:11.508583 kubelet[2094]: W1213 08:48:11.508408 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.218.120:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:11.508583 kubelet[2094]: E1213 08:48:11.508469 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.218.120:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:11.511788 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 13 08:48:11.528773 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 13 08:48:11.534472 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 13 08:48:11.541273 kubelet[2094]: E1213 08:48:11.541173 2094 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.2.1-7-8969f226b5\" not found" Dec 13 08:48:11.545009 kubelet[2094]: I1213 08:48:11.544968 2094 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 08:48:11.546044 kubelet[2094]: I1213 08:48:11.545518 2094 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 13 08:48:11.546044 kubelet[2094]: I1213 08:48:11.545541 2094 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 08:48:11.548256 kubelet[2094]: I1213 08:48:11.548138 2094 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 08:48:11.559016 kubelet[2094]: E1213 08:48:11.558865 2094 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.2.1-7-8969f226b5\" not found" Dec 13 08:48:11.625475 systemd[1]: Created slice kubepods-burstable-pod622e436345fa15bb38e07e4eec4a3c86.slice - libcontainer container kubepods-burstable-pod622e436345fa15bb38e07e4eec4a3c86.slice. Dec 13 08:48:11.648369 kubelet[2094]: E1213 08:48:11.648316 2094 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.218.120:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.2.1-7-8969f226b5?timeout=10s\": dial tcp 64.23.218.120:6443: connect: connection refused" interval="400ms" Dec 13 08:48:11.649836 kubelet[2094]: I1213 08:48:11.649315 2094 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.649836 kubelet[2094]: I1213 08:48:11.649471 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fc20cd769925e8225cbbf811df9d72ca-kubeconfig\") pod \"kube-scheduler-ci-4081.2.1-7-8969f226b5\" (UID: \"fc20cd769925e8225cbbf811df9d72ca\") " pod="kube-system/kube-scheduler-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.649836 kubelet[2094]: E1213 08:48:11.649759 2094 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.218.120:6443/api/v1/nodes\": dial tcp 64.23.218.120:6443: connect: connection refused" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.650214 kubelet[2094]: I1213 08:48:11.649870 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.650214 kubelet[2094]: I1213 08:48:11.649929 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.650612 kubelet[2094]: I1213 08:48:11.650427 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-k8s-certs\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.653785 kubelet[2094]: I1213 08:48:11.651069 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-kubeconfig\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.651866 systemd[1]: Created slice kubepods-burstable-pod8a8699c44b49427809b746ecea30fd41.slice - libcontainer container kubepods-burstable-pod8a8699c44b49427809b746ecea30fd41.slice. Dec 13 08:48:11.655716 kubelet[2094]: I1213 08:48:11.654624 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.655716 kubelet[2094]: I1213 08:48:11.654691 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-ca-certs\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.655716 kubelet[2094]: I1213 08:48:11.654735 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-ca-certs\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.655716 kubelet[2094]: I1213 08:48:11.654760 2094 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-k8s-certs\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.676576 systemd[1]: Created slice kubepods-burstable-podfc20cd769925e8225cbbf811df9d72ca.slice - libcontainer container kubepods-burstable-podfc20cd769925e8225cbbf811df9d72ca.slice. Dec 13 08:48:11.852009 kubelet[2094]: I1213 08:48:11.851840 2094 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.852963 kubelet[2094]: E1213 08:48:11.852910 2094 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.218.120:6443/api/v1/nodes\": dial tcp 64.23.218.120:6443: connect: connection refused" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:11.941085 kubelet[2094]: E1213 08:48:11.941024 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:11.942560 containerd[1457]: time="2024-12-13T08:48:11.942477957Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.2.1-7-8969f226b5,Uid:622e436345fa15bb38e07e4eec4a3c86,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:11.970864 kubelet[2094]: E1213 08:48:11.969518 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:11.981113 kubelet[2094]: E1213 08:48:11.980743 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:11.981621 containerd[1457]: time="2024-12-13T08:48:11.981573318Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.2.1-7-8969f226b5,Uid:8a8699c44b49427809b746ecea30fd41,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:11.983534 containerd[1457]: time="2024-12-13T08:48:11.983177387Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.2.1-7-8969f226b5,Uid:fc20cd769925e8225cbbf811df9d72ca,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:12.050509 kubelet[2094]: E1213 08:48:12.050424 2094 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.218.120:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.2.1-7-8969f226b5?timeout=10s\": dial tcp 64.23.218.120:6443: connect: connection refused" interval="800ms" Dec 13 08:48:12.259187 kubelet[2094]: I1213 08:48:12.258390 2094 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:12.259187 kubelet[2094]: E1213 08:48:12.258887 2094 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.218.120:6443/api/v1/nodes\": dial tcp 64.23.218.120:6443: connect: connection refused" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:12.328001 kubelet[2094]: W1213 08:48:12.327892 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.218.120:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:12.328363 kubelet[2094]: E1213 08:48:12.328325 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.218.120:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:12.599276 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3596499795.mount: Deactivated successfully. Dec 13 08:48:12.633697 containerd[1457]: time="2024-12-13T08:48:12.632917675Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 08:48:12.638259 containerd[1457]: time="2024-12-13T08:48:12.638162520Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 08:48:12.648512 containerd[1457]: time="2024-12-13T08:48:12.646675929Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 08:48:12.653300 containerd[1457]: time="2024-12-13T08:48:12.653207994Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 08:48:12.655636 containerd[1457]: time="2024-12-13T08:48:12.655427260Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 08:48:12.672621 containerd[1457]: time="2024-12-13T08:48:12.672534996Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Dec 13 08:48:12.676984 containerd[1457]: time="2024-12-13T08:48:12.676878124Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Dec 13 08:48:12.683716 containerd[1457]: time="2024-12-13T08:48:12.683637398Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 13 08:48:12.685452 containerd[1457]: time="2024-12-13T08:48:12.685363890Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 703.450376ms" Dec 13 08:48:12.693413 containerd[1457]: time="2024-12-13T08:48:12.693338018Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 710.048369ms" Dec 13 08:48:12.713990 containerd[1457]: time="2024-12-13T08:48:12.713420429Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 770.809788ms" Dec 13 08:48:12.741370 kubelet[2094]: W1213 08:48:12.741247 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.218.120:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.2.1-7-8969f226b5&limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:12.741820 kubelet[2094]: E1213 08:48:12.741761 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.218.120:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.2.1-7-8969f226b5&limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:12.853318 kubelet[2094]: E1213 08:48:12.853025 2094 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.218.120:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.2.1-7-8969f226b5?timeout=10s\": dial tcp 64.23.218.120:6443: connect: connection refused" interval="1.6s" Dec 13 08:48:12.857645 kubelet[2094]: W1213 08:48:12.857547 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.218.120:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:12.858278 kubelet[2094]: E1213 08:48:12.858232 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.218.120:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:12.907950 kubelet[2094]: W1213 08:48:12.906512 2094 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.218.120:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.218.120:6443: connect: connection refused Dec 13 08:48:12.907950 kubelet[2094]: E1213 08:48:12.906580 2094 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.218.120:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:13.062070 kubelet[2094]: I1213 08:48:13.061348 2094 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:13.062070 kubelet[2094]: E1213 08:48:13.062001 2094 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.218.120:6443/api/v1/nodes\": dial tcp 64.23.218.120:6443: connect: connection refused" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:13.072506 containerd[1457]: time="2024-12-13T08:48:13.071696350Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:13.072506 containerd[1457]: time="2024-12-13T08:48:13.071852744Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:13.072506 containerd[1457]: time="2024-12-13T08:48:13.071881036Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.072506 containerd[1457]: time="2024-12-13T08:48:13.072276000Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.085354 containerd[1457]: time="2024-12-13T08:48:13.084590426Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:13.085354 containerd[1457]: time="2024-12-13T08:48:13.084694197Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:13.085354 containerd[1457]: time="2024-12-13T08:48:13.084724073Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.085354 containerd[1457]: time="2024-12-13T08:48:13.084945921Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.089656 containerd[1457]: time="2024-12-13T08:48:13.088628221Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:13.089656 containerd[1457]: time="2024-12-13T08:48:13.088720486Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:13.089656 containerd[1457]: time="2024-12-13T08:48:13.088742567Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.089656 containerd[1457]: time="2024-12-13T08:48:13.088946935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:13.121219 systemd[1]: Started cri-containerd-884ef9e0c146ba07eb15d36da204b2e76a2a8ec8b9a8dc3ecbfee368bbc82dec.scope - libcontainer container 884ef9e0c146ba07eb15d36da204b2e76a2a8ec8b9a8dc3ecbfee368bbc82dec. Dec 13 08:48:13.148490 systemd[1]: Started cri-containerd-76ab8cf08680b132bf76d5f880a8ea48f0739cecfa24bdf135bded497d066176.scope - libcontainer container 76ab8cf08680b132bf76d5f880a8ea48f0739cecfa24bdf135bded497d066176. Dec 13 08:48:13.159514 systemd[1]: Started cri-containerd-446e6178126d7b8641a5c09b262b2e7bc7ef844b7d5ec70ff206b534298ba7a2.scope - libcontainer container 446e6178126d7b8641a5c09b262b2e7bc7ef844b7d5ec70ff206b534298ba7a2. Dec 13 08:48:13.260952 containerd[1457]: time="2024-12-13T08:48:13.260649013Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.2.1-7-8969f226b5,Uid:fc20cd769925e8225cbbf811df9d72ca,Namespace:kube-system,Attempt:0,} returns sandbox id \"76ab8cf08680b132bf76d5f880a8ea48f0739cecfa24bdf135bded497d066176\"" Dec 13 08:48:13.266286 kubelet[2094]: E1213 08:48:13.266206 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:13.275791 containerd[1457]: time="2024-12-13T08:48:13.275729194Z" level=info msg="CreateContainer within sandbox \"76ab8cf08680b132bf76d5f880a8ea48f0739cecfa24bdf135bded497d066176\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 13 08:48:13.285241 containerd[1457]: time="2024-12-13T08:48:13.285162700Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.2.1-7-8969f226b5,Uid:622e436345fa15bb38e07e4eec4a3c86,Namespace:kube-system,Attempt:0,} returns sandbox id \"884ef9e0c146ba07eb15d36da204b2e76a2a8ec8b9a8dc3ecbfee368bbc82dec\"" Dec 13 08:48:13.287332 kubelet[2094]: E1213 08:48:13.287247 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:13.289580 containerd[1457]: time="2024-12-13T08:48:13.289529733Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.2.1-7-8969f226b5,Uid:8a8699c44b49427809b746ecea30fd41,Namespace:kube-system,Attempt:0,} returns sandbox id \"446e6178126d7b8641a5c09b262b2e7bc7ef844b7d5ec70ff206b534298ba7a2\"" Dec 13 08:48:13.292919 kubelet[2094]: E1213 08:48:13.292584 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:13.294235 containerd[1457]: time="2024-12-13T08:48:13.294147941Z" level=info msg="CreateContainer within sandbox \"884ef9e0c146ba07eb15d36da204b2e76a2a8ec8b9a8dc3ecbfee368bbc82dec\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 13 08:48:13.296070 containerd[1457]: time="2024-12-13T08:48:13.295916951Z" level=info msg="CreateContainer within sandbox \"446e6178126d7b8641a5c09b262b2e7bc7ef844b7d5ec70ff206b534298ba7a2\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 13 08:48:13.378508 containerd[1457]: time="2024-12-13T08:48:13.377238693Z" level=info msg="CreateContainer within sandbox \"884ef9e0c146ba07eb15d36da204b2e76a2a8ec8b9a8dc3ecbfee368bbc82dec\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2592a342a6b592c96da8448dc2663d079f7317dc4e1ed1effd2e457104d958ef\"" Dec 13 08:48:13.382898 containerd[1457]: time="2024-12-13T08:48:13.381594026Z" level=info msg="StartContainer for \"2592a342a6b592c96da8448dc2663d079f7317dc4e1ed1effd2e457104d958ef\"" Dec 13 08:48:13.392364 containerd[1457]: time="2024-12-13T08:48:13.392292075Z" level=info msg="CreateContainer within sandbox \"76ab8cf08680b132bf76d5f880a8ea48f0739cecfa24bdf135bded497d066176\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"87c8c990d0517c2ffd4bde8858d5400b292aa32109828f95aa45a999f039d37e\"" Dec 13 08:48:13.393627 containerd[1457]: time="2024-12-13T08:48:13.393576776Z" level=info msg="StartContainer for \"87c8c990d0517c2ffd4bde8858d5400b292aa32109828f95aa45a999f039d37e\"" Dec 13 08:48:13.397101 containerd[1457]: time="2024-12-13T08:48:13.397045297Z" level=info msg="CreateContainer within sandbox \"446e6178126d7b8641a5c09b262b2e7bc7ef844b7d5ec70ff206b534298ba7a2\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f0f414a3d86b6290f5df828c7eabcd7d065bea35fd3a9da2fec01c12bca63578\"" Dec 13 08:48:13.399832 containerd[1457]: time="2024-12-13T08:48:13.398438308Z" level=info msg="StartContainer for \"f0f414a3d86b6290f5df828c7eabcd7d065bea35fd3a9da2fec01c12bca63578\"" Dec 13 08:48:13.475735 systemd[1]: Started cri-containerd-87c8c990d0517c2ffd4bde8858d5400b292aa32109828f95aa45a999f039d37e.scope - libcontainer container 87c8c990d0517c2ffd4bde8858d5400b292aa32109828f95aa45a999f039d37e. Dec 13 08:48:13.490466 systemd[1]: Started cri-containerd-2592a342a6b592c96da8448dc2663d079f7317dc4e1ed1effd2e457104d958ef.scope - libcontainer container 2592a342a6b592c96da8448dc2663d079f7317dc4e1ed1effd2e457104d958ef. Dec 13 08:48:13.516392 systemd[1]: Started cri-containerd-f0f414a3d86b6290f5df828c7eabcd7d065bea35fd3a9da2fec01c12bca63578.scope - libcontainer container f0f414a3d86b6290f5df828c7eabcd7d065bea35fd3a9da2fec01c12bca63578. Dec 13 08:48:13.529700 kubelet[2094]: E1213 08:48:13.526983 2094 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://64.23.218.120:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 64.23.218.120:6443: connect: connection refused" logger="UnhandledError" Dec 13 08:48:13.650079 containerd[1457]: time="2024-12-13T08:48:13.649916669Z" level=info msg="StartContainer for \"87c8c990d0517c2ffd4bde8858d5400b292aa32109828f95aa45a999f039d37e\" returns successfully" Dec 13 08:48:13.666612 containerd[1457]: time="2024-12-13T08:48:13.666551687Z" level=info msg="StartContainer for \"2592a342a6b592c96da8448dc2663d079f7317dc4e1ed1effd2e457104d958ef\" returns successfully" Dec 13 08:48:13.728757 containerd[1457]: time="2024-12-13T08:48:13.728591427Z" level=info msg="StartContainer for \"f0f414a3d86b6290f5df828c7eabcd7d065bea35fd3a9da2fec01c12bca63578\" returns successfully" Dec 13 08:48:14.572911 kubelet[2094]: E1213 08:48:14.572701 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:14.580036 kubelet[2094]: E1213 08:48:14.579993 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:14.585854 kubelet[2094]: E1213 08:48:14.585110 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:14.665682 kubelet[2094]: I1213 08:48:14.665009 2094 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:15.584825 kubelet[2094]: E1213 08:48:15.584746 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:15.586735 kubelet[2094]: E1213 08:48:15.586683 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:15.589638 kubelet[2094]: E1213 08:48:15.589362 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:16.589156 kubelet[2094]: E1213 08:48:16.589073 2094 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:17.024935 kubelet[2094]: E1213 08:48:17.024751 2094 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.2.1-7-8969f226b5\" not found" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:17.197091 kubelet[2094]: I1213 08:48:17.196904 2094 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:17.197091 kubelet[2094]: E1213 08:48:17.196974 2094 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4081.2.1-7-8969f226b5\": node \"ci-4081.2.1-7-8969f226b5\" not found" Dec 13 08:48:17.411488 kubelet[2094]: I1213 08:48:17.411411 2094 apiserver.go:52] "Watching apiserver" Dec 13 08:48:17.444927 kubelet[2094]: I1213 08:48:17.444748 2094 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 13 08:48:19.621508 systemd[1]: Reloading requested from client PID 2372 ('systemctl') (unit session-5.scope)... Dec 13 08:48:19.621538 systemd[1]: Reloading... Dec 13 08:48:19.795846 zram_generator::config[2414]: No configuration found. Dec 13 08:48:20.069202 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 08:48:20.260394 systemd[1]: Reloading finished in 638 ms. Dec 13 08:48:20.335352 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:48:20.354121 systemd[1]: kubelet.service: Deactivated successfully. Dec 13 08:48:20.354469 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:20.355115 systemd[1]: kubelet.service: Consumed 1.346s CPU time, 112.6M memory peak, 0B memory swap peak. Dec 13 08:48:20.368347 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 13 08:48:20.597165 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 13 08:48:20.609277 (kubelet)[2462]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 13 08:48:20.766313 kubelet[2462]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 08:48:20.766862 kubelet[2462]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 08:48:20.766947 kubelet[2462]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 08:48:20.767213 kubelet[2462]: I1213 08:48:20.767147 2462 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 08:48:20.785897 kubelet[2462]: I1213 08:48:20.784575 2462 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Dec 13 08:48:20.785897 kubelet[2462]: I1213 08:48:20.784626 2462 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 08:48:20.788786 kubelet[2462]: I1213 08:48:20.788736 2462 server.go:929] "Client rotation is on, will bootstrap in background" Dec 13 08:48:20.795162 kubelet[2462]: I1213 08:48:20.795090 2462 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 13 08:48:20.807988 kubelet[2462]: I1213 08:48:20.807683 2462 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 08:48:20.817669 kubelet[2462]: E1213 08:48:20.817584 2462 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Dec 13 08:48:20.818175 kubelet[2462]: I1213 08:48:20.817755 2462 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Dec 13 08:48:20.825516 kubelet[2462]: I1213 08:48:20.825354 2462 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 08:48:20.826952 kubelet[2462]: I1213 08:48:20.825771 2462 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 13 08:48:20.826952 kubelet[2462]: I1213 08:48:20.826072 2462 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 08:48:20.826952 kubelet[2462]: I1213 08:48:20.826133 2462 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.2.1-7-8969f226b5","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 13 08:48:20.826952 kubelet[2462]: I1213 08:48:20.826448 2462 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 08:48:20.827519 kubelet[2462]: I1213 08:48:20.826470 2462 container_manager_linux.go:300] "Creating device plugin manager" Dec 13 08:48:20.827519 kubelet[2462]: I1213 08:48:20.826527 2462 state_mem.go:36] "Initialized new in-memory state store" Dec 13 08:48:20.831849 kubelet[2462]: I1213 08:48:20.831793 2462 kubelet.go:408] "Attempting to sync node with API server" Dec 13 08:48:20.832930 kubelet[2462]: I1213 08:48:20.832097 2462 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 08:48:20.832930 kubelet[2462]: I1213 08:48:20.832162 2462 kubelet.go:314] "Adding apiserver pod source" Dec 13 08:48:20.832930 kubelet[2462]: I1213 08:48:20.832181 2462 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 08:48:20.834524 kubelet[2462]: I1213 08:48:20.834488 2462 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Dec 13 08:48:20.838530 kubelet[2462]: I1213 08:48:20.838078 2462 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 08:48:20.841255 kubelet[2462]: I1213 08:48:20.839472 2462 server.go:1269] "Started kubelet" Dec 13 08:48:20.859717 kubelet[2462]: I1213 08:48:20.859246 2462 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 08:48:20.885293 kubelet[2462]: I1213 08:48:20.885217 2462 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 08:48:20.893897 kubelet[2462]: I1213 08:48:20.893831 2462 server.go:460] "Adding debug handlers to kubelet server" Dec 13 08:48:20.902460 kubelet[2462]: I1213 08:48:20.902378 2462 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 08:48:20.904853 kubelet[2462]: I1213 08:48:20.903764 2462 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 08:48:20.908138 kubelet[2462]: I1213 08:48:20.908087 2462 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 13 08:48:20.911551 kubelet[2462]: I1213 08:48:20.911514 2462 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 13 08:48:20.912175 kubelet[2462]: E1213 08:48:20.912139 2462 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.2.1-7-8969f226b5\" not found" Dec 13 08:48:20.918409 kubelet[2462]: I1213 08:48:20.918367 2462 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 13 08:48:20.918908 kubelet[2462]: I1213 08:48:20.918885 2462 reconciler.go:26] "Reconciler: start to sync state" Dec 13 08:48:20.919136 kubelet[2462]: I1213 08:48:20.919091 2462 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 08:48:20.922458 kubelet[2462]: I1213 08:48:20.921508 2462 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 08:48:20.922458 kubelet[2462]: I1213 08:48:20.921571 2462 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 08:48:20.922458 kubelet[2462]: I1213 08:48:20.921607 2462 kubelet.go:2321] "Starting kubelet main sync loop" Dec 13 08:48:20.922458 kubelet[2462]: E1213 08:48:20.921696 2462 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 08:48:20.932268 kubelet[2462]: I1213 08:48:20.932224 2462 factory.go:221] Registration of the systemd container factory successfully Dec 13 08:48:20.934458 kubelet[2462]: I1213 08:48:20.933705 2462 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 08:48:20.944580 kubelet[2462]: E1213 08:48:20.944533 2462 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 08:48:20.946960 kubelet[2462]: I1213 08:48:20.946917 2462 factory.go:221] Registration of the containerd container factory successfully Dec 13 08:48:21.026839 kubelet[2462]: E1213 08:48:21.026131 2462 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 13 08:48:21.032139 kubelet[2462]: I1213 08:48:21.032058 2462 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 08:48:21.032407 kubelet[2462]: I1213 08:48:21.032383 2462 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 08:48:21.032533 kubelet[2462]: I1213 08:48:21.032519 2462 state_mem.go:36] "Initialized new in-memory state store" Dec 13 08:48:21.033647 kubelet[2462]: I1213 08:48:21.033605 2462 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 13 08:48:21.033872 kubelet[2462]: I1213 08:48:21.033789 2462 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 13 08:48:21.034004 kubelet[2462]: I1213 08:48:21.033988 2462 policy_none.go:49] "None policy: Start" Dec 13 08:48:21.035586 kubelet[2462]: I1213 08:48:21.035550 2462 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 08:48:21.037424 kubelet[2462]: I1213 08:48:21.035944 2462 state_mem.go:35] "Initializing new in-memory state store" Dec 13 08:48:21.037424 kubelet[2462]: I1213 08:48:21.036294 2462 state_mem.go:75] "Updated machine memory state" Dec 13 08:48:21.049493 kubelet[2462]: I1213 08:48:21.048819 2462 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 08:48:21.050057 kubelet[2462]: I1213 08:48:21.049881 2462 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 13 08:48:21.050057 kubelet[2462]: I1213 08:48:21.049913 2462 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 08:48:21.052507 kubelet[2462]: I1213 08:48:21.051344 2462 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 08:48:21.161388 kubelet[2462]: I1213 08:48:21.161228 2462 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.182824 kubelet[2462]: I1213 08:48:21.182473 2462 kubelet_node_status.go:111] "Node was previously registered" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.182824 kubelet[2462]: I1213 08:48:21.182675 2462 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.257653 kubelet[2462]: W1213 08:48:21.256737 2462 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Dec 13 08:48:21.261319 kubelet[2462]: W1213 08:48:21.260450 2462 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Dec 13 08:48:21.261319 kubelet[2462]: W1213 08:48:21.261205 2462 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Dec 13 08:48:21.335970 kubelet[2462]: I1213 08:48:21.335230 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.335970 kubelet[2462]: I1213 08:48:21.335292 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-k8s-certs\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.335970 kubelet[2462]: I1213 08:48:21.335572 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.335970 kubelet[2462]: I1213 08:48:21.335618 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-k8s-certs\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.335970 kubelet[2462]: I1213 08:48:21.335646 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-ca-certs\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.336521 kubelet[2462]: I1213 08:48:21.335672 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-kubeconfig\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.336521 kubelet[2462]: I1213 08:48:21.335699 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8a8699c44b49427809b746ecea30fd41-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.2.1-7-8969f226b5\" (UID: \"8a8699c44b49427809b746ecea30fd41\") " pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.336521 kubelet[2462]: I1213 08:48:21.335725 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fc20cd769925e8225cbbf811df9d72ca-kubeconfig\") pod \"kube-scheduler-ci-4081.2.1-7-8969f226b5\" (UID: \"fc20cd769925e8225cbbf811df9d72ca\") " pod="kube-system/kube-scheduler-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.336521 kubelet[2462]: I1213 08:48:21.335758 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/622e436345fa15bb38e07e4eec4a3c86-ca-certs\") pod \"kube-apiserver-ci-4081.2.1-7-8969f226b5\" (UID: \"622e436345fa15bb38e07e4eec4a3c86\") " pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" Dec 13 08:48:21.559555 kubelet[2462]: E1213 08:48:21.559452 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.561968 kubelet[2462]: E1213 08:48:21.561624 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.561968 kubelet[2462]: E1213 08:48:21.561788 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.834730 kubelet[2462]: I1213 08:48:21.834177 2462 apiserver.go:52] "Watching apiserver" Dec 13 08:48:21.895845 update_engine[1449]: I20241213 08:48:21.895605 1449 update_attempter.cc:509] Updating boot flags... Dec 13 08:48:21.925883 kubelet[2462]: I1213 08:48:21.919751 2462 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 13 08:48:21.974591 kubelet[2462]: E1213 08:48:21.969947 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.974591 kubelet[2462]: E1213 08:48:21.971116 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.978741 kubelet[2462]: E1213 08:48:21.978475 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:21.993914 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2509) Dec 13 08:48:22.261268 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2512) Dec 13 08:48:22.265833 kubelet[2462]: I1213 08:48:22.263838 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.2.1-7-8969f226b5" podStartSLOduration=1.263777964 podStartE2EDuration="1.263777964s" podCreationTimestamp="2024-12-13 08:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:22.118914586 +0000 UTC m=+1.470034334" watchObservedRunningTime="2024-12-13 08:48:22.263777964 +0000 UTC m=+1.614897716" Dec 13 08:48:22.395550 kubelet[2462]: I1213 08:48:22.394938 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.2.1-7-8969f226b5" podStartSLOduration=1.39489483 podStartE2EDuration="1.39489483s" podCreationTimestamp="2024-12-13 08:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:22.271495554 +0000 UTC m=+1.622615307" watchObservedRunningTime="2024-12-13 08:48:22.39489483 +0000 UTC m=+1.746014604" Dec 13 08:48:22.975401 kubelet[2462]: E1213 08:48:22.973087 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:23.104051 sudo[1612]: pam_unix(sudo:session): session closed for user root Dec 13 08:48:23.113604 sshd[1609]: pam_unix(sshd:session): session closed for user core Dec 13 08:48:23.122255 systemd[1]: sshd@4-64.23.218.120:22-147.75.109.163:34812.service: Deactivated successfully. Dec 13 08:48:23.129362 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 08:48:23.130369 systemd[1]: session-5.scope: Consumed 5.091s CPU time, 156.3M memory peak, 0B memory swap peak. Dec 13 08:48:23.131679 systemd-logind[1448]: Session 5 logged out. Waiting for processes to exit. Dec 13 08:48:23.139396 systemd-logind[1448]: Removed session 5. Dec 13 08:48:24.451404 kubelet[2462]: I1213 08:48:24.451170 2462 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 13 08:48:24.453908 kubelet[2462]: I1213 08:48:24.452241 2462 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 13 08:48:24.454199 containerd[1457]: time="2024-12-13T08:48:24.451826709Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 08:48:24.760756 kubelet[2462]: E1213 08:48:24.758866 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:24.790205 kubelet[2462]: I1213 08:48:24.790026 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.2.1-7-8969f226b5" podStartSLOduration=3.789998148 podStartE2EDuration="3.789998148s" podCreationTimestamp="2024-12-13 08:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:22.395740156 +0000 UTC m=+1.746859911" watchObservedRunningTime="2024-12-13 08:48:24.789998148 +0000 UTC m=+4.141117900" Dec 13 08:48:24.978296 kubelet[2462]: E1213 08:48:24.977915 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:25.485497 kubelet[2462]: W1213 08:48:25.485354 2462 reflector.go:561] object-"kube-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4081.2.1-7-8969f226b5" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4081.2.1-7-8969f226b5' and this object Dec 13 08:48:25.485497 kubelet[2462]: E1213 08:48:25.485452 2462 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:ci-4081.2.1-7-8969f226b5\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081.2.1-7-8969f226b5' and this object" logger="UnhandledError" Dec 13 08:48:25.492209 kubelet[2462]: W1213 08:48:25.489875 2462 reflector.go:561] object-"kube-system"/"kube-proxy": failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:ci-4081.2.1-7-8969f226b5" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4081.2.1-7-8969f226b5' and this object Dec 13 08:48:25.492209 kubelet[2462]: E1213 08:48:25.489938 2462 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"kube-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-proxy\" is forbidden: User \"system:node:ci-4081.2.1-7-8969f226b5\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081.2.1-7-8969f226b5' and this object" logger="UnhandledError" Dec 13 08:48:25.493683 systemd[1]: Created slice kubepods-burstable-pod8aba3296_8949_4a5f_802a_43b5179c8048.slice - libcontainer container kubepods-burstable-pod8aba3296_8949_4a5f_802a_43b5179c8048.slice. Dec 13 08:48:25.512830 systemd[1]: Created slice kubepods-besteffort-pod43c02f1d_788c_4f05_948e_3ee10e1c277b.slice - libcontainer container kubepods-besteffort-pod43c02f1d_788c_4f05_948e_3ee10e1c277b.slice. Dec 13 08:48:25.579653 kubelet[2462]: I1213 08:48:25.579591 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/8aba3296-8949-4a5f-802a-43b5179c8048-flannel-cfg\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.580239 kubelet[2462]: I1213 08:48:25.580036 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/43c02f1d-788c-4f05-948e-3ee10e1c277b-kube-proxy\") pod \"kube-proxy-5q97x\" (UID: \"43c02f1d-788c-4f05-948e-3ee10e1c277b\") " pod="kube-system/kube-proxy-5q97x" Dec 13 08:48:25.580239 kubelet[2462]: I1213 08:48:25.580118 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/43c02f1d-788c-4f05-948e-3ee10e1c277b-xtables-lock\") pod \"kube-proxy-5q97x\" (UID: \"43c02f1d-788c-4f05-948e-3ee10e1c277b\") " pod="kube-system/kube-proxy-5q97x" Dec 13 08:48:25.580239 kubelet[2462]: I1213 08:48:25.580148 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjbxq\" (UniqueName: \"kubernetes.io/projected/43c02f1d-788c-4f05-948e-3ee10e1c277b-kube-api-access-bjbxq\") pod \"kube-proxy-5q97x\" (UID: \"43c02f1d-788c-4f05-948e-3ee10e1c277b\") " pod="kube-system/kube-proxy-5q97x" Dec 13 08:48:25.580239 kubelet[2462]: I1213 08:48:25.580208 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvh76\" (UniqueName: \"kubernetes.io/projected/8aba3296-8949-4a5f-802a-43b5179c8048-kube-api-access-kvh76\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.580881 kubelet[2462]: I1213 08:48:25.580570 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43c02f1d-788c-4f05-948e-3ee10e1c277b-lib-modules\") pod \"kube-proxy-5q97x\" (UID: \"43c02f1d-788c-4f05-948e-3ee10e1c277b\") " pod="kube-system/kube-proxy-5q97x" Dec 13 08:48:25.580881 kubelet[2462]: I1213 08:48:25.580664 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8aba3296-8949-4a5f-802a-43b5179c8048-run\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.580881 kubelet[2462]: I1213 08:48:25.580701 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/8aba3296-8949-4a5f-802a-43b5179c8048-cni-plugin\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.580881 kubelet[2462]: I1213 08:48:25.580749 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/8aba3296-8949-4a5f-802a-43b5179c8048-cni\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.581321 kubelet[2462]: I1213 08:48:25.581022 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8aba3296-8949-4a5f-802a-43b5179c8048-xtables-lock\") pod \"kube-flannel-ds-2924x\" (UID: \"8aba3296-8949-4a5f-802a-43b5179c8048\") " pod="kube-flannel/kube-flannel-ds-2924x" Dec 13 08:48:25.807238 kubelet[2462]: E1213 08:48:25.804593 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:25.807445 containerd[1457]: time="2024-12-13T08:48:25.805667133Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-2924x,Uid:8aba3296-8949-4a5f-802a-43b5179c8048,Namespace:kube-flannel,Attempt:0,}" Dec 13 08:48:25.887260 kubelet[2462]: E1213 08:48:25.887189 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:25.900482 containerd[1457]: time="2024-12-13T08:48:25.899970454Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:25.900482 containerd[1457]: time="2024-12-13T08:48:25.900070815Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:25.900482 containerd[1457]: time="2024-12-13T08:48:25.900096029Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:25.900482 containerd[1457]: time="2024-12-13T08:48:25.900238154Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:25.961328 systemd[1]: Started cri-containerd-a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8.scope - libcontainer container a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8. Dec 13 08:48:25.981207 kubelet[2462]: E1213 08:48:25.981045 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:25.981207 kubelet[2462]: E1213 08:48:25.981048 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:26.041040 containerd[1457]: time="2024-12-13T08:48:26.040725112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-2924x,Uid:8aba3296-8949-4a5f-802a-43b5179c8048,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\"" Dec 13 08:48:26.042405 kubelet[2462]: E1213 08:48:26.041902 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:26.046507 containerd[1457]: time="2024-12-13T08:48:26.046363867Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Dec 13 08:48:26.685053 kubelet[2462]: E1213 08:48:26.684999 2462 configmap.go:193] Couldn't get configMap kube-system/kube-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 13 08:48:26.687270 kubelet[2462]: E1213 08:48:26.686187 2462 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/43c02f1d-788c-4f05-948e-3ee10e1c277b-kube-proxy podName:43c02f1d-788c-4f05-948e-3ee10e1c277b nodeName:}" failed. No retries permitted until 2024-12-13 08:48:27.186146915 +0000 UTC m=+6.537266662 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/43c02f1d-788c-4f05-948e-3ee10e1c277b-kube-proxy") pod "kube-proxy-5q97x" (UID: "43c02f1d-788c-4f05-948e-3ee10e1c277b") : failed to sync configmap cache: timed out waiting for the condition Dec 13 08:48:27.195100 kubelet[2462]: E1213 08:48:27.192883 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:27.343828 kubelet[2462]: E1213 08:48:27.327633 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:27.344047 containerd[1457]: time="2024-12-13T08:48:27.343418747Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5q97x,Uid:43c02f1d-788c-4f05-948e-3ee10e1c277b,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:27.412338 containerd[1457]: time="2024-12-13T08:48:27.411724049Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:27.412338 containerd[1457]: time="2024-12-13T08:48:27.411894943Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:27.412338 containerd[1457]: time="2024-12-13T08:48:27.411929086Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:27.412338 containerd[1457]: time="2024-12-13T08:48:27.412169420Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:27.454323 systemd[1]: Started cri-containerd-aee9339a3db18afda90f99c72a27de78bd3089b06ec6db3fcd67565e6fe4a9df.scope - libcontainer container aee9339a3db18afda90f99c72a27de78bd3089b06ec6db3fcd67565e6fe4a9df. Dec 13 08:48:27.499573 containerd[1457]: time="2024-12-13T08:48:27.499090732Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5q97x,Uid:43c02f1d-788c-4f05-948e-3ee10e1c277b,Namespace:kube-system,Attempt:0,} returns sandbox id \"aee9339a3db18afda90f99c72a27de78bd3089b06ec6db3fcd67565e6fe4a9df\"" Dec 13 08:48:27.501007 kubelet[2462]: E1213 08:48:27.500610 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:27.507532 containerd[1457]: time="2024-12-13T08:48:27.507304918Z" level=info msg="CreateContainer within sandbox \"aee9339a3db18afda90f99c72a27de78bd3089b06ec6db3fcd67565e6fe4a9df\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 08:48:27.580733 containerd[1457]: time="2024-12-13T08:48:27.580489228Z" level=info msg="CreateContainer within sandbox \"aee9339a3db18afda90f99c72a27de78bd3089b06ec6db3fcd67565e6fe4a9df\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4a10803dfcfd80f9856eb4250c3e01b9ed1bbe7c17acc6da4a4b1b43af1f8962\"" Dec 13 08:48:27.582866 containerd[1457]: time="2024-12-13T08:48:27.582110992Z" level=info msg="StartContainer for \"4a10803dfcfd80f9856eb4250c3e01b9ed1bbe7c17acc6da4a4b1b43af1f8962\"" Dec 13 08:48:27.630175 systemd[1]: Started cri-containerd-4a10803dfcfd80f9856eb4250c3e01b9ed1bbe7c17acc6da4a4b1b43af1f8962.scope - libcontainer container 4a10803dfcfd80f9856eb4250c3e01b9ed1bbe7c17acc6da4a4b1b43af1f8962. Dec 13 08:48:27.686869 containerd[1457]: time="2024-12-13T08:48:27.686750214Z" level=info msg="StartContainer for \"4a10803dfcfd80f9856eb4250c3e01b9ed1bbe7c17acc6da4a4b1b43af1f8962\" returns successfully" Dec 13 08:48:27.987979 kubelet[2462]: E1213 08:48:27.987931 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:27.989325 kubelet[2462]: E1213 08:48:27.988625 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:28.036885 kubelet[2462]: I1213 08:48:28.036673 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-5q97x" podStartSLOduration=3.036644039 podStartE2EDuration="3.036644039s" podCreationTimestamp="2024-12-13 08:48:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:28.035358496 +0000 UTC m=+7.386478248" watchObservedRunningTime="2024-12-13 08:48:28.036644039 +0000 UTC m=+7.387763786" Dec 13 08:48:28.359784 containerd[1457]: time="2024-12-13T08:48:28.358489409Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:28.363961 containerd[1457]: time="2024-12-13T08:48:28.363865084Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3852937" Dec 13 08:48:28.367201 containerd[1457]: time="2024-12-13T08:48:28.367101591Z" level=info msg="ImageCreate event name:\"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:28.386014 containerd[1457]: time="2024-12-13T08:48:28.380528501Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:28.386384 containerd[1457]: time="2024-12-13T08:48:28.386322197Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3842055\" in 2.33980964s" Dec 13 08:48:28.386510 containerd[1457]: time="2024-12-13T08:48:28.386488867Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\"" Dec 13 08:48:28.395380 containerd[1457]: time="2024-12-13T08:48:28.395331105Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Dec 13 08:48:28.435647 containerd[1457]: time="2024-12-13T08:48:28.435586550Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b\"" Dec 13 08:48:28.437070 containerd[1457]: time="2024-12-13T08:48:28.437019767Z" level=info msg="StartContainer for \"995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b\"" Dec 13 08:48:28.498286 systemd[1]: Started cri-containerd-995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b.scope - libcontainer container 995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b. Dec 13 08:48:28.539762 systemd[1]: cri-containerd-995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b.scope: Deactivated successfully. Dec 13 08:48:28.548429 containerd[1457]: time="2024-12-13T08:48:28.548227150Z" level=info msg="StartContainer for \"995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b\" returns successfully" Dec 13 08:48:28.588296 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b-rootfs.mount: Deactivated successfully. Dec 13 08:48:28.630302 containerd[1457]: time="2024-12-13T08:48:28.630024905Z" level=info msg="shim disconnected" id=995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b namespace=k8s.io Dec 13 08:48:28.630302 containerd[1457]: time="2024-12-13T08:48:28.630175085Z" level=warning msg="cleaning up after shim disconnected" id=995030b9a6bdb5cff584f6a9efd4a5666913a80604b2fe4f9fd7b0efe83eeb4b namespace=k8s.io Dec 13 08:48:28.630302 containerd[1457]: time="2024-12-13T08:48:28.630189213Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 08:48:28.993650 kubelet[2462]: E1213 08:48:28.992470 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:28.995440 containerd[1457]: time="2024-12-13T08:48:28.995354729Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Dec 13 08:48:31.200326 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount640782226.mount: Deactivated successfully. Dec 13 08:48:32.481978 containerd[1457]: time="2024-12-13T08:48:32.481850409Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:32.486495 containerd[1457]: time="2024-12-13T08:48:32.486359821Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26866358" Dec 13 08:48:32.493631 containerd[1457]: time="2024-12-13T08:48:32.493521035Z" level=info msg="ImageCreate event name:\"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:32.503672 containerd[1457]: time="2024-12-13T08:48:32.503358139Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 13 08:48:32.506651 containerd[1457]: time="2024-12-13T08:48:32.506117537Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26855532\" in 3.510680015s" Dec 13 08:48:32.506651 containerd[1457]: time="2024-12-13T08:48:32.506180684Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\"" Dec 13 08:48:32.512461 containerd[1457]: time="2024-12-13T08:48:32.512195274Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Dec 13 08:48:32.575566 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2812360305.mount: Deactivated successfully. Dec 13 08:48:32.585546 containerd[1457]: time="2024-12-13T08:48:32.585445561Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7\"" Dec 13 08:48:32.588597 containerd[1457]: time="2024-12-13T08:48:32.587087141Z" level=info msg="StartContainer for \"0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7\"" Dec 13 08:48:32.639240 systemd[1]: Started cri-containerd-0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7.scope - libcontainer container 0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7. Dec 13 08:48:32.681772 systemd[1]: cri-containerd-0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7.scope: Deactivated successfully. Dec 13 08:48:32.691204 containerd[1457]: time="2024-12-13T08:48:32.691142739Z" level=info msg="StartContainer for \"0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7\" returns successfully" Dec 13 08:48:32.703162 kubelet[2462]: I1213 08:48:32.703097 2462 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Dec 13 08:48:32.747751 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7-rootfs.mount: Deactivated successfully. Dec 13 08:48:32.837643 kubelet[2462]: I1213 08:48:32.837490 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b63fe421-5289-459d-bcc4-3705855e8220-config-volume\") pod \"coredns-6f6b679f8f-b6fmk\" (UID: \"b63fe421-5289-459d-bcc4-3705855e8220\") " pod="kube-system/coredns-6f6b679f8f-b6fmk" Dec 13 08:48:32.837643 kubelet[2462]: I1213 08:48:32.837571 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsmmk\" (UniqueName: \"kubernetes.io/projected/b63fe421-5289-459d-bcc4-3705855e8220-kube-api-access-wsmmk\") pod \"coredns-6f6b679f8f-b6fmk\" (UID: \"b63fe421-5289-459d-bcc4-3705855e8220\") " pod="kube-system/coredns-6f6b679f8f-b6fmk" Dec 13 08:48:32.837643 kubelet[2462]: I1213 08:48:32.837603 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2349356-5fe3-4dc4-b467-7bdcd3062e4f-config-volume\") pod \"coredns-6f6b679f8f-rpq6r\" (UID: \"d2349356-5fe3-4dc4-b467-7bdcd3062e4f\") " pod="kube-system/coredns-6f6b679f8f-rpq6r" Dec 13 08:48:32.837643 kubelet[2462]: I1213 08:48:32.837627 2462 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57blg\" (UniqueName: \"kubernetes.io/projected/d2349356-5fe3-4dc4-b467-7bdcd3062e4f-kube-api-access-57blg\") pod \"coredns-6f6b679f8f-rpq6r\" (UID: \"d2349356-5fe3-4dc4-b467-7bdcd3062e4f\") " pod="kube-system/coredns-6f6b679f8f-rpq6r" Dec 13 08:48:32.864871 systemd[1]: Created slice kubepods-burstable-podd2349356_5fe3_4dc4_b467_7bdcd3062e4f.slice - libcontainer container kubepods-burstable-podd2349356_5fe3_4dc4_b467_7bdcd3062e4f.slice. Dec 13 08:48:32.873356 containerd[1457]: time="2024-12-13T08:48:32.873224346Z" level=info msg="shim disconnected" id=0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7 namespace=k8s.io Dec 13 08:48:32.873356 containerd[1457]: time="2024-12-13T08:48:32.873306735Z" level=warning msg="cleaning up after shim disconnected" id=0fc55c07f854ac722b4e9743cc20862bd2f2516ca088b398052cbeec134c02c7 namespace=k8s.io Dec 13 08:48:32.873356 containerd[1457]: time="2024-12-13T08:48:32.873329311Z" level=info msg="cleaning up dead shim" namespace=k8s.io Dec 13 08:48:32.883155 systemd[1]: Created slice kubepods-burstable-podb63fe421_5289_459d_bcc4_3705855e8220.slice - libcontainer container kubepods-burstable-podb63fe421_5289_459d_bcc4_3705855e8220.slice. Dec 13 08:48:33.007747 kubelet[2462]: E1213 08:48:33.007529 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:33.015322 containerd[1457]: time="2024-12-13T08:48:33.015120536Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Dec 13 08:48:33.068282 containerd[1457]: time="2024-12-13T08:48:33.068205021Z" level=info msg="CreateContainer within sandbox \"a71fcdc9dcdbd7617ff9744c19ef3ec72beb0e422b205eee2b3bd1ebe01835d8\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"c27bc2309665928fbda4fadcc49fa22d1ac57f828aa56fbc14ae95fc93897514\"" Dec 13 08:48:33.072878 containerd[1457]: time="2024-12-13T08:48:33.070748558Z" level=info msg="StartContainer for \"c27bc2309665928fbda4fadcc49fa22d1ac57f828aa56fbc14ae95fc93897514\"" Dec 13 08:48:33.138855 systemd[1]: Started cri-containerd-c27bc2309665928fbda4fadcc49fa22d1ac57f828aa56fbc14ae95fc93897514.scope - libcontainer container c27bc2309665928fbda4fadcc49fa22d1ac57f828aa56fbc14ae95fc93897514. Dec 13 08:48:33.177265 kubelet[2462]: E1213 08:48:33.177167 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:33.179207 containerd[1457]: time="2024-12-13T08:48:33.179128891Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-rpq6r,Uid:d2349356-5fe3-4dc4-b467-7bdcd3062e4f,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:33.190424 kubelet[2462]: E1213 08:48:33.190381 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:33.193477 containerd[1457]: time="2024-12-13T08:48:33.192624096Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-b6fmk,Uid:b63fe421-5289-459d-bcc4-3705855e8220,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:33.212005 containerd[1457]: time="2024-12-13T08:48:33.211891417Z" level=info msg="StartContainer for \"c27bc2309665928fbda4fadcc49fa22d1ac57f828aa56fbc14ae95fc93897514\" returns successfully" Dec 13 08:48:33.316933 containerd[1457]: time="2024-12-13T08:48:33.316790987Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-rpq6r,Uid:d2349356-5fe3-4dc4-b467-7bdcd3062e4f,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"bf9adacb004c8f40504d70f52d5b84c7ce721e6c58d27a96e8a32e5d712b134f\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Dec 13 08:48:33.318047 kubelet[2462]: E1213 08:48:33.317247 2462 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf9adacb004c8f40504d70f52d5b84c7ce721e6c58d27a96e8a32e5d712b134f\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Dec 13 08:48:33.318047 kubelet[2462]: E1213 08:48:33.317343 2462 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf9adacb004c8f40504d70f52d5b84c7ce721e6c58d27a96e8a32e5d712b134f\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-rpq6r" Dec 13 08:48:33.318047 kubelet[2462]: E1213 08:48:33.317374 2462 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bf9adacb004c8f40504d70f52d5b84c7ce721e6c58d27a96e8a32e5d712b134f\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-rpq6r" Dec 13 08:48:33.318047 kubelet[2462]: E1213 08:48:33.317452 2462 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-rpq6r_kube-system(d2349356-5fe3-4dc4-b467-7bdcd3062e4f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-rpq6r_kube-system(d2349356-5fe3-4dc4-b467-7bdcd3062e4f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bf9adacb004c8f40504d70f52d5b84c7ce721e6c58d27a96e8a32e5d712b134f\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-6f6b679f8f-rpq6r" podUID="d2349356-5fe3-4dc4-b467-7bdcd3062e4f" Dec 13 08:48:33.328756 containerd[1457]: time="2024-12-13T08:48:33.328419676Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-b6fmk,Uid:b63fe421-5289-459d-bcc4-3705855e8220,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d60d96d631eb78322399f44dd9cd89eff3b7a2d08418c83d156bc1427b035ee1\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Dec 13 08:48:33.329032 kubelet[2462]: E1213 08:48:33.328883 2462 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d60d96d631eb78322399f44dd9cd89eff3b7a2d08418c83d156bc1427b035ee1\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Dec 13 08:48:33.329032 kubelet[2462]: E1213 08:48:33.328979 2462 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d60d96d631eb78322399f44dd9cd89eff3b7a2d08418c83d156bc1427b035ee1\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-b6fmk" Dec 13 08:48:33.329032 kubelet[2462]: E1213 08:48:33.329013 2462 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d60d96d631eb78322399f44dd9cd89eff3b7a2d08418c83d156bc1427b035ee1\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-b6fmk" Dec 13 08:48:33.329839 kubelet[2462]: E1213 08:48:33.329082 2462 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-b6fmk_kube-system(b63fe421-5289-459d-bcc4-3705855e8220)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-b6fmk_kube-system(b63fe421-5289-459d-bcc4-3705855e8220)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d60d96d631eb78322399f44dd9cd89eff3b7a2d08418c83d156bc1427b035ee1\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-6f6b679f8f-b6fmk" podUID="b63fe421-5289-459d-bcc4-3705855e8220" Dec 13 08:48:34.013155 kubelet[2462]: E1213 08:48:34.013103 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:34.035327 kubelet[2462]: I1213 08:48:34.034224 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-2924x" podStartSLOduration=2.570104785 podStartE2EDuration="9.034196879s" podCreationTimestamp="2024-12-13 08:48:25 +0000 UTC" firstStartedPulling="2024-12-13 08:48:26.044316855 +0000 UTC m=+5.395436653" lastFinishedPulling="2024-12-13 08:48:32.508409016 +0000 UTC m=+11.859528747" observedRunningTime="2024-12-13 08:48:34.033869958 +0000 UTC m=+13.384989713" watchObservedRunningTime="2024-12-13 08:48:34.034196879 +0000 UTC m=+13.385316631" Dec 13 08:48:34.313757 systemd-networkd[1365]: flannel.1: Link UP Dec 13 08:48:34.313773 systemd-networkd[1365]: flannel.1: Gained carrier Dec 13 08:48:35.016424 kubelet[2462]: E1213 08:48:35.015792 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:35.567922 systemd-networkd[1365]: flannel.1: Gained IPv6LL Dec 13 08:48:44.924345 kubelet[2462]: E1213 08:48:44.924282 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:44.926459 containerd[1457]: time="2024-12-13T08:48:44.926399638Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-b6fmk,Uid:b63fe421-5289-459d-bcc4-3705855e8220,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:45.041776 systemd-networkd[1365]: cni0: Link UP Dec 13 08:48:45.041790 systemd-networkd[1365]: cni0: Gained carrier Dec 13 08:48:45.049764 systemd-networkd[1365]: cni0: Lost carrier Dec 13 08:48:45.053161 systemd-networkd[1365]: veth5c67af1c: Link UP Dec 13 08:48:45.055026 kernel: cni0: port 1(veth5c67af1c) entered blocking state Dec 13 08:48:45.055199 kernel: cni0: port 1(veth5c67af1c) entered disabled state Dec 13 08:48:45.056238 kernel: veth5c67af1c: entered allmulticast mode Dec 13 08:48:45.061199 kernel: veth5c67af1c: entered promiscuous mode Dec 13 08:48:45.061306 kernel: cni0: port 1(veth5c67af1c) entered blocking state Dec 13 08:48:45.063035 kernel: cni0: port 1(veth5c67af1c) entered forwarding state Dec 13 08:48:45.063163 kernel: cni0: port 1(veth5c67af1c) entered disabled state Dec 13 08:48:45.079104 kernel: cni0: port 1(veth5c67af1c) entered blocking state Dec 13 08:48:45.079224 kernel: cni0: port 1(veth5c67af1c) entered forwarding state Dec 13 08:48:45.079124 systemd-networkd[1365]: veth5c67af1c: Gained carrier Dec 13 08:48:45.083617 systemd-networkd[1365]: cni0: Gained carrier Dec 13 08:48:45.105137 containerd[1457]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc00001a938), "name":"cbr0", "type":"bridge"} Dec 13 08:48:45.105137 containerd[1457]: delegateAdd: netconf sent to delegate plugin: Dec 13 08:48:45.155691 containerd[1457]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2024-12-13T08:48:45.155140902Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:45.155691 containerd[1457]: time="2024-12-13T08:48:45.155237372Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:45.155691 containerd[1457]: time="2024-12-13T08:48:45.155257049Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:45.155691 containerd[1457]: time="2024-12-13T08:48:45.155386841Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:45.195957 systemd[1]: run-containerd-runc-k8s.io-e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae-runc.lp4P7y.mount: Deactivated successfully. Dec 13 08:48:45.208188 systemd[1]: Started cri-containerd-e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae.scope - libcontainer container e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae. Dec 13 08:48:45.286460 containerd[1457]: time="2024-12-13T08:48:45.286399799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-b6fmk,Uid:b63fe421-5289-459d-bcc4-3705855e8220,Namespace:kube-system,Attempt:0,} returns sandbox id \"e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae\"" Dec 13 08:48:45.297442 kubelet[2462]: E1213 08:48:45.297387 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:45.302197 containerd[1457]: time="2024-12-13T08:48:45.302143922Z" level=info msg="CreateContainer within sandbox \"e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 13 08:48:45.366628 containerd[1457]: time="2024-12-13T08:48:45.366520046Z" level=info msg="CreateContainer within sandbox \"e1b2e80cdfdcf8e88b235caa74b954495b3107e102a7b9c6c898c2bd1d5d88ae\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"eb98cf1099200b175210f9d6d26a4925fb038f6efae3ace413f4bf0b0cdeab90\"" Dec 13 08:48:45.367711 containerd[1457]: time="2024-12-13T08:48:45.367643888Z" level=info msg="StartContainer for \"eb98cf1099200b175210f9d6d26a4925fb038f6efae3ace413f4bf0b0cdeab90\"" Dec 13 08:48:45.417186 systemd[1]: Started cri-containerd-eb98cf1099200b175210f9d6d26a4925fb038f6efae3ace413f4bf0b0cdeab90.scope - libcontainer container eb98cf1099200b175210f9d6d26a4925fb038f6efae3ace413f4bf0b0cdeab90. Dec 13 08:48:45.468592 containerd[1457]: time="2024-12-13T08:48:45.468417440Z" level=info msg="StartContainer for \"eb98cf1099200b175210f9d6d26a4925fb038f6efae3ace413f4bf0b0cdeab90\" returns successfully" Dec 13 08:48:45.923736 kubelet[2462]: E1213 08:48:45.923059 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:45.925435 containerd[1457]: time="2024-12-13T08:48:45.925380182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-rpq6r,Uid:d2349356-5fe3-4dc4-b467-7bdcd3062e4f,Namespace:kube-system,Attempt:0,}" Dec 13 08:48:46.001478 kernel: cni0: port 2(veth08a23009) entered blocking state Dec 13 08:48:46.004324 kernel: cni0: port 2(veth08a23009) entered disabled state Dec 13 08:48:46.004405 kernel: veth08a23009: entered allmulticast mode Dec 13 08:48:46.007062 kernel: veth08a23009: entered promiscuous mode Dec 13 08:48:46.011213 kernel: cni0: port 2(veth08a23009) entered blocking state Dec 13 08:48:46.011362 kernel: cni0: port 2(veth08a23009) entered forwarding state Dec 13 08:48:46.010737 systemd-networkd[1365]: veth08a23009: Link UP Dec 13 08:48:46.032145 systemd-networkd[1365]: veth08a23009: Gained carrier Dec 13 08:48:46.036983 containerd[1457]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc000020938), "name":"cbr0", "type":"bridge"} Dec 13 08:48:46.036983 containerd[1457]: delegateAdd: netconf sent to delegate plugin: Dec 13 08:48:46.070977 kubelet[2462]: E1213 08:48:46.070439 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:46.084943 containerd[1457]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2024-12-13T08:48:46.084483341Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 08:48:46.084943 containerd[1457]: time="2024-12-13T08:48:46.084655910Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 08:48:46.084943 containerd[1457]: time="2024-12-13T08:48:46.084683382Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:46.096508 containerd[1457]: time="2024-12-13T08:48:46.096255062Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 08:48:46.133834 kubelet[2462]: I1213 08:48:46.131802 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-b6fmk" podStartSLOduration=21.131774649 podStartE2EDuration="21.131774649s" podCreationTimestamp="2024-12-13 08:48:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:46.120830363 +0000 UTC m=+25.471950115" watchObservedRunningTime="2024-12-13 08:48:46.131774649 +0000 UTC m=+25.482894409" Dec 13 08:48:46.153172 systemd[1]: Started cri-containerd-531a2101bfd51c1eb178d1781d1ad3e32a73a6d6f13e4ffaaeeb03b94d178cd2.scope - libcontainer container 531a2101bfd51c1eb178d1781d1ad3e32a73a6d6f13e4ffaaeeb03b94d178cd2. Dec 13 08:48:46.254715 containerd[1457]: time="2024-12-13T08:48:46.254645938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-rpq6r,Uid:d2349356-5fe3-4dc4-b467-7bdcd3062e4f,Namespace:kube-system,Attempt:0,} returns sandbox id \"531a2101bfd51c1eb178d1781d1ad3e32a73a6d6f13e4ffaaeeb03b94d178cd2\"" Dec 13 08:48:46.254993 systemd-networkd[1365]: cni0: Gained IPv6LL Dec 13 08:48:46.256287 systemd-networkd[1365]: veth5c67af1c: Gained IPv6LL Dec 13 08:48:46.282070 kubelet[2462]: E1213 08:48:46.281973 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:46.288581 containerd[1457]: time="2024-12-13T08:48:46.288401863Z" level=info msg="CreateContainer within sandbox \"531a2101bfd51c1eb178d1781d1ad3e32a73a6d6f13e4ffaaeeb03b94d178cd2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 13 08:48:46.344255 containerd[1457]: time="2024-12-13T08:48:46.344176434Z" level=info msg="CreateContainer within sandbox \"531a2101bfd51c1eb178d1781d1ad3e32a73a6d6f13e4ffaaeeb03b94d178cd2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"305e52aa5da49d114136d09d0ce4fb1f8a81d44b3e16e37c650031379e36e55a\"" Dec 13 08:48:46.346835 containerd[1457]: time="2024-12-13T08:48:46.345522818Z" level=info msg="StartContainer for \"305e52aa5da49d114136d09d0ce4fb1f8a81d44b3e16e37c650031379e36e55a\"" Dec 13 08:48:46.393248 systemd[1]: Started cri-containerd-305e52aa5da49d114136d09d0ce4fb1f8a81d44b3e16e37c650031379e36e55a.scope - libcontainer container 305e52aa5da49d114136d09d0ce4fb1f8a81d44b3e16e37c650031379e36e55a. Dec 13 08:48:46.442738 containerd[1457]: time="2024-12-13T08:48:46.442623231Z" level=info msg="StartContainer for \"305e52aa5da49d114136d09d0ce4fb1f8a81d44b3e16e37c650031379e36e55a\" returns successfully" Dec 13 08:48:47.074644 kubelet[2462]: E1213 08:48:47.074563 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:47.076944 kubelet[2462]: E1213 08:48:47.075438 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:47.098396 kubelet[2462]: I1213 08:48:47.097152 2462 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-rpq6r" podStartSLOduration=22.097124975 podStartE2EDuration="22.097124975s" podCreationTimestamp="2024-12-13 08:48:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 08:48:47.096893399 +0000 UTC m=+26.448013155" watchObservedRunningTime="2024-12-13 08:48:47.097124975 +0000 UTC m=+26.448244726" Dec 13 08:48:47.663181 systemd-networkd[1365]: veth08a23009: Gained IPv6LL Dec 13 08:48:48.077414 kubelet[2462]: E1213 08:48:48.076929 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:48.077414 kubelet[2462]: E1213 08:48:48.077267 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:48:49.080578 kubelet[2462]: E1213 08:48:49.080375 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:08.994031 systemd[1]: Started sshd@5-64.23.218.120:22-147.75.109.163:57498.service - OpenSSH per-connection server daemon (147.75.109.163:57498). Dec 13 08:49:09.067425 sshd[3449]: Accepted publickey for core from 147.75.109.163 port 57498 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:09.070612 sshd[3449]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:09.079216 systemd-logind[1448]: New session 6 of user core. Dec 13 08:49:09.093662 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 13 08:49:09.281288 sshd[3449]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:09.287602 systemd[1]: sshd@5-64.23.218.120:22-147.75.109.163:57498.service: Deactivated successfully. Dec 13 08:49:09.291697 systemd[1]: session-6.scope: Deactivated successfully. Dec 13 08:49:09.295215 systemd-logind[1448]: Session 6 logged out. Waiting for processes to exit. Dec 13 08:49:09.297233 systemd-logind[1448]: Removed session 6. Dec 13 08:49:14.307413 systemd[1]: Started sshd@6-64.23.218.120:22-147.75.109.163:57514.service - OpenSSH per-connection server daemon (147.75.109.163:57514). Dec 13 08:49:14.392513 sshd[3484]: Accepted publickey for core from 147.75.109.163 port 57514 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:14.395571 sshd[3484]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:14.406025 systemd-logind[1448]: New session 7 of user core. Dec 13 08:49:14.415205 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 13 08:49:14.606591 sshd[3484]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:14.614002 systemd[1]: sshd@6-64.23.218.120:22-147.75.109.163:57514.service: Deactivated successfully. Dec 13 08:49:14.616949 systemd[1]: session-7.scope: Deactivated successfully. Dec 13 08:49:14.619170 systemd-logind[1448]: Session 7 logged out. Waiting for processes to exit. Dec 13 08:49:14.620528 systemd-logind[1448]: Removed session 7. Dec 13 08:49:19.624574 systemd[1]: Started sshd@7-64.23.218.120:22-147.75.109.163:45282.service - OpenSSH per-connection server daemon (147.75.109.163:45282). Dec 13 08:49:19.704742 sshd[3523]: Accepted publickey for core from 147.75.109.163 port 45282 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:19.708189 sshd[3523]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:19.725160 systemd-logind[1448]: New session 8 of user core. Dec 13 08:49:19.735310 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 13 08:49:19.944572 sshd[3523]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:19.961067 systemd[1]: sshd@7-64.23.218.120:22-147.75.109.163:45282.service: Deactivated successfully. Dec 13 08:49:19.968022 systemd[1]: session-8.scope: Deactivated successfully. Dec 13 08:49:19.972051 systemd-logind[1448]: Session 8 logged out. Waiting for processes to exit. Dec 13 08:49:19.983906 systemd[1]: Started sshd@8-64.23.218.120:22-147.75.109.163:45286.service - OpenSSH per-connection server daemon (147.75.109.163:45286). Dec 13 08:49:19.990629 systemd-logind[1448]: Removed session 8. Dec 13 08:49:20.052113 sshd[3554]: Accepted publickey for core from 147.75.109.163 port 45286 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:20.054723 sshd[3554]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:20.063003 systemd-logind[1448]: New session 9 of user core. Dec 13 08:49:20.069255 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 13 08:49:20.308043 sshd[3554]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:20.324707 systemd[1]: sshd@8-64.23.218.120:22-147.75.109.163:45286.service: Deactivated successfully. Dec 13 08:49:20.330183 systemd[1]: session-9.scope: Deactivated successfully. Dec 13 08:49:20.331754 systemd-logind[1448]: Session 9 logged out. Waiting for processes to exit. Dec 13 08:49:20.344427 systemd[1]: Started sshd@9-64.23.218.120:22-147.75.109.163:45302.service - OpenSSH per-connection server daemon (147.75.109.163:45302). Dec 13 08:49:20.347750 systemd-logind[1448]: Removed session 9. Dec 13 08:49:20.457374 sshd[3566]: Accepted publickey for core from 147.75.109.163 port 45302 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:20.459078 sshd[3566]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:20.471145 systemd-logind[1448]: New session 10 of user core. Dec 13 08:49:20.475822 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 13 08:49:20.651906 sshd[3566]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:20.658399 systemd[1]: sshd@9-64.23.218.120:22-147.75.109.163:45302.service: Deactivated successfully. Dec 13 08:49:20.663747 systemd[1]: session-10.scope: Deactivated successfully. Dec 13 08:49:20.665832 systemd-logind[1448]: Session 10 logged out. Waiting for processes to exit. Dec 13 08:49:20.667737 systemd-logind[1448]: Removed session 10. Dec 13 08:49:25.312352 systemd[1]: Started sshd@10-64.23.218.120:22-60.191.20.210:23456.service - OpenSSH per-connection server daemon (60.191.20.210:23456). Dec 13 08:49:25.686775 systemd[1]: Started sshd@11-64.23.218.120:22-147.75.109.163:45304.service - OpenSSH per-connection server daemon (147.75.109.163:45304). Dec 13 08:49:25.745718 sshd[3604]: Accepted publickey for core from 147.75.109.163 port 45304 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:25.749587 sshd[3604]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:25.761214 systemd-logind[1448]: New session 11 of user core. Dec 13 08:49:25.766260 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 13 08:49:25.965877 sshd[3604]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:25.972647 systemd[1]: sshd@11-64.23.218.120:22-147.75.109.163:45304.service: Deactivated successfully. Dec 13 08:49:25.976566 systemd[1]: session-11.scope: Deactivated successfully. Dec 13 08:49:25.979439 systemd-logind[1448]: Session 11 logged out. Waiting for processes to exit. Dec 13 08:49:25.982508 systemd-logind[1448]: Removed session 11. Dec 13 08:49:30.993076 systemd[1]: Started sshd@12-64.23.218.120:22-147.75.109.163:34676.service - OpenSSH per-connection server daemon (147.75.109.163:34676). Dec 13 08:49:31.051084 sshd[3640]: Accepted publickey for core from 147.75.109.163 port 34676 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:31.053509 sshd[3640]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:31.064094 systemd-logind[1448]: New session 12 of user core. Dec 13 08:49:31.071206 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 13 08:49:31.245783 sshd[3640]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:31.251676 systemd-logind[1448]: Session 12 logged out. Waiting for processes to exit. Dec 13 08:49:31.252475 systemd[1]: sshd@12-64.23.218.120:22-147.75.109.163:34676.service: Deactivated successfully. Dec 13 08:49:31.256226 systemd[1]: session-12.scope: Deactivated successfully. Dec 13 08:49:31.258370 systemd-logind[1448]: Removed session 12. Dec 13 08:49:34.942170 kubelet[2462]: E1213 08:49:34.942100 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:35.923074 kubelet[2462]: E1213 08:49:35.923021 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:36.267387 systemd[1]: Started sshd@13-64.23.218.120:22-147.75.109.163:58922.service - OpenSSH per-connection server daemon (147.75.109.163:58922). Dec 13 08:49:36.353558 sshd[3673]: Accepted publickey for core from 147.75.109.163 port 58922 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:36.356671 sshd[3673]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:36.365735 systemd-logind[1448]: New session 13 of user core. Dec 13 08:49:36.374165 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 13 08:49:36.554218 sshd[3673]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:36.564334 systemd[1]: sshd@13-64.23.218.120:22-147.75.109.163:58922.service: Deactivated successfully. Dec 13 08:49:36.568339 systemd[1]: session-13.scope: Deactivated successfully. Dec 13 08:49:36.569810 systemd-logind[1448]: Session 13 logged out. Waiting for processes to exit. Dec 13 08:49:36.571734 systemd-logind[1448]: Removed session 13. Dec 13 08:49:41.577478 systemd[1]: Started sshd@14-64.23.218.120:22-147.75.109.163:58930.service - OpenSSH per-connection server daemon (147.75.109.163:58930). Dec 13 08:49:41.636439 sshd[3707]: Accepted publickey for core from 147.75.109.163 port 58930 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:41.639142 sshd[3707]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:41.647400 systemd-logind[1448]: New session 14 of user core. Dec 13 08:49:41.657217 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 13 08:49:41.834214 sshd[3707]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:41.848441 systemd[1]: sshd@14-64.23.218.120:22-147.75.109.163:58930.service: Deactivated successfully. Dec 13 08:49:41.853746 systemd[1]: session-14.scope: Deactivated successfully. Dec 13 08:49:41.857848 systemd-logind[1448]: Session 14 logged out. Waiting for processes to exit. Dec 13 08:49:41.859874 systemd-logind[1448]: Removed session 14. Dec 13 08:49:41.866513 systemd[1]: Started sshd@15-64.23.218.120:22-147.75.109.163:58934.service - OpenSSH per-connection server daemon (147.75.109.163:58934). Dec 13 08:49:41.931855 sshd[3719]: Accepted publickey for core from 147.75.109.163 port 58934 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:41.934153 sshd[3719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:41.945261 systemd-logind[1448]: New session 15 of user core. Dec 13 08:49:41.949249 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 13 08:49:42.543141 sshd[3719]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:42.554608 systemd[1]: sshd@15-64.23.218.120:22-147.75.109.163:58934.service: Deactivated successfully. Dec 13 08:49:42.560662 systemd[1]: session-15.scope: Deactivated successfully. Dec 13 08:49:42.562514 systemd-logind[1448]: Session 15 logged out. Waiting for processes to exit. Dec 13 08:49:42.571424 systemd[1]: Started sshd@16-64.23.218.120:22-147.75.109.163:58948.service - OpenSSH per-connection server daemon (147.75.109.163:58948). Dec 13 08:49:42.576313 systemd-logind[1448]: Removed session 15. Dec 13 08:49:42.663862 sshd[3730]: Accepted publickey for core from 147.75.109.163 port 58948 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:42.666243 sshd[3730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:42.674872 systemd-logind[1448]: New session 16 of user core. Dec 13 08:49:42.682214 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 13 08:49:44.779849 sshd[3730]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:44.801008 systemd[1]: sshd@16-64.23.218.120:22-147.75.109.163:58948.service: Deactivated successfully. Dec 13 08:49:44.808590 systemd[1]: session-16.scope: Deactivated successfully. Dec 13 08:49:44.816183 systemd-logind[1448]: Session 16 logged out. Waiting for processes to exit. Dec 13 08:49:44.827314 systemd[1]: Started sshd@17-64.23.218.120:22-147.75.109.163:58960.service - OpenSSH per-connection server daemon (147.75.109.163:58960). Dec 13 08:49:44.832474 systemd-logind[1448]: Removed session 16. Dec 13 08:49:44.897182 sshd[3753]: Accepted publickey for core from 147.75.109.163 port 58960 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:44.900147 sshd[3753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:44.908588 systemd-logind[1448]: New session 17 of user core. Dec 13 08:49:44.920229 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 13 08:49:45.311337 sshd[3753]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:45.332937 systemd[1]: sshd@17-64.23.218.120:22-147.75.109.163:58960.service: Deactivated successfully. Dec 13 08:49:45.338424 systemd[1]: session-17.scope: Deactivated successfully. Dec 13 08:49:45.340178 systemd-logind[1448]: Session 17 logged out. Waiting for processes to exit. Dec 13 08:49:45.350591 systemd[1]: Started sshd@18-64.23.218.120:22-147.75.109.163:58976.service - OpenSSH per-connection server daemon (147.75.109.163:58976). Dec 13 08:49:45.354695 systemd-logind[1448]: Removed session 17. Dec 13 08:49:45.430759 sshd[3781]: Accepted publickey for core from 147.75.109.163 port 58976 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:45.433488 sshd[3781]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:45.441238 systemd-logind[1448]: New session 18 of user core. Dec 13 08:49:45.450235 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 13 08:49:45.627985 sshd[3781]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:45.635535 systemd[1]: sshd@18-64.23.218.120:22-147.75.109.163:58976.service: Deactivated successfully. Dec 13 08:49:45.640415 systemd[1]: session-18.scope: Deactivated successfully. Dec 13 08:49:45.642123 systemd-logind[1448]: Session 18 logged out. Waiting for processes to exit. Dec 13 08:49:45.643956 systemd-logind[1448]: Removed session 18. Dec 13 08:49:48.933029 kubelet[2462]: E1213 08:49:48.932959 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:49.923399 kubelet[2462]: E1213 08:49:49.923253 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:50.665646 systemd[1]: Started sshd@19-64.23.218.120:22-147.75.109.163:39188.service - OpenSSH per-connection server daemon (147.75.109.163:39188). Dec 13 08:49:50.731593 sshd[3814]: Accepted publickey for core from 147.75.109.163 port 39188 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:50.734296 sshd[3814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:50.742525 systemd-logind[1448]: New session 19 of user core. Dec 13 08:49:50.751161 systemd[1]: Started session-19.scope - Session 19 of User core. Dec 13 08:49:50.924458 sshd[3814]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:50.932732 systemd[1]: sshd@19-64.23.218.120:22-147.75.109.163:39188.service: Deactivated successfully. Dec 13 08:49:50.936773 systemd[1]: session-19.scope: Deactivated successfully. Dec 13 08:49:50.938850 systemd-logind[1448]: Session 19 logged out. Waiting for processes to exit. Dec 13 08:49:50.940817 systemd-logind[1448]: Removed session 19. Dec 13 08:49:51.923148 kubelet[2462]: E1213 08:49:51.923025 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:53.923668 kubelet[2462]: E1213 08:49:53.923619 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:54.925069 kubelet[2462]: E1213 08:49:54.924349 2462 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Dec 13 08:49:55.949854 systemd[1]: Started sshd@20-64.23.218.120:22-147.75.109.163:39194.service - OpenSSH per-connection server daemon (147.75.109.163:39194). Dec 13 08:49:56.016140 sshd[3850]: Accepted publickey for core from 147.75.109.163 port 39194 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:49:56.018932 sshd[3850]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:49:56.026313 systemd-logind[1448]: New session 20 of user core. Dec 13 08:49:56.033112 systemd[1]: Started session-20.scope - Session 20 of User core. Dec 13 08:49:56.191501 sshd[3850]: pam_unix(sshd:session): session closed for user core Dec 13 08:49:56.197231 systemd-logind[1448]: Session 20 logged out. Waiting for processes to exit. Dec 13 08:49:56.198684 systemd[1]: sshd@20-64.23.218.120:22-147.75.109.163:39194.service: Deactivated successfully. Dec 13 08:49:56.203302 systemd[1]: session-20.scope: Deactivated successfully. Dec 13 08:49:56.206055 systemd-logind[1448]: Removed session 20. Dec 13 08:49:57.272482 sshd[3602]: kex_exchange_identification: read: Connection reset by peer Dec 13 08:49:57.272482 sshd[3602]: Connection reset by 60.191.20.210 port 23456 Dec 13 08:49:57.274522 systemd[1]: sshd@10-64.23.218.120:22-60.191.20.210:23456.service: Deactivated successfully. Dec 13 08:50:01.211475 systemd[1]: Started sshd@21-64.23.218.120:22-147.75.109.163:41930.service - OpenSSH per-connection server daemon (147.75.109.163:41930). Dec 13 08:50:01.269005 sshd[3888]: Accepted publickey for core from 147.75.109.163 port 41930 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:50:01.271627 sshd[3888]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:50:01.280148 systemd-logind[1448]: New session 21 of user core. Dec 13 08:50:01.292192 systemd[1]: Started session-21.scope - Session 21 of User core. Dec 13 08:50:01.455260 sshd[3888]: pam_unix(sshd:session): session closed for user core Dec 13 08:50:01.462169 systemd-logind[1448]: Session 21 logged out. Waiting for processes to exit. Dec 13 08:50:01.463015 systemd[1]: sshd@21-64.23.218.120:22-147.75.109.163:41930.service: Deactivated successfully. Dec 13 08:50:01.466306 systemd[1]: session-21.scope: Deactivated successfully. Dec 13 08:50:01.468230 systemd-logind[1448]: Removed session 21. Dec 13 08:50:06.479394 systemd[1]: Started sshd@22-64.23.218.120:22-147.75.109.163:35062.service - OpenSSH per-connection server daemon (147.75.109.163:35062). Dec 13 08:50:06.546322 sshd[3922]: Accepted publickey for core from 147.75.109.163 port 35062 ssh2: RSA SHA256:GmRBCjv5DLbtT++ktFQz5R9M6+onrAQ9dTcgZ+NRPZM Dec 13 08:50:06.549077 sshd[3922]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 13 08:50:06.564213 systemd-logind[1448]: New session 22 of user core. Dec 13 08:50:06.569136 systemd[1]: Started session-22.scope - Session 22 of User core. Dec 13 08:50:06.745875 sshd[3922]: pam_unix(sshd:session): session closed for user core Dec 13 08:50:06.752311 systemd[1]: sshd@22-64.23.218.120:22-147.75.109.163:35062.service: Deactivated successfully. Dec 13 08:50:06.756099 systemd[1]: session-22.scope: Deactivated successfully. Dec 13 08:50:06.758288 systemd-logind[1448]: Session 22 logged out. Waiting for processes to exit. Dec 13 08:50:06.760994 systemd-logind[1448]: Removed session 22.